fix anthropic llm caller routing (#2241)
This commit is contained in:
@@ -676,7 +676,7 @@ class LLMCaller:
|
|||||||
timeout: float = settings.LLM_CONFIG_TIMEOUT,
|
timeout: float = settings.LLM_CONFIG_TIMEOUT,
|
||||||
**active_parameters: dict[str, Any],
|
**active_parameters: dict[str, Any],
|
||||||
) -> ModelResponse | CustomStreamWrapper | AnthropicMessage:
|
) -> ModelResponse | CustomStreamWrapper | AnthropicMessage:
|
||||||
if self.llm_key and self.llm_key.startswith("ANTHROPIC"):
|
if self.llm_key and "ANTHROPIC" in self.llm_key:
|
||||||
return await self._call_anthropic(messages, tools, timeout, **active_parameters)
|
return await self._call_anthropic(messages, tools, timeout, **active_parameters)
|
||||||
|
|
||||||
return await litellm.acompletion(
|
return await litellm.acompletion(
|
||||||
|
|||||||
Reference in New Issue
Block a user