use anthropic beta client (#2235)

This commit is contained in:
Shuchang Zheng
2025-04-28 16:24:12 +08:00
committed by GitHub
parent 2f0e6e5eb3
commit a8d237deee
3 changed files with 20 additions and 8 deletions

View File

@@ -1433,7 +1433,6 @@ class ForgeAgent:
raw_response=True,
betas=["computer-use-2025-01-24"],
)
LOG.info("Anthropic response", llm_response=llm_response)
assistant_content = llm_response["content"]
llm_caller.message_history.append({"role": "assistant", "content": assistant_content})

View File

@@ -6,7 +6,7 @@ from typing import Any
import litellm
import structlog
from anthropic.types.message import Message as AnthropicMessage
from anthropic.types.beta.beta_message import BetaMessage as AnthropicMessage
from jinja2 import Template
from litellm.utils import CustomStreamWrapper, ModelResponse
@@ -678,7 +678,7 @@ class LLMCaller:
) -> AnthropicMessage:
max_tokens = active_parameters.get("max_completion_tokens") or active_parameters.get("max_tokens") or 4096
model_name = self.llm_config.model_name.replace("bedrock/", "").replace("anthropic/", "")
return await app.ANTHROPIC_CLIENT.messages.create(
response = await app.ANTHROPIC_CLIENT.beta.messages.create(
max_tokens=max_tokens,
messages=messages,
model=model_name,
@@ -686,6 +686,8 @@ class LLMCaller:
timeout=timeout,
betas=active_parameters.get("betas", None),
)
LOG.info("Anthropic response", response=response)
return response
class LLMCallerManager: