Jon/model name massage (#2559)

This commit is contained in:
Shuchang Zheng
2025-05-31 19:34:30 -07:00
committed by GitHub
parent b4d5837196
commit 2167d88c20
11 changed files with 67 additions and 60 deletions

View File

@@ -264,7 +264,7 @@ class Settings(BaseSettings):
SKYVERN_BASE_URL: str = "https://api.skyvern.com"
SKYVERN_API_KEY: str = "PLACEHOLDER"
def get_model_name_to_llm_key(self) -> dict[str, str]:
def get_model_name_to_llm_key(self) -> dict[str, dict[str, str]]:
"""
Keys are model names available to blocks in the frontend. These map to key names
in LLMConfigRegistry._configs.
@@ -272,22 +272,40 @@ class Settings(BaseSettings):
if self.is_cloud_environment():
return {
"Gemini 2.5": "GEMINI_2.5_PRO_PREVIEW",
"Gemini 2.5 Flash": "VERTEX_GEMINI_2.5_FLASH_PREVIEW_05_20",
"GPT 4.1": "OPENAI_GPT4_1",
"GPT o3-mini": "OPENAI_O3_MINI",
"bedrock/us.anthropic.claude-opus-4-20250514-v1:0": "BEDROCK_ANTHROPIC_CLAUDE4_OPUS_INFERENCE_PROFILE",
"bedrock/us.anthropic.claude-sonnet-4-20250514-v1:0": "BEDROCK_ANTHROPIC_CLAUDE4_SONNET_INFERENCE_PROFILE",
"gemini-2.5-pro-preview-05-06": {"llm_key": "VERTEX_GEMINI_2.5_PRO_PREVIEW", "label": "Gemini 2.5 Pro"},
"gemini-2.5-flash-preview-05-20": {
"llm_key": "VERTEX_GEMINI_2.5_FLASH_PREVIEW_05_20",
"label": "Gemini 2.5 Flash",
},
"azure/gpt-4.1": {"llm_key": "AZURE_OPENAI_GPT4_1", "label": "GPT 4.1"},
"azure/o3-mini": {"llm_key": "AZURE_OPENAI_O3_MINI", "label": "GPT O3 Mini"},
"us.anthropic.claude-opus-4-20250514-v1:0": {
"llm_key": "BEDROCK_ANTHROPIC_CLAUDE4_OPUS_INFERENCE_PROFILE",
"label": "Anthropic Claude 4 Opus",
},
"us.anthropic.claude-sonnet-4-20250514-v1:0": {
"llm_key": "BEDROCK_ANTHROPIC_CLAUDE4_SONNET_INFERENCE_PROFILE",
"label": "Anthropic Claude 4 Sonnet",
},
}
else:
# TODO: apparently the list for OSS is to be much larger
return {
"Gemini 2.5": "GEMINI_2.5_PRO_PREVIEW",
"Gemini 2.5 Flash": "VERTEX_GEMINI_2.5_FLASH_PREVIEW_05_20",
"GPT 4.1": "OPENAI_GPT4_1",
"GPT o3-mini": "OPENAI_O3_MINI",
"bedrock/us.anthropic.claude-opus-4-20250514-v1:0": "BEDROCK_ANTHROPIC_CLAUDE4_OPUS_INFERENCE_PROFILE",
"bedrock/us.anthropic.claude-sonnet-4-20250514-v1:0": "BEDROCK_ANTHROPIC_CLAUDE4_SONNET_INFERENCE_PROFILE",
"gemini-2.5-pro-preview-05-06": {"llm_key": "VERTEX_GEMINI_2.5_PRO_PREVIEW", "label": "Gemini 2.5 Pro"},
"gemini-2.5-flash-preview-05-20": {
"llm_key": "VERTEX_GEMINI_2.5_FLASH_PREVIEW_05_20",
"label": "Gemini 2.5 Flash",
},
"azure/gpt-4.1": {"llm_key": "AZURE_OPENAI_GPT4_1", "label": "GPT 4.1"},
"azure/o3-mini": {"llm_key": "AZURE_OPENAI_O3_MINI", "label": "GPT O3 Mini"},
"us.anthropic.claude-opus-4-20250514-v1:0": {
"llm_key": "BEDROCK_ANTHROPIC_CLAUDE4_OPUS_INFERENCE_PROFILE",
"label": "Anthropic Claude 4 Opus",
},
"us.anthropic.claude-sonnet-4-20250514-v1:0": {
"llm_key": "BEDROCK_ANTHROPIC_CLAUDE4_SONNET_INFERENCE_PROFILE",
"label": "Anthropic Claude 4 Sonnet",
},
}
def is_cloud_environment(self) -> bool:

View File

@@ -867,12 +867,13 @@ class ForgeAgent:
else:
if engine in CUA_ENGINES:
self.async_operation_pool.run_operation(task.task_id, AgentPhase.llm)
json_response = await app.LLM_API_HANDLER(
prompt=extract_action_prompt,
prompt_name="extract-actions",
step=step,
screenshots=scraped_page.screenshots,
llm_key_override=llm_caller.llm_key if llm_caller else None,
llm_key_override=task.llm_key,
)
try:
json_response = await self.handle_potential_verification_code(

View File

@@ -810,9 +810,9 @@ async def models() -> ModelsResponse:
Get a list of available models.
"""
mapping = settings.get_model_name_to_llm_key()
models = list(mapping.keys())
just_labels = {k: v["label"] for k, v in mapping.items() if "anthropic" not in k.lower()}
return ModelsResponse(models=models)
return ModelsResponse(models=just_labels)
@legacy_base_router.post(

View File

@@ -57,10 +57,10 @@ class TaskV2(BaseModel):
"""
if self.model:
model_name = self.model.get("name")
model_name = self.model.get("model_name")
if model_name:
mapping = settings.get_model_name_to_llm_key()
llm_key = mapping.get(model_name)
llm_key = mapping.get(model_name, {}).get("llm_key")
if llm_key:
return llm_key

View File

@@ -248,10 +248,10 @@ class Task(TaskBase):
Otherwise return `None`.
"""
if self.model:
model_name = self.model.get("name")
model_name = self.model.get("model_name")
if model_name:
mapping = settings.get_model_name_to_llm_key()
return mapping.get(model_name)
return mapping.get(model_name, {}).get("llm_key")
return None
@@ -370,4 +370,4 @@ class SortDirection(StrEnum):
class ModelsResponse(BaseModel):
models: list[str]
models: dict[str, str]

View File

@@ -5,11 +5,10 @@ from typing import Any, List
from pydantic import BaseModel, field_validator
from typing_extensions import deprecated
from skyvern.config import settings
from skyvern.forge.sdk.schemas.files import FileInfo
from skyvern.forge.sdk.schemas.task_v2 import TaskV2
from skyvern.forge.sdk.workflow.exceptions import WorkflowDefinitionHasDuplicateBlockLabels
from skyvern.forge.sdk.workflow.models.block import Block, BlockTypeVar
from skyvern.forge.sdk.workflow.models.block import BlockTypeVar
from skyvern.forge.sdk.workflow.models.parameter import PARAMETER_TYPE
from skyvern.schemas.runs import ProxyLocation
from skyvern.utils.url_validators import validate_url
@@ -82,34 +81,6 @@ class Workflow(BaseModel):
modified_at: datetime
deleted_at: datetime | None = None
def determine_llm_key(self, *, block: Block | None = None) -> str | None:
"""
Determine the LLM key override to use for a block, if it has one.
It has one if:
- it defines one, or
- the workflow it is a part of (if applicable) defines one
"""
mapping = settings.get_model_name_to_llm_key()
if block:
model_name = (block.model or {}).get("name")
if model_name:
llm_key = mapping.get(model_name)
if llm_key:
return llm_key
workflow_model_name = (self.model or {}).get("name")
if workflow_model_name:
llm_key = mapping.get(workflow_model_name)
if llm_key:
return llm_key
return None
class WorkflowRunStatus(StrEnum):
created = "created"