From 5d7d668252fc88414f453d772c1740006534c5e4 Mon Sep 17 00:00:00 2001 From: pedrohsdb Date: Fri, 24 Oct 2025 16:45:58 -0700 Subject: [PATCH] point flash and flash lite to stable (#3816) --- skyvern/config.py | 8 ++++---- skyvern/forge/sdk/api/llm/config_registry.py | 8 ++++---- 2 files changed, 8 insertions(+), 8 deletions(-) diff --git a/skyvern/config.py b/skyvern/config.py index 65e3d2fc..ca401819 100644 --- a/skyvern/config.py +++ b/skyvern/config.py @@ -395,11 +395,11 @@ class Settings(BaseSettings): if self.is_cloud_environment(): return { "gemini-2.5-pro-preview-05-06": {"llm_key": "VERTEX_GEMINI_2.5_PRO", "label": "Gemini 2.5 Pro"}, - "gemini-2.5-flash-preview-09-2025": { + "gemini-2.5-flash": { "llm_key": "VERTEX_GEMINI_2.5_FLASH", "label": "Gemini 2.5 Flash", }, - "gemini-2.5-flash-lite-preview-09-2025": { + "gemini-2.5-flash-lite": { "llm_key": "VERTEX_GEMINI_2.5_FLASH_LITE", "label": "Gemini 2.5 Flash Lite", }, @@ -431,11 +431,11 @@ class Settings(BaseSettings): # TODO: apparently the list for OSS is to be much larger return { "gemini-2.5-pro-preview-05-06": {"llm_key": "VERTEX_GEMINI_2.5_PRO", "label": "Gemini 2.5 Pro"}, - "gemini-2.5-flash-preview-09-2025": { + "gemini-2.5-flash": { "llm_key": "VERTEX_GEMINI_2.5_FLASH", "label": "Gemini 2.5 Flash", }, - "gemini-2.5-flash-lite-preview-09-2025": { + "gemini-2.5-flash-lite": { "llm_key": "VERTEX_GEMINI_2.5_FLASH_LITE", "label": "Gemini 2.5 Flash Lite", }, diff --git a/skyvern/forge/sdk/api/llm/config_registry.py b/skyvern/forge/sdk/api/llm/config_registry.py index 20a126d2..b2ff73e9 100644 --- a/skyvern/forge/sdk/api/llm/config_registry.py +++ b/skyvern/forge/sdk/api/llm/config_registry.py @@ -1190,14 +1190,14 @@ if settings.ENABLE_VERTEX_AI and settings.VERTEX_CREDENTIALS: LLMConfigRegistry.register_config( "VERTEX_GEMINI_2.5_FLASH", LLMConfig( - "vertex_ai/gemini-2.5-flash-preview-09-2025", + "vertex_ai/gemini-2.5-flash", ["VERTEX_CREDENTIALS"], supports_vision=True, add_assistant_prefix=False, max_completion_tokens=65535, litellm_params=LiteLLMParams( vertex_credentials=settings.VERTEX_CREDENTIALS, - api_base=f"{api_base}/gemini-2.5-flash-preview-09-2025" if api_base else None, + api_base=f"{api_base}/gemini-2.5-flash" if api_base else None, vertex_location=settings.VERTEX_LOCATION, thinking={ "budget_tokens": settings.GEMINI_THINKING_BUDGET, @@ -1209,14 +1209,14 @@ if settings.ENABLE_VERTEX_AI and settings.VERTEX_CREDENTIALS: LLMConfigRegistry.register_config( "VERTEX_GEMINI_2.5_FLASH_LITE", LLMConfig( - "vertex_ai/gemini-2.5-flash-lite-preview-09-2025", + "vertex_ai/gemini-2.5-flash-lite", ["VERTEX_CREDENTIALS"], supports_vision=True, add_assistant_prefix=False, max_completion_tokens=65535, litellm_params=LiteLLMParams( vertex_credentials=settings.VERTEX_CREDENTIALS, - api_base=f"{api_base}/gemini-2.5-flash-lite-preview-09-2025" if api_base else None, + api_base=f"{api_base}/gemini-2.5-flash-lite" if api_base else None, vertex_location=settings.VERTEX_LOCATION, thinking={ "budget_tokens": settings.GEMINI_THINKING_BUDGET,