From 20a611a9350f04bb96f4ede2dfaf76a7fec9e323 Mon Sep 17 00:00:00 2001 From: Kerem Yilmaz Date: Wed, 24 Jul 2024 23:47:02 -0700 Subject: [PATCH] Use perf-counter instead of time (#643) --- skyvern/forge/sdk/api/llm/api_handler_factory.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/skyvern/forge/sdk/api/llm/api_handler_factory.py b/skyvern/forge/sdk/api/llm/api_handler_factory.py index 21b9db55..7b23deb1 100644 --- a/skyvern/forge/sdk/api/llm/api_handler_factory.py +++ b/skyvern/forge/sdk/api/llm/api_handler_factory.py @@ -195,7 +195,7 @@ class LLMAPIHandlerFactory: # TODO (kerem): add a retry mechanism to this call (acompletion_with_retries) # TODO (kerem): use litellm fallbacks? https://litellm.vercel.app/docs/tutorials/fallbacks#how-does-completion_with_fallbacks-work LOG.info("Calling LLM API", llm_key=llm_key, model=llm_config.model_name) - t_llm_request = time.time() + t_llm_request = time.perf_counter() response = await litellm.acompletion( model=llm_config.model_name, messages=messages, @@ -206,7 +206,7 @@ class LLMAPIHandlerFactory: except openai.OpenAIError as e: raise LLMProviderError(llm_key) from e except CancelledError: - t_llm_cancelled = time.time() + t_llm_cancelled = time.perf_counter() LOG.error( "LLM request got cancelled", llm_key=llm_key,