catch litellm no deployment error (token ratelimit / ValueError) and raise LLMProviderErrorRetryableTask (#758)
This commit is contained in:
@@ -108,6 +108,13 @@ class LLMAPIHandlerFactory:
|
|||||||
LOG.info("LLM API call successful", llm_key=llm_key, model=llm_config.model_name)
|
LOG.info("LLM API call successful", llm_key=llm_key, model=llm_config.model_name)
|
||||||
except litellm.exceptions.APIError as e:
|
except litellm.exceptions.APIError as e:
|
||||||
raise LLMProviderErrorRetryableTask(llm_key) from e
|
raise LLMProviderErrorRetryableTask(llm_key) from e
|
||||||
|
except ValueError as e:
|
||||||
|
LOG.exception(
|
||||||
|
"LLM token limit exceeded",
|
||||||
|
llm_key=llm_key,
|
||||||
|
model=main_model_group,
|
||||||
|
)
|
||||||
|
raise LLMProviderErrorRetryableTask(llm_key) from e
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
LOG.exception(
|
LOG.exception(
|
||||||
"LLM request failed unexpectedly",
|
"LLM request failed unexpectedly",
|
||||||
|
|||||||
Reference in New Issue
Block a user