diff --git a/backend/onyx/llm/utils.py b/backend/onyx/llm/utils.py index 81bb1146fc..5b0c805d84 100644 --- a/backend/onyx/llm/utils.py +++ b/backend/onyx/llm/utils.py @@ -498,15 +498,21 @@ def get_llm_contextual_cost( num_input_tokens += num_tokens + num_docs * DOCUMENT_SUMMARY_TOKEN_ESTIMATE num_output_tokens += num_docs * MAX_CONTEXT_TOKENS - usd_per_prompt, usd_per_completion = litellm.cost_per_token( - model=llm.config.model_name, - prompt_tokens=num_input_tokens, - completion_tokens=num_output_tokens, - ) + try: + usd_per_prompt, usd_per_completion = litellm.cost_per_token( + model=llm.config.model_name, + prompt_tokens=num_input_tokens, + completion_tokens=num_output_tokens, + ) + except Exception: + logger.exception( + f"An unexpected error occurred while calculating cost for model {llm.config.model_name} (potentially due to malformed name). Assuming cost is 0." + ) + return 0 + # Costs are in USD dollars per million tokens return usd_per_prompt + usd_per_completion - def get_llm_max_tokens( model_map: dict, model_name: str,