Skip to content

Releases: BerriAI/litellm

v1.65.0.rc

28 Mar 22:58
0865e52
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.64.1-nightly...v1.65.0.rc

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.65.0.rc

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 540.0 624.3702232376279 5.474375767460597 0.0 1637 0 487.16235000006236 2776.3750889999983
Aggregated Failed ❌ 540.0 624.3702232376279 5.474375767460597 0.0 1637 0 487.16235000006236 2776.3750889999983

v1.65.0-nightly

28 Mar 22:08
0865e52
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.64.1-nightly...v1.65.0-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.65.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 530.0 578.5163737856723 5.6521867286945735 0.0 1689 0 483.77581299996564 4841.4478800000325
Aggregated Failed ❌ 530.0 578.5163737856723 5.6521867286945735 0.0 1689 0 483.77581299996564 4841.4478800000325

v1.64.1.dev1

28 Mar 07:34
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.64.1-nightly...v1.64.1.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.64.1.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.64.1.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.64.1.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 530.0 588.3618842036032 5.564625107213408 0.0 1665 0 469.31838699998707 4872.995692000018
Aggregated Failed ❌ 530.0 588.3618842036032 5.564625107213408 0.0 1665 0 469.31838699998707 4872.995692000018

v1.64.1-nightly

26 Mar 19:47
Compare
Choose a tag to compare

What's Changed

Full Changelog: 1.64.0.dev1...v1.64.1-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.64.1-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 530.0 583.0955007522234 5.641309914749418 0.0 1687 0 483.3096179999643 5048.277267999993
Aggregated Failed ❌ 530.0 583.0955007522234 5.641309914749418 0.0 1687 0 483.3096179999643 5048.277267999993

1.64.0.dev1

25 Mar 16:19
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.63.14-nightly...1.64.0.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-1.64.0.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 530.0 595.2345953515114 5.639441403576418 0.0 1687 0 488.5956009999859 4882.238139000037
Aggregated Failed ❌ 530.0 595.2345953515114 5.639441403576418 0.0 1687 0 488.5956009999859 4882.238139000037

v1.63.14-stable.patch1

25 Mar 07:17
Compare
Choose a tag to compare

Full Changelog: v1.63.14-stable...v1.63.14-stable.patch1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.63.14-stable.patch1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 240.0 255.95973818329688 6.163881077009807 0.0 1844 0 213.38151600002675 2969.717206000041
Aggregated Passed ✅ 240.0 255.95973818329688 6.163881077009807 0.0 1844 0 213.38151600002675 2969.717206000041

v1.63.14-stable

23 Mar 03:02
Compare
Choose a tag to compare

Full Changelog: v1.63.14-nightly...v1.63.14-stable

🚨 Issue identified on Bedrock streaming usage calculation - fixed in v1.63.14-stable.patch 1 - https://github.com/BerriAI/litellm/releases/tag/v1.63.14-stable.patch1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.63.14-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 190.0 207.58892887694674 6.219076098228175 0.0 1861 0 165.81997500003354 2278.4240120000163
Aggregated Passed ✅ 190.0 207.58892887694674 6.219076098228175 0.0 1861 0 165.81997500003354 2278.4240120000163

v1.63.14.rc

22 Mar 22:06
Compare
Choose a tag to compare

What's Changed

  • Modify completion handler for SageMaker to use payload from prepared_request by @andjsmi in #9326
  • Arize integration Fix by @nate-mar in #9338
  • Fix get_llm_provider error + add model name to tpm/rpm cache key (enables wildcard models to work w/ usage-based-routing) by @krrishdholakia in #9355
  • Litellm dev 03 18 2025 p1 v3 by @krrishdholakia in #9354
  • [Bug Fix] Arize AI Logging Integration with LiteLLM by @ishaan-jaff in #9352
  • build(model_prices_and_context_window.json): fix azure gpt-4o pricing by @krrishdholakia in #9361
  • Contributor PR: Fix sagemaker too little data for content error by @krrishdholakia in #9335
  • [Feat] - API - Allow using dynamic Arize AI Spaces on LiteLLM by @ishaan-jaff in #9353
  • fix(common_utils.py): handle cris only model by @krrishdholakia in #9363
  • docs(litellm_proxy): correct parameter assignment in litellm proxy docs by @colesmcintosh in #9375
  • Feature flag checking LiteLLM_CredentialsTable by @krrishdholakia in #9376
  • fix(lowest_tpm_rpm_v2.py): support batch writing increments to redis by @krrishdholakia in #9357
  • Support 'prisma migrate' for db schema changes by @krrishdholakia in #9382
  • Fix latency redis by @emerzon in #9387
  • Revert "Fix latency redis" by @krrishdholakia in #9388
  • build(model_prices_and_context_window.json): add o1-pro pricing by @krrishdholakia in #9397
  • [Bug Fix] - Azure OpenAI - ensure SSL verification runs by @ishaan-jaff in #9341
  • [Feat] - Allow building custom prompt management integration by @ishaan-jaff in #9384
  • Litellm fix icons by @azdolinski in #9374
  • [UI Improvement] Use local icons for model providers instead of downloading them by @ishaan-jaff in #9399
  • fix(internal_user_endpoints.py): re-introduce upsert on user not found by @krrishdholakia in #9395
  • docs: Include Phoenix Page into sidebar under observability by @SrilakshmiC in #9332
  • fix(types/utils.py): support openai 'file' message type by @krrishdholakia in #9402
  • feat: Add support for custom OPENROUTER_API_BASE via get_secret in co… by @graysonchen in #9369
  • fix: VertexAI outputDimensionality configuration by @JamesGuthrie in #9366
  • docs(vertexai): fix typo in required env variables by @Shiti in #9325
  • Update perplexity.md by @jollyolliel in #9290
  • fix: VoyageAI prompt_token always empty by @lucasra1 in #9260
  • build(deps): bump litellm from 1.55.3 to 1.61.15 in /cookbook/litellm-ollama-docker-image by @dependabot in #9422
  • [Feat] OpenAI o1-pro Responses API streaming support by @ishaan-jaff in #9419
  • [Feat] Add OpenAI o1-pro support on Responses API by @ishaan-jaff in #9415
  • [Docs - Draft] LiteLLM x MCP Interface by @ishaan-jaff in #9435
  • support returning api-base on pass-through endpoints + consistently return 404 if team not found in DB by @krrishdholakia in #9439
  • fix(handle_error.py): make cooldown error more descriptive by @krrishdholakia in #9438
  • Consistent anthropic response_format streaming/non-streaming behaviour by @krrishdholakia in #9437
  • New Azure Models (GPT-4.5-Preview, Mistral Small 3.1) by @emerzon in #9453
  • Set max size limit to in-memory cache item - prevents OOM errors by @krrishdholakia in #9448
  • fix(model_param_helper.py): update _get_litellm_supported_transcription_kwargs() to use proper annotations from TranscriptionCreateParamsNonStreaming & ``TranscriptionCreateParamsStreaming` by @hsaeed3 in #9451
  • [Feat] LiteLLM x MCP Bridge - Use MCP Tools with LiteLLM by @ishaan-jaff in #9436
  • fix(model_param_helper.py): update _get_litellm_supported_transcription_kwargs() to use proper annotations from TranscriptionCreateParamsNonStreaming & `TranscriptionCreateParamsStreaming by @ishaan-jaff in #9455

New Contributors

Full Changelog: v1.63.12-nightly...v1.63.14.rc

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.14.rc

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 230.0 249.89683940656786 6.2068378536570386 0.0 1857 0 196.31636200006142 3500.2345190000597
Aggregated Passed ✅ 230.0 249.89683940656786 6.2068378536570386 0.0 1857 0 196.31636200006142 3500.2345190000597

v1.63.14-nightly

22 Mar 06:35
Compare
Choose a tag to compare

What's Changed

  • Modify completion handler for SageMaker to use payload from prepared_request by @andjsmi in #9326
  • Arize integration Fix by @nate-mar in #9338
  • Fix get_llm_provider error + add model name to tpm/rpm cache key (enables wildcard models to work w/ usage-based-routing) by @krrishdholakia in #9355
  • Litellm dev 03 18 2025 p1 v3 by @krrishdholakia in #9354
  • [Bug Fix] Arize AI Logging Integration with LiteLLM by @ishaan-jaff in #9352
  • build(model_prices_and_context_window.json): fix azure gpt-4o pricing by @krrishdholakia in #9361
  • Contributor PR: Fix sagemaker too little data for content error by @krrishdholakia in #9335
  • [Feat] - API - Allow using dynamic Arize AI Spaces on LiteLLM by @ishaan-jaff in #9353
  • fix(common_utils.py): handle cris only model by @krrishdholakia in #9363
  • docs(litellm_proxy): correct parameter assignment in litellm proxy docs by @colesmcintosh in #9375
  • Feature flag checking LiteLLM_CredentialsTable by @krrishdholakia in #9376
  • fix(lowest_tpm_rpm_v2.py): support batch writing increments to redis by @krrishdholakia in #9357
  • Support 'prisma migrate' for db schema changes by @krrishdholakia in #9382
  • Fix latency redis by @emerzon in #9387
  • Revert "Fix latency redis" by @krrishdholakia in #9388
  • build(model_prices_and_context_window.json): add o1-pro pricing by @krrishdholakia in #9397
  • [Bug Fix] - Azure OpenAI - ensure SSL verification runs by @ishaan-jaff in #9341
  • [Feat] - Allow building custom prompt management integration by @ishaan-jaff in #9384
  • Litellm fix icons by @azdolinski in #9374
  • [UI Improvement] Use local icons for model providers instead of downloading them by @ishaan-jaff in #9399
  • fix(internal_user_endpoints.py): re-introduce upsert on user not found by @krrishdholakia in #9395
  • docs: Include Phoenix Page into sidebar under observability by @SrilakshmiC in #9332
  • fix(types/utils.py): support openai 'file' message type by @krrishdholakia in #9402
  • feat: Add support for custom OPENROUTER_API_BASE via get_secret in co… by @graysonchen in #9369
  • fix: VertexAI outputDimensionality configuration by @JamesGuthrie in #9366
  • docs(vertexai): fix typo in required env variables by @Shiti in #9325
  • Update perplexity.md by @jollyolliel in #9290
  • fix: VoyageAI prompt_token always empty by @lucasra1 in #9260
  • build(deps): bump litellm from 1.55.3 to 1.61.15 in /cookbook/litellm-ollama-docker-image by @dependabot in #9422
  • [Feat] OpenAI o1-pro Responses API streaming support by @ishaan-jaff in #9419
  • [Feat] Add OpenAI o1-pro support on Responses API by @ishaan-jaff in #9415
  • [Docs - Draft] LiteLLM x MCP Interface by @ishaan-jaff in #9435
  • support returning api-base on pass-through endpoints + consistently return 404 if team not found in DB by @krrishdholakia in #9439
  • fix(handle_error.py): make cooldown error more descriptive by @krrishdholakia in #9438
  • Consistent anthropic response_format streaming/non-streaming behaviour by @krrishdholakia in #9437
  • New Azure Models (GPT-4.5-Preview, Mistral Small 3.1) by @emerzon in #9453
  • Set max size limit to in-memory cache item - prevents OOM errors by @krrishdholakia in #9448
  • fix(model_param_helper.py): update _get_litellm_supported_transcription_kwargs() to use proper annotations from TranscriptionCreateParamsNonStreaming & ``TranscriptionCreateParamsStreaming` by @hsaeed3 in #9451
  • [Feat] LiteLLM x MCP Bridge - Use MCP Tools with LiteLLM by @ishaan-jaff in #9436
  • fix(model_param_helper.py): update _get_litellm_supported_transcription_kwargs() to use proper annotations from TranscriptionCreateParamsNonStreaming & `TranscriptionCreateParamsStreaming by @ishaan-jaff in #9455

New Contributors

Full Changelog: v1.63.12-nightly...v1.63.14-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.14-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 210.0 225.80703369680768 6.282456357875219 0.0 1880 0 182.8801230000181 2803.2175499999994
Aggregated Passed ✅ 210.0 225.80703369680768 6.282456357875219 0.0 1880 0 182.8801230000181 2803.2175499999994

v1.63.12-nightly

19 Mar 16:20
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.63.11-nightly...v1.63.12-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.12-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 240.0 266.99215766025395 6.081096542088128 0.0 1819 0 211.12568599994574 4206.960361000029
Aggregated Passed ✅ 240.0 266.99215766025395 6.081096542088128 0.0 1819 0 211.12568599994574 4206.960361000029