Skip to content

Releases: BerriAI/litellm

v1.74.9.rc.1

29 Jul 01:37
Compare
Choose a tag to compare

What's Changed

  • Litellm release notes 07 27 2025 p1 by @krrishdholakia in #13027
  • VertexAI - camelcase optional params for image generation + Anthropic - streaming, always ensure assistant role set on only first chunk by @krrishdholakia in #12889
  • Bulk User Edit - additional improvements - edit all users + set 'no-default-models' on all users by @krrishdholakia in #12925
  • add X-Initiator header for GitHub Copilot to reduce premium requests by @ckoehler in #13016
  • docs - openwebui show how to include reasoning content for gemini models by @ishaan-jaff in #13060

New Contributors

Full Changelog: v1.74.9.rc-draft...v1.74.9.rc.1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.9.rc.1

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.9.rc.1

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed βœ… 130.0 168.0418430074381 6.293403780045943 0.0 1882 0 97.96855799999094 1049.3644050000057
Aggregated Passed βœ… 130.0 168.0418430074381 6.293403780045943 0.0 1882 0 97.96855799999094 1049.3644050000057

v1.74.9-stable

01 Aug 21:40
Compare
Choose a tag to compare

What's Changed

  • Litellm release notes 07 27 2025 p1 by @krrishdholakia in #13027
  • VertexAI - camelcase optional params for image generation + Anthropic - streaming, always ensure assistant role set on only first chunk by @krrishdholakia in #12889
  • Bulk User Edit - additional improvements - edit all users + set 'no-default-models' on all users by @krrishdholakia in #12925
  • add X-Initiator header for GitHub Copilot to reduce premium requests by @ckoehler in #13016
  • docs - openwebui show how to include reasoning content for gemini models by @ishaan-jaff in #13060

New Contributors

Full Changelog: v1.74.9.rc-draft...v1.74.9-stable

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.9-stable

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed βœ… 110.0 151.02663461606545 6.447232352830848 0.0 1930 0 81.53728299998875 1408.0881720000207
Aggregated Passed βœ… 110.0 151.02663461606545 6.447232352830848 0.0 1930 0 81.53728299998875 1408.0881720000207

litellm_v1.65.4-dev_fix

28 Jul 23:32
Compare
Choose a tag to compare

Full Changelog: v1.65.4-stable...litellm_v1.65.4-dev_fix

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.65.4-dev_fix

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed βœ… 150.0 182.3086370464627 6.335489722473191 0.0 1894 0 112.72804899999755 1087.5013490000072
Aggregated Passed βœ… 150.0 182.3086370464627 6.335489722473191 0.0 1894 0 112.72804899999755 1087.5013490000072

v1.74.9.rc-draft

26 Jul 18:46
b921e22
Compare
Choose a tag to compare
v1.74.9.rc-draft Pre-release
Pre-release

What's Changed

New Contributors

  • @magicalne made their first contribution in h...
Read more

v1.74.7-stable.patch.2

25 Jul 17:25
Compare
Choose a tag to compare

Full Changelog: v1.74.7-stable.patch.1...v1.74.7-stable.patch.2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.7-stable.patch.2

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.7-stable.patch.2

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.7-stable.patch.2

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed βœ… 120.0 157.13796241417324 6.362622204908533 0.0 1905 0 82.22530700004427 1211.5445880000095
Aggregated Passed βœ… 120.0 157.13796241417324 6.362622204908533 0.0 1905 0 82.22530700004427 1211.5445880000095

v1.74.7-stable.patch.1

25 Jul 15:13
Compare
Choose a tag to compare

Full Changelog: v1.74.7-stable...v1.74.7-stable.patch.1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.7-stable.patch.1

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed βœ… 170.0 210.81530860833013 6.340641152265373 0.0 1897 0 139.2534420000402 1204.8251080000227
Aggregated Passed βœ… 170.0 210.81530860833013 6.340641152265373 0.0 1897 0 139.2534420000402 1204.8251080000227

v1.74.7-stable

25 Jul 02:32
Compare
Choose a tag to compare

Full Changelog: v1.74.7.rc.1...v1.74.7-stable

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.7-stable

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed βœ… 120.0 155.293356068966 6.395530747523885 0.0 1914 0 79.63155700002744 1028.7689699999873
Aggregated Passed βœ… 120.0 155.293356068966 6.395530747523885 0.0 1914 0 79.63155700002744 1028.7689699999873

v1.74.8-nightly

24 Jul 00:32
Compare
Choose a tag to compare

What's Changed

  • Fix SSO Logout | Create Unified Login Page with SSO and Username/Password Options by @NANDINI-star in #12703
  • docs(moonshot): correct base url and document CN-specific endpoint by @magicalne in #12804
  • Adding HolmesGPT to projects using LiteLLM by @pavangudiwada in #12798
  • feat: Add Lambda AI provider support by @colesmcintosh in #12817
  • [UI Bug Fix] Show correct guardrails when editing a team by @ishaan-jaff in #12823
  • [Docs] Show correct list of vertex ai mistral models by @ishaan-jaff in #12828
  • [LLM Translation - GH Copilot] added dynamic endpoint support by @jugaldb in #12827
  • feat: add Morph provider support by @colesmcintosh in #12821
  • fix: remove deprecated groq/qwen-qwq-32b and add qwen/qwen3-32b by @aholmberg in #12831
  • Add Google Cloud Model Armor guardrail documentation by @colesmcintosh in #12814
  • [Feat] Add Recraft Image Generation API Support - New LLM Provider by @ishaan-jaff in #12832
  • [Bug Fix] - gemini leaking FD for sync calls with litellm.completion by @ishaan-jaff in #12824
  • [Feat] Add fireworks - fireworks/models/kimi-k2-instruct by @ishaan-jaff in #12837
  • [LLM Translation] add qwen-vl-plus by @jugaldb in #12829
  • [Azure OpenAI Feature] - Support DefaultAzureCredential without hard-coded environment variables by @ishaan-jaff in #12841
  • Docs - litellm benchmarks by @ishaan-jaff in #12842
  • fix(watsonx): IBM Watsonx - use correct parameter name for tool choice by @Tomas2D in #9980
  • feat: add Hyperbolic provider support by @colesmcintosh in #12826
  • Litellm batch cost tracking debug by @krrishdholakia in #12782
  • Passthrough Auth - make Auth checks OSS + Anthropic - only show 'reasoning_effort' for supported models by @krrishdholakia in #12847
  • build(deps): bump form-data from 4.0.0 to 4.0.4 in /ui/litellm-dashboard by @dependabot[bot] in #12851
  • Fix team_member_budget update logic by @krrishdholakia in #12843
  • Fix async retryer on .acompletion() + forward clientside headers - filter out content-type from clientside request (causes llm api call to hang) by @krrishdholakia in #12848
  • Openrouter - filter out cache_control flag for non-anthropic models (allows usage with claude code) by @krrishdholakia in #12850
  • Improvements on the Regenerate Key Flow by @NANDINI-star in #12788
  • Replace non-root Dockerfile base with Alpine multi-stage build; by @mdiloreto in #12707
  • build(deps): bump form-data from 4.0.3 to 4.0.4 in /docs/my-website by @dependabot[bot] in #12867
  • Bug fix - Azure KeyVault not in image, add azure-keyvault==4.2.0 to Docker img by @ishaan-jaff in #12873
  • [Separate Health App] Pass through cmd args via supervisord by @jugaldb in #12871
  • [Feat] Add Recraft API - Image Edits Support by @ishaan-jaff in #12874
  • [LLM Translation] Litellm gemini 2.0 live support by @jugaldb in #12839
  • [Feat] Add cost tracking for new vertex_ai/llama-3 API models by @ishaan-jaff in #12878
  • [Feat] - Track cost + add tags for health checks done by LiteLLM Proxy by @ishaan-jaff in #12880
  • Fix async retryer on .acompletion() + forward clientside headers - filter out content-type from clientside request (causes llm api call to hang) by @krrishdholakia in #12886
  • Request Headers - support x-litellm-num-retries + Usage - support usage by model group by @krrishdholakia in #12890
  • [Docs ]Litellm mcp access group doc by @jugaldb in #12883
  • [Add health check] add architecture diagram by @jugaldb in #12879
  • rm retired anthropic models from model_prices_and_context_window.json by @staeiou in #12864
  • feat: extended /v1/models endpoint, now it returns with fallbacks on demand by @murad-khafizov in #12811
  • feat: Add Pillar Security guardrail integration by @eagle-p in #12791
  • Swagger - bump DOMPurify version + add back Support local swagger bundle (good for airgapped environments) + support local swagger on custom root paths by @krrishdholakia in #12911
  • [Feat] LiteLLM x Cloudzero integration - Allow exporting spend to cloudzero by @ishaan-jaff in #12908

New Contributors

Full Changelog: v1.74.7.rc.1...v1.74.8-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.8-nightly

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed βœ… 170.0 208.0958998087373 6.271169991247407 0.0 1877 0 130.56155500004252 1168.738483000027
Aggregated Passed βœ… 170.0 208.0958998087373 6.271169991247407 0.0 1877 0 130.56155500004252 1168.738483000027

v1.74.3-stable.patch.4

23 Jul 03:17
Compare
Choose a tag to compare

Full Changelog: v1.74.3-stable.patch.3...v1.74.3-stable.patch.4

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.3-stable.patch.4

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed βœ… 160.0 204.61626603598248 6.22322538904882 0.0 1862 0 127.11543199998232 1627.7963420000106
Aggregated Passed βœ… 160.0 204.61626603598248 6.22322538904882 0.0 1862 0 127.11543199998232 1627.7963420000106

v1.72.2-stable.debug

22 Jul 06:05
Compare
Choose a tag to compare

Full Changelog: v1.72.2-stable...v1.72.2-stable.debug

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.72.2-stable.debug

Don't want to maintain your internal proxy? get in touch πŸŽ‰

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed βœ… 130.0 168.6584858474577 6.307054920240884 0.0 1888 0 95.40891199998214 1772.0808789999865
Aggregated Passed βœ… 130.0 168.6584858474577 6.307054920240884 0.0 1888 0 95.40891199998214 1772.0808789999865