Skip to content

Releases: BerriAI/litellm

v1.73.2.dev1

24 Jun 22:02
Compare
Choose a tag to compare

What's Changed

  • VertexAI Anthropic passthrough cost calc fixes + Filter litellm params from request sent to passthrough endpoint by @krrishdholakia in #11992
  • Fix custom pricing logging + Gemini - only use accepted format values + Gemini - cache tools if passing alongside cached content by @krrishdholakia in #11989
  • Fix unpack_defs handling of nested $ref inside anyOf items by @colesmcintosh in #11964
  • #response_format NVIDIA-NIM add response_format to OpenAI parameters … by @shagunb-acn in #12003
  • Add Azure o3-pro Pricing by @marty-sullivan in #11990
  • [Bug Fix] SCIM - Ensure new user roles are applied by @ishaan-jaff in #12015

Full Changelog: v1.73.1-nightly...v1.73.2.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.73.2.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 267.8382003869747 6.2096771619800935 0.0 1858 0 214.47131599995828 1466.6541370000346
Aggregated Passed ✅ 250.0 267.8382003869747 6.2096771619800935 0.0 1858 0 214.47131599995828 1466.6541370000346

v1.73.1-nightly

24 Jun 16:17
Compare
Choose a tag to compare

What's Changed

  • Fix SambaNova 'created' field validation error - handle float timestamps by @neubig in #11971
  • Docs - Add Recommended Machine Specifications by @ishaan-jaff in #11980
  • fix: make response api support Azure Authentication method by @hsuyuming in #11941
  • feat: add Last Success column to health check table by @colesmcintosh in #11903
  • Add GitHub Actions workflow for LLM translation testing artifacts by @colesmcintosh in #11780
  • Fix markdown table not rendering properly by @mukesh-dream11 in #11969
  • [Fix] - Check HTTP_PROXY vars in networking requests by @ishaan-jaff in #11947
  • Proxy UI MCP Auth passthrough by @wagnerjt in #11968
  • fix unrecognised parameter reasoning_effort by @Shankyg in #11838
  • Fixing watsonx error: 'model_id' or 'model' cannot be specified in the request body for models in a deployment space by @cbjuan in #11854
  • [Bug Fix] Perplexity - LiteLLM doesn't support 'web_search_options' for Perplexity' Sonar Pro model by @ishaan-jaff in #11983
  • feat: implement Perplexity citation tokens and search queries cost calculation by @colesmcintosh in #11938
  • [Feat] Enterprise - Allow dynamically disabling callbacks in request headers by @ishaan-jaff in #11985
  • Add Mistral 3.2 24B to model mapping by @colesmcintosh in #11926
  • [Feat] Add List Callbacks API Endpoint by @ishaan-jaff in #11987
  • fix: fix test_get_azure_ad_token_with_oidc_token testcase issue by @hsuyuming in #11988
  • [Bug Fix] Bedrock Guardrail - Don't raise exception on intervene action by @ishaan-jaff in #11875

New Contributors

Full Changelog: v1.73.0.rc.1...v1.73.1-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.73.1-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.73.1-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 269.80153099125215 6.123419901585826 0.0 1829 0 217.6905329999954 1336.1768169999948
Aggregated Passed ✅ 250.0 269.80153099125215 6.123419901585826 0.0 1829 0 217.6905329999954 1336.1768169999948

v1.73.0.rc.1

22 Jun 06:43
Compare
Choose a tag to compare

What's Changed

  • (Tutorial) Onboard Users for AI Exploration by @krrishdholakia in #11955
  • Management Fixes - don't apply default internal user settings to admins + preserve all model access for teams with empty model list, when team model added + /v2/model/info fixes by @krrishdholakia in #11957

Full Changelog: v1.73.0-nightly...v1.73.0.rc.1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.73.0.rc.1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 220.0 240.58426726332922 6.145106667198675 0.0 1838 0 196.0181700000021 1838.010895000025
Aggregated Passed ✅ 220.0 240.58426726332922 6.145106667198675 0.0 1838 0 196.0181700000021 1838.010895000025

v1.73.0-nightly

22 Jun 01:45
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.72.9-nightly...v1.73.0-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.73.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.73.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 230.0 252.55233764162196 6.182759830384375 0.0 1850 0 208.6453730000244 1743.1928639999796
Aggregated Passed ✅ 230.0 252.55233764162196 6.182759830384375 0.0 1850 0 208.6453730000244 1743.1928639999796

v1.72.9-nightly

21 Jun 04:53
Compare
Choose a tag to compare

What's Changed

  • [Feat] MCP - Allow connecting to MCP with authentication headers + Allow clients to specify MCP headers (#11890) by @ishaan-jaff in #11891
  • [Fix] Networking - allow using CA Bundles by @ishaan-jaff in #11906
  • [Feat] Add AWS Bedrock profiles for the APAC region by @lgruen-vcgs in #11883
  • bumps the anthropic package by @rinormaloku in #11851
  • Add deployment annotations by @InvisibleMan1306 in #11849
  • Enhance Mistral API: Add support for parallel tool calls by @njbrake in #11770
  • [UI] QA Items for adding pass through endpoints by @ishaan-jaff in #11909
  • build(model_prices_and_context_window.json): mark all gemini-2.5 models support pdf input + Set anthropic custom llm provider property by @krrishdholakia in #11907
  • fix(proxy_server.py): fix loading ui on custom root path by @krrishdholakia in #11912
  • LiteLLM SDK <-> Proxy improvement (don't transform message client-side) + Bedrock - handle qs:.. in base64 file data + Tag Management - support adding public model names by @krrishdholakia in #11908
  • Add success modal for health check responses by @colesmcintosh in #11899
  • Volcengine - thinking param support + Azure - handle more gpt custom naming patterns by @krrishdholakia in #11914
  • [Feat] Model Cost Map - Add gemini-2.5-pro and set gemini-2.5-pro supports_reasoning=True by @ishaan-jaff in #11927
  • [Feat] UI Allow testing /v1/messages on the Test Key Page by @ishaan-jaff in #11930
  • Feat/add delete callback by @jtong99 in #11654
  • add ciphers in command and pass to hypercorn for proxy by @frankzye in #11916
  • [Bug Fix] Fix model_group tracked for /v1/messages and /moderations by @ishaan-jaff in #11933
  • [Bug Fix] Cost tracking and logging via the /v1/messages API are not working when using Claude Code by @ishaan-jaff in #11928
  • [Feat] Add Azure Codex Models on LiteLLM + new /v1 preview Azure OpenAI API by @ishaan-jaff in #11934
  • [Feat] UI QA: Pass through endpoints by @ishaan-jaff in #11939

New Contributors

Full Changelog: v1.72.7-nightly...v1.72.9-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.72.9-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 240.0 259.91299670392254 6.2187422072270495 0.0 1861 0 210.9276310000041 1676.9406920000165
Aggregated Passed ✅ 240.0 259.91299670392254 6.2187422072270495 0.0 1861 0 210.9276310000041 1676.9406920000165

v1.72.7-nightly

20 Jun 02:47
ad2e230
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.72.6.dev1...v1.72.7-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.72.7-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 190.0 211.70463782970805 6.29948833219958 0.0 1885 0 169.2135669999857 2108.276391000004
Aggregated Passed ✅ 190.0 211.70463782970805 6.29948833219958 0.0 1885 0 169.2135669999857 2108.276391000004

v1.72.6-stable

19 Jun 20:08
56aaaf7
Compare
Choose a tag to compare

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.72.6-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 269.27781931947453 6.111834388077504 0.0 1828 0 215.86210600003142 1630.9297619999938
Aggregated Passed ✅ 250.0 269.27781931947453 6.111834388077504 0.0 1828 0 215.86210600003142 1630.9297619999938

What's Changed

Read more

v1.72.6.post1-nightly

18 Jun 04:41
Compare
Choose a tag to compare

Full Changelog: v1.72.6.dev1...v1.72.6.post1-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.72.6.post1-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 190.0 209.8013988365269 6.275681933110413 0.0 1878 0 167.48262099997646 1487.4784890000115
Aggregated Passed ✅ 190.0 209.8013988365269 6.275681933110413 0.0 1878 0 167.48262099997646 1487.4784890000115

v1.72.6.devSCIM

18 Jun 17:44
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.72.6.dev1...v1.72.6.devSCIM

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.72.6.devSCIM

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 190.0 215.1720637640139 6.262237464870193 0.0 1873 0 171.28891599998042 1800.7898239999918
Aggregated Passed ✅ 190.0 215.1720637640139 6.262237464870193 0.0 1873 0 171.28891599998042 1800.7898239999918

v1.72.6.SCIM2

18 Jun 21:19
Compare
Choose a tag to compare

Full Changelog: v1.72.6.devSCIM...v1.72.6.SCIM2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.72.6.SCIM2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 190.0 213.45712869978374 6.190773809263607 0.0 1852 0 171.36217200004467 1296.009626
Aggregated Passed ✅ 190.0 213.45712869978374 6.190773809263607 0.0 1852 0 171.36217200004467 1296.009626