Releases: BerriAI/litellm
v1.74.9-stable.patch.1
Full Changelog: v1.74.9-stable...v1.74.9-stable.patch.1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.9-stable.patch.1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 160.0 | 195.02149162499995 | 6.2583306558386465 | 0.0 | 1872 | 0 | 124.22932999999148 | 1133.3790610000278 |
Aggregated | Passed ✅ | 160.0 | 195.02149162499995 | 6.2583306558386465 | 0.0 | 1872 | 0 | 124.22932999999148 | 1133.3790610000278 |
v1.74.15-nightly
What's Changed
- Add advanced date picker to all the tabs on the usage page by @NANDINI-star in #13221
- Add Perplexity citation annotations support by @kankute-sameer in #13225
- fix: role chaining and session name with webauthentication for aws bedrock by @RichardoC in #13205
- Fix missing extra_headers support for vLLM/openai_like embeddings by @colesmcintosh in #13198
- litellm/proxy: preserve model order of /v1/models and /model_group/info by @AlexanderYastrebov in #13178
- Prompt Management - abstract prompt templates away from model list (enables permission management on prompt templates) by @krrishdholakia in #13219
- [QA Fixes for MCP] - Ensure MCPs load + don't run a health check everytime we load MCPs on UI by @ishaan-jaff in #13228
- Revert "fix: role chaining and session name with webauthentication for aws bedrock" by @krrishdholakia in #13230
- Prompt Management (2/2) - New
/prompt/list
endpoint + key-based access to prompt templates by @krrishdholakia in #13218 - [UI QA Fixes] Stable release by @ishaan-jaff in #13231
New Contributors
- @kankute-sameer made their first contribution in #13225
- @AlexanderYastrebov made their first contribution in #13178
Full Changelog: v1.74.14-nightly...v1.74.15-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.15-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 160.0 | 201.90947012894696 | 6.350644454200707 | 0.0 | 1900 | 0 | 128.45073899995896 | 1495.8115740000153 |
Aggregated | Passed ✅ | 160.0 | 201.90947012894696 | 6.350644454200707 | 0.0 | 1900 | 0 | 128.45073899995896 | 1495.8115740000153 |
1.74.15.rc.1
What's Changed
- Add advanced date picker to all the tabs on the usage page by @NANDINI-star in #13221
- Add Perplexity citation annotations support by @kankute-sameer in #13225
- fix: role chaining and session name with webauthentication for aws bedrock by @RichardoC in #13205
- Fix missing extra_headers support for vLLM/openai_like embeddings by @colesmcintosh in #13198
- litellm/proxy: preserve model order of /v1/models and /model_group/info by @AlexanderYastrebov in #13178
- Prompt Management - abstract prompt templates away from model list (enables permission management on prompt templates) by @krrishdholakia in #13219
- [QA Fixes for MCP] - Ensure MCPs load + don't run a health check everytime we load MCPs on UI by @ishaan-jaff in #13228
- Revert "fix: role chaining and session name with webauthentication for aws bedrock" by @krrishdholakia in #13230
- Prompt Management (2/2) - New
/prompt/list
endpoint + key-based access to prompt templates by @krrishdholakia in #13218 - [UI QA Fixes] Stable release by @ishaan-jaff in #13231
New Contributors
- @kankute-sameer made their first contribution in #13225
- @AlexanderYastrebov made their first contribution in #13178
Full Changelog: v1.74.14-nightly...1.74.15.rc.1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-1.74.15.rc.1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 160.0 | 194.82821616116973 | 6.285152111114945 | 0.0 | 1880 | 0 | 124.01886500003911 | 1206.761389999997 |
Aggregated | Passed ✅ | 160.0 | 194.82821616116973 | 6.285152111114945 | 0.0 | 1880 | 0 | 124.01886500003911 | 1206.761389999997 |
v1.74.14.dev1
Full Changelog: v1.74.14-nightly...v1.74.14.dev1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.14.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.14.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 160.0 | 199.19181043322646 | 6.258277007293016 | 0.0 | 1872 | 0 | 130.81277300000238 | 1170.560404000014 |
Aggregated | Passed ✅ | 160.0 | 199.19181043322646 | 6.258277007293016 | 0.0 | 1872 | 0 | 130.81277300000238 | 1170.560404000014 |
v1.74.14-nightly
What's Changed
- [Bug Fix] Pass through logging handler VertexAI - ensure multimodal embedding responses are logged by @ishaan-jaff in #13050
- Remove duplicate test case verifying field filtering logic by @emmanuel-ferdman in #13023
- Properly parse json options for key generation in the UI by @5731la in #12989
- fix: correct CompletionRequest messages type to match OpenAI API spec by @restato in #12980
- chore: Improve docs for cost tracking by @RichardoC in #12976
- feat(langfuse-otel): Add comprehensive metadata support to Langfuse OpenTelemetry integration by @strickvl in #12956
- [MCP Gateway] MCP tools fix scrolling issue by @jugaldb in #13015
- [MCP gateway] add url namespacing docs by @jugaldb in #13063
- [Bug Fix] The model gemini-2.5-flash with the merge_reasoning_content_in_choices parameter does not work by @ishaan-jaff in #13066
- [Feat] Add Google AI Studio Imagen4 model family by @ishaan-jaff in #13065
- Default Usage Chart Date Range: Last 7 Days by @NANDINI-star in #12917
- Added handling for pwd protected cert files in AOAI CertificateCreden… by @Ne0-1 in #12995
- docs: add Qwen Code CLI tutorial by @colesmcintosh in #12915
- Azure
api_version="preview"
support + Bedrock cost tracking via Anthropic/v1/messages
by @krrishdholakia in #13072 - Remove extraneous
s
in docs by @maxrabin in #13079 - Fix list team v2 security check by @tofarr in #13094
- [MCP gateway] add pre and during call hooks init by @jugaldb in #13067
- fix: improve MCP server URL validation to support internal/Kubernetes URLs by @colesmcintosh in #13099
- [Bug Fix] Gemini-CLI - The Gemini Custom API request has an incorrect authorization format by @ishaan-jaff in #13098
- set default value for mcp namespace tool name in spend table to prevent duplicate entry in table by @lvuna in #12894
- [Feat] Allow using query_params for setting API Key for generateContent routes by @ishaan-jaff in #13100
- [LLM translation] add openrouter grok4 by @jugaldb in #13018
- [Infra] Looses MCP python version restrictions by @ishaan-jaff in #13102
- fix: always use choice index=0 for Anthropic streaming responses by @Maximgitman in #12666
- BUGFIX: Jitter should be added not multiplied (#12877) by @pathikrit in #12901
- fix: helm migration job not running schema update by @huetterma in #12809
- [Bug]: Set user from token user_id for OpenMeter integration by @betterthanbreakfast in #13029
- Revert "[Bug]: Set user from token user_id for OpenMeter integration" by @krrishdholakia in #13107
- Fix fallback delete by @phosae in #12606
- Fix/gemini api key environment variable support by @sahusiddharth in #12507
- [MCP Gateway] Add protocol headers by @jugaldb in #13062
- Fix token counter to ignore unsupported keys like prefix (#11791) by @Amit-kr26 in #11954
- Custom Auth - bubble up custom exceptions by @krrishdholakia in #13093
- feat: Add dot notation support for all JWT fields by @colesmcintosh in #13013
- [Feat] MLFlow Logging - Allow adding tags for ML Flow logging requests by @ishaan-jaff in #13108
- [LLM translation] Add support for bedrock computer use by @jugaldb in #12948
- After selecting date range show loader on usage cost charts by @NANDINI-star in #13113
- Revert "[LLM translation] Add support for bedrock computer use" by @krrishdholakia in #13118
- fix(model_checks.py): handle custom values in wildcard model name (e.g. genai/test/*) by @krrishdholakia in #13116
- move to use_prisma_migrate by default + resolve team-only models on auth checks + UI - add sagemaker on UI by @krrishdholakia in #13117
- fix tool aws bedrock call index when the function only have optional arg by @jnhyperion in #13115
- [MCP Gateway] add health check endpoints for MCP by @jugaldb in #13106
- [MCP Protocol header] fix issue with clients protocol header by @jugaldb in #13112
- Added Voyage, Jinai, Deepinfra and VolcEngine providers on the UI by @NANDINI-star in #13131
- [Proxy UI] fix object permission for orgs by @jugaldb in #13142
- New Advanced Date Range Picker Component by @NANDINI-star in #13141
- [Feat] UI + Backend add a tab for use agent activity by @ishaan-jaff in #13146
- [LLM translation] Fix bedrock computer use by @jugaldb in #13143
- [MCP Guardrails] move pre and during hooks to ProxyLoggin by @jugaldb in #13109
- [Feat] v2 updates - tracking DAU, WAU, MAU for coding tool usage + show Daily Usage per User by @ishaan-jaff in #13147
- Litellm explore postgres db ci cd by @krrishdholakia in #13156
- [MCP Gateway] Litellm mcp client list fail by @jugaldb in #13114
- [MCP Gateway] fix migrations by @jugaldb in #13157
- build(config.yml): migrate build_and_test to ci/cd pg db by @krrishdholakia in #13166
- add framework name to UserAgent header in AWS Bedrock API call by @0x-fang in #13159
- fix: remove obsolete attribute
version
in docker compose by @kowyo in #13172 - [Feat] Background Health Checks - Allow disabling background health checks for a specific by @ishaan-jaff in #13186
- [Proxy Startup]fix db config through envs by @jugaldb in #13111
- fix: support negative indexes in cache_control_injection_points for Anthropic Claude (#10226) by @AnandKhinvasara in #13187
- [Bug Fix] Gemini-CLI Integration - ensure tool calling works as expected on generateContent by @ishaan-jaff in #13189
- [Bug Fix] Infra - ensure that stale Prisma clients disconnect DB connection by @ishaan-jaff in #13140
- [Feat] Allow redacting message / response content for specific logging integrations - DD LLM Observability by @ishaan-jaff in #13158
- Litellm fix fallbacks UI by @ishaan-jaff in #13191
- Fix langfuse test patch path causing CI failures by @colesmcintosh in #13192
- [Docs] Add details on when to use specific health endpoints by @ishaan-jaff in #13193
- Fix - using managed files w/ OTEL + UI - add model group alias on UI by @krrishdholakia in #13171
- Anthropic - working mid-stream fallbacks by @krrishdholakia in #13149
- Anthropic - mid stream fallbacks p2 (add token usage across both calls) by @krrishdholakia in #13170
- [UI QA] QA - Agent Activity Tab by @ishaan-jaff in #13203
- Fix/panw prisma airs post call hook by @jroberts2600 in #13185
- feat(helm): allow helm hooks for migrations job by @unique-jakub in #13174
- add openssl in apk install in runtime stage in dockerfile.non_root by @mdiloreto in #13168
- Allow to redifine LLM base api URL in the pass trough endpoints by @tyumentsev4 in #13134
- Fix API Key Being Logged by @aayush-malviya-acquia in #12978
- Index.md - cleanup docs by @krrishdholakia in #13215
- [LLM translation] Fix bedrock computer use #13143 by @jugaldb in #13150
- [QA] Viewing Agent Activity Headers on UI Usage Page by @ishaan-jaff in #13212
- [Separate Health App] Update Helm Deployment.yaml by @jugaldb in #13162
- [Proxy]fix key mgmt by @jugaldb in #13148
- [LLM] fix model reload on model update by @jugaldb in #13216
- [LLM] - suppress httpx logging...
v1.74.12-nightly
What's Changed
- [Bug Fix] Pass through logging handler VertexAI - ensure multimodal embedding responses are logged by @ishaan-jaff in #13050
- Remove duplicate test case verifying field filtering logic by @emmanuel-ferdman in #13023
- Properly parse json options for key generation in the UI by @5731la in #12989
- fix: correct CompletionRequest messages type to match OpenAI API spec by @restato in #12980
- chore: Improve docs for cost tracking by @RichardoC in #12976
- feat(langfuse-otel): Add comprehensive metadata support to Langfuse OpenTelemetry integration by @strickvl in #12956
- [MCP Gateway] MCP tools fix scrolling issue by @jugaldb in #13015
- [MCP gateway] add url namespacing docs by @jugaldb in #13063
- [Bug Fix] The model gemini-2.5-flash with the merge_reasoning_content_in_choices parameter does not work by @ishaan-jaff in #13066
- [Feat] Add Google AI Studio Imagen4 model family by @ishaan-jaff in #13065
- Default Usage Chart Date Range: Last 7 Days by @NANDINI-star in #12917
- Added handling for pwd protected cert files in AOAI CertificateCreden… by @Ne0-1 in #12995
- docs: add Qwen Code CLI tutorial by @colesmcintosh in #12915
- Azure
api_version="preview"
support + Bedrock cost tracking via Anthropic/v1/messages
by @krrishdholakia in #13072 - Remove extraneous
s
in docs by @maxrabin in #13079 - Fix list team v2 security check by @tofarr in #13094
- [MCP gateway] add pre and during call hooks init by @jugaldb in #13067
- fix: improve MCP server URL validation to support internal/Kubernetes URLs by @colesmcintosh in #13099
- [Bug Fix] Gemini-CLI - The Gemini Custom API request has an incorrect authorization format by @ishaan-jaff in #13098
- set default value for mcp namespace tool name in spend table to prevent duplicate entry in table by @lvuna in #12894
- [Feat] Allow using query_params for setting API Key for generateContent routes by @ishaan-jaff in #13100
- [LLM translation] add openrouter grok4 by @jugaldb in #13018
- [Infra] Looses MCP python version restrictions by @ishaan-jaff in #13102
- fix: always use choice index=0 for Anthropic streaming responses by @Maximgitman in #12666
- BUGFIX: Jitter should be added not multiplied (#12877) by @pathikrit in #12901
- fix: helm migration job not running schema update by @huetterma in #12809
- [Bug]: Set user from token user_id for OpenMeter integration by @betterthanbreakfast in #13029
- Revert "[Bug]: Set user from token user_id for OpenMeter integration" by @krrishdholakia in #13107
- Fix fallback delete by @phosae in #12606
- Fix/gemini api key environment variable support by @sahusiddharth in #12507
- [MCP Gateway] Add protocol headers by @jugaldb in #13062
- Fix token counter to ignore unsupported keys like prefix (#11791) by @Amit-kr26 in #11954
- Custom Auth - bubble up custom exceptions by @krrishdholakia in #13093
- feat: Add dot notation support for all JWT fields by @colesmcintosh in #13013
- [Feat] MLFlow Logging - Allow adding tags for ML Flow logging requests by @ishaan-jaff in #13108
- [LLM translation] Add support for bedrock computer use by @jugaldb in #12948
- After selecting date range show loader on usage cost charts by @NANDINI-star in #13113
- Revert "[LLM translation] Add support for bedrock computer use" by @krrishdholakia in #13118
- fix(model_checks.py): handle custom values in wildcard model name (e.g. genai/test/*) by @krrishdholakia in #13116
- move to use_prisma_migrate by default + resolve team-only models on auth checks + UI - add sagemaker on UI by @krrishdholakia in #13117
- fix tool aws bedrock call index when the function only have optional arg by @jnhyperion in #13115
- [MCP Gateway] add health check endpoints for MCP by @jugaldb in #13106
- [MCP Protocol header] fix issue with clients protocol header by @jugaldb in #13112
- Added Voyage, Jinai, Deepinfra and VolcEngine providers on the UI by @NANDINI-star in #13131
- [Proxy UI] fix object permission for orgs by @jugaldb in #13142
- New Advanced Date Range Picker Component by @NANDINI-star in #13141
- [Feat] UI + Backend add a tab for use agent activity by @ishaan-jaff in #13146
- [LLM translation] Fix bedrock computer use by @jugaldb in #13143
- [MCP Guardrails] move pre and during hooks to ProxyLoggin by @jugaldb in #13109
- [Feat] v2 updates - tracking DAU, WAU, MAU for coding tool usage + show Daily Usage per User by @ishaan-jaff in #13147
- Litellm explore postgres db ci cd by @krrishdholakia in #13156
- [MCP Gateway] Litellm mcp client list fail by @jugaldb in #13114
- [MCP Gateway] fix migrations by @jugaldb in #13157
- build(config.yml): migrate build_and_test to ci/cd pg db by @krrishdholakia in #13166
- add framework name to UserAgent header in AWS Bedrock API call by @0x-fang in #13159
- fix: remove obsolete attribute
version
in docker compose by @kowyo in #13172
New Contributors
- @5731la made their first contribution in #12989
- @restato made their first contribution in #12980
- @strickvl made their first contribution in #12956
- @Ne0-1 made their first contribution in #12995
- @maxrabin made their first contribution in #13079
- @lvuna made their first contribution in #12894
- @Maximgitman made their first contribution in #12666
- @pathikrit made their first contribution in #12901
- @huetterma made their first contribution in #12809
- @betterthanbreakfast made their first contribution in #13029
- @phosae made their first contribution in #12606
- @sahusiddharth made their first contribution in #12507
- @Amit-kr26 made their first contribution in #11954
- @kowyo made their first contribution in #13172
Full Changelog: v1.74.9.rc.1...v1.74.12-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.12-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 120.0 | 157.82532362996275 | 6.312836133958831 | 0.0 | 1889 | 0 | 83.03327700002683 | 1049.7224119999373 |
Aggregated | Passed ✅ | 120.0 | 157.82532362996275 | 6.312836133958831 | 0.0 | 1889 | 0 | 83.03327700002683 | 1049.7224119999373 |
v1.74.9.rc.1
What's Changed
- Litellm release notes 07 27 2025 p1 by @krrishdholakia in #13027
- VertexAI - camelcase optional params for image generation + Anthropic - streaming, always ensure assistant role set on only first chunk by @krrishdholakia in #12889
- Bulk User Edit - additional improvements - edit all users + set 'no-default-models' on all users by @krrishdholakia in #12925
- add X-Initiator header for GitHub Copilot to reduce premium requests by @ckoehler in #13016
- docs - openwebui show how to include reasoning content for gemini models by @ishaan-jaff in #13060
New Contributors
Full Changelog: v1.74.9.rc-draft...v1.74.9.rc.1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.9.rc.1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.9.rc.1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 130.0 | 168.0418430074381 | 6.293403780045943 | 0.0 | 1882 | 0 | 97.96855799999094 | 1049.3644050000057 |
Aggregated | Passed ✅ | 130.0 | 168.0418430074381 | 6.293403780045943 | 0.0 | 1882 | 0 | 97.96855799999094 | 1049.3644050000057 |
v1.74.9-stable
What's Changed
- Litellm release notes 07 27 2025 p1 by @krrishdholakia in #13027
- VertexAI - camelcase optional params for image generation + Anthropic - streaming, always ensure assistant role set on only first chunk by @krrishdholakia in #12889
- Bulk User Edit - additional improvements - edit all users + set 'no-default-models' on all users by @krrishdholakia in #12925
- add X-Initiator header for GitHub Copilot to reduce premium requests by @ckoehler in #13016
- docs - openwebui show how to include reasoning content for gemini models by @ishaan-jaff in #13060
New Contributors
Full Changelog: v1.74.9.rc-draft...v1.74.9-stable
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.9-stable
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 110.0 | 151.02663461606545 | 6.447232352830848 | 0.0 | 1930 | 0 | 81.53728299998875 | 1408.0881720000207 |
Aggregated | Passed ✅ | 110.0 | 151.02663461606545 | 6.447232352830848 | 0.0 | 1930 | 0 | 81.53728299998875 | 1408.0881720000207 |
litellm_v1.65.4-dev_fix
Full Changelog: v1.65.4-stable...litellm_v1.65.4-dev_fix
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.65.4-dev_fix
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 150.0 | 182.3086370464627 | 6.335489722473191 | 0.0 | 1894 | 0 | 112.72804899999755 | 1087.5013490000072 |
Aggregated | Passed ✅ | 150.0 | 182.3086370464627 | 6.335489722473191 | 0.0 | 1894 | 0 | 112.72804899999755 | 1087.5013490000072 |
v1.74.9.rc-draft
What's Changed
- Fix SSO Logout | Create Unified Login Page with SSO and Username/Password Options by @NANDINI-star in #12703
- docs(moonshot): correct base url and document CN-specific endpoint by @magicalne in #12804
- Adding HolmesGPT to projects using LiteLLM by @pavangudiwada in #12798
- feat: Add Lambda AI provider support by @colesmcintosh in #12817
- [UI Bug Fix] Show correct guardrails when editing a team by @ishaan-jaff in #12823
- [Docs] Show correct list of vertex ai mistral models by @ishaan-jaff in #12828
- [LLM Translation - GH Copilot] added dynamic endpoint support by @jugaldb in #12827
- feat: add Morph provider support by @colesmcintosh in #12821
- fix: remove deprecated groq/qwen-qwq-32b and add qwen/qwen3-32b by @aholmberg in #12831
- Add Google Cloud Model Armor guardrail documentation by @colesmcintosh in #12814
- [Feat] Add Recraft Image Generation API Support - New LLM Provider by @ishaan-jaff in #12832
- [Bug Fix] - gemini leaking FD for sync calls with litellm.completion by @ishaan-jaff in #12824
- [Feat] Add fireworks -
fireworks/models/kimi-k2-instruct
by @ishaan-jaff in #12837 - [LLM Translation] add qwen-vl-plus by @jugaldb in #12829
- [Azure OpenAI Feature] - Support DefaultAzureCredential without hard-coded environment variables by @ishaan-jaff in #12841
- Docs - litellm benchmarks by @ishaan-jaff in #12842
- fix(watsonx): IBM Watsonx - use correct parameter name for tool choice by @Tomas2D in #9980
- feat: add Hyperbolic provider support by @colesmcintosh in #12826
- Litellm batch cost tracking debug by @krrishdholakia in #12782
- Passthrough Auth - make Auth checks OSS + Anthropic - only show 'reasoning_effort' for supported models by @krrishdholakia in #12847
- build(deps): bump form-data from 4.0.0 to 4.0.4 in /ui/litellm-dashboard by @dependabot[bot] in #12851
- Fix team_member_budget update logic by @krrishdholakia in #12843
- Fix async retryer on
.acompletion()
+ forward clientside headers - filter outcontent-type
from clientside request (causes llm api call to hang) by @krrishdholakia in #12848 - Openrouter - filter out cache_control flag for non-anthropic models (allows usage with claude code) by @krrishdholakia in #12850
- Improvements on the Regenerate Key Flow by @NANDINI-star in #12788
- Replace non-root Dockerfile base with Alpine multi-stage build; by @mdiloreto in #12707
- build(deps): bump form-data from 4.0.3 to 4.0.4 in /docs/my-website by @dependabot[bot] in #12867
- Bug fix - Azure KeyVault not in image, add
azure-keyvault==4.2.0
to Docker img by @ishaan-jaff in #12873 - [Separate Health App] Pass through cmd args via supervisord by @jugaldb in #12871
- [Feat] Add Recraft API - Image Edits Support by @ishaan-jaff in #12874
- [LLM Translation] Litellm gemini 2.0 live support by @jugaldb in #12839
- [Feat] Add cost tracking for new vertex_ai/llama-3 API models by @ishaan-jaff in #12878
- [Feat] - Track cost + add tags for health checks done by LiteLLM Proxy by @ishaan-jaff in #12880
- Fix async retryer on .acompletion() + forward clientside headers - filter out content-type from clientside request (causes llm api call to hang) by @krrishdholakia in #12886
- Request Headers - support
x-litellm-num-retries
+ Usage - support usage by model group by @krrishdholakia in #12890 - [Docs ]Litellm mcp access group doc by @jugaldb in #12883
- [Add health check] add architecture diagram by @jugaldb in #12879
- rm retired anthropic models from model_prices_and_context_window.json by @staeiou in #12864
- feat: extended
/v1/models
endpoint, now it returns with fallbacks on demand by @murad-khafizov in #12811 - feat: Add Pillar Security guardrail integration by @eagle-p in #12791
- Swagger - bump DOMPurify version + add back Support local swagger bundle (good for airgapped environments) + support local swagger on custom root paths by @krrishdholakia in #12911
- [Feat] LiteLLM x Cloudzero integration - Allow exporting spend to cloudzero by @ishaan-jaff in #12908
- [Feat] LiteLLM CloudZero Integration updates - using LiteLLM_SpendLogs Table by @ishaan-jaff in #12922
- Add GA version of gemini 2.5 flash lite for both vertex and gemini by @apoorv-sharma in #12920
- [UI] Allow setting up CloudZero Usage through LiteLLM UI by @ishaan-jaff in #12923
- Update control_plane_and_data_plane.md by @superpoussin22 in #12939
- Prometheus - tags, fix '[tag]="false"' when tag is set by @krrishdholakia in #12916
- Proxy - specify
key_type
- allows specifying if key can call LLM API routes vs. Management routes only by @krrishdholakia in #12909 - fix(internal_user_endpoints.py): delete member from team table on
/user/delete
by @krrishdholakia in #12926 - [Feat] Backend Router - Add Auto-Router powered by
semantic-router
by @ishaan-jaff in #12955 - [LLM Translation] - Bug fix Anthropic Tool calling by @ishaan-jaff in #12959
- [LLM Translation] added new realtime model for openai by @jugaldb in #12946
- [Feat] UI - Allow Adding LiteLLM Auto Router on UI by @ishaan-jaff in #12960
- [Feat] Edit Auto Router Settings on UI by @ishaan-jaff in #12966
- [Feat] Add inpainting support and corresponding tests for Amazon Nova… by @SantoshDhaladhuli in #12949
- docs: added documentation about metadata exposed over the
/v1/models
endpoint by @murad-khafizov in #12942 - Fix: Shorten Gemini tool_call_id for Open AI compatibility by @tonga54 in #12941
- feat: add openrouter/qwen/qwen3-coder model configuration by @colesmcintosh in #12910
- fix(auth_utils): make header comparison case-insensitive by @sings-to-bees-on-wednesdays in #12950
- GuardrailsAI: use validatedOutput to allow usage of "fix" guards by @DmitriyAlergant in #12891
- Show global retry policy on UI by @krrishdholakia in #12969
- Guardrails - support model-level guardrails by @krrishdholakia in #12968
- build(github/manual_pypi_publish.yml): manual workflow to publish pip package - used for pushing dev releases by @krrishdholakia in #12985
- [LLM Translation] Add bytedance/ui-tars-1.5-7b on openrouter by @jugaldb in #12882
- clean and verify key before inserting by @jugaldb in #12840
- [LLM Translation] fix query params for realtime api intent by @jugaldb in #12838
- [MCP Gateway] Move cost tracking and permission management to OSS by @jugaldb in #12988
- build: update pip package by @krrishdholakia in #12998
- Fix issue writing db by @emerzon in #13001
- [BUG Fix] Cannot pickle coroutine object by @jugaldb in #13005
- [MCP Gateway] add Litellm mcp alias for prefixing by @jugaldb in #12994
- Fixup ollama model listing (again) by @matthid in #13008
- [Vector Store] make vector store permission management OSS by @jugaldb in #12990
- [FEAT] Model-Guardrails: Add on UI by @krrishdholakia in #13006
- UI SSO - fix reset env var when ui_access_mode is updated by @krrishdholakia in #13011
- [MCP Gateway] Litellm mcp multi header propagation by @jugaldb in #13003
New Contributors
- @magicalne made their first contribution in h...