v1.72.9-nightly
What's Changed
- [Feat] MCP - Allow connecting to MCP with authentication headers + Allow clients to specify MCP headers (#11890) by @ishaan-jaff in #11891
- [Fix] Networking - allow using CA Bundles by @ishaan-jaff in #11906
- [Feat] Add AWS Bedrock profiles for the APAC region by @lgruen-vcgs in #11883
- bumps the anthropic package by @rinormaloku in #11851
- Add deployment annotations by @InvisibleMan1306 in #11849
- Enhance Mistral API: Add support for parallel tool calls by @njbrake in #11770
- [UI] QA Items for adding pass through endpoints by @ishaan-jaff in #11909
- build(model_prices_and_context_window.json): mark all gemini-2.5 models support pdf input + Set anthropic custom llm provider property by @krrishdholakia in #11907
- fix(proxy_server.py): fix loading ui on custom root path by @krrishdholakia in #11912
- LiteLLM SDK <-> Proxy improvement (don't transform message client-side) + Bedrock - handle
qs:..
in base64 file data + Tag Management - support adding public model names by @krrishdholakia in #11908 - Add success modal for health check responses by @colesmcintosh in #11899
- Volcengine - thinking param support + Azure - handle more gpt custom naming patterns by @krrishdholakia in #11914
- [Feat] Model Cost Map - Add
gemini-2.5-pro
and setgemini-2.5-pro
supports_reasoning=True by @ishaan-jaff in #11927 - [Feat] UI Allow testing /v1/messages on the Test Key Page by @ishaan-jaff in #11930
- Feat/add delete callback by @jtong99 in #11654
- add ciphers in command and pass to hypercorn for proxy by @frankzye in #11916
- [Bug Fix] Fix model_group tracked for /v1/messages and /moderations by @ishaan-jaff in #11933
- [Bug Fix] Cost tracking and logging via the /v1/messages API are not working when using Claude Code by @ishaan-jaff in #11928
- [Feat] Add Azure Codex Models on LiteLLM + new /v1 preview Azure OpenAI API by @ishaan-jaff in #11934
- [Feat] UI QA: Pass through endpoints by @ishaan-jaff in #11939
New Contributors
- @lgruen-vcgs made their first contribution in #11883
- @rinormaloku made their first contribution in #11851
- @InvisibleMan1306 made their first contribution in #11849
Full Changelog: v1.72.7-nightly...v1.72.9-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.72.9-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 240.0 | 259.91299670392254 | 6.2187422072270495 | 0.0 | 1861 | 0 | 210.9276310000041 | 1676.9406920000165 |
Aggregated | Passed ✅ | 240.0 | 259.91299670392254 | 6.2187422072270495 | 0.0 | 1861 | 0 | 210.9276310000041 | 1676.9406920000165 |