Skip to content

v1.72.9-nightly

Compare
Choose a tag to compare
@github-actions github-actions released this 21 Jun 04:53

What's Changed

  • [Feat] MCP - Allow connecting to MCP with authentication headers + Allow clients to specify MCP headers (#11890) by @ishaan-jaff in #11891
  • [Fix] Networking - allow using CA Bundles by @ishaan-jaff in #11906
  • [Feat] Add AWS Bedrock profiles for the APAC region by @lgruen-vcgs in #11883
  • bumps the anthropic package by @rinormaloku in #11851
  • Add deployment annotations by @InvisibleMan1306 in #11849
  • Enhance Mistral API: Add support for parallel tool calls by @njbrake in #11770
  • [UI] QA Items for adding pass through endpoints by @ishaan-jaff in #11909
  • build(model_prices_and_context_window.json): mark all gemini-2.5 models support pdf input + Set anthropic custom llm provider property by @krrishdholakia in #11907
  • fix(proxy_server.py): fix loading ui on custom root path by @krrishdholakia in #11912
  • LiteLLM SDK <-> Proxy improvement (don't transform message client-side) + Bedrock - handle qs:.. in base64 file data + Tag Management - support adding public model names by @krrishdholakia in #11908
  • Add success modal for health check responses by @colesmcintosh in #11899
  • Volcengine - thinking param support + Azure - handle more gpt custom naming patterns by @krrishdholakia in #11914
  • [Feat] Model Cost Map - Add gemini-2.5-pro and set gemini-2.5-pro supports_reasoning=True by @ishaan-jaff in #11927
  • [Feat] UI Allow testing /v1/messages on the Test Key Page by @ishaan-jaff in #11930
  • Feat/add delete callback by @jtong99 in #11654
  • add ciphers in command and pass to hypercorn for proxy by @frankzye in #11916
  • [Bug Fix] Fix model_group tracked for /v1/messages and /moderations by @ishaan-jaff in #11933
  • [Bug Fix] Cost tracking and logging via the /v1/messages API are not working when using Claude Code by @ishaan-jaff in #11928
  • [Feat] Add Azure Codex Models on LiteLLM + new /v1 preview Azure OpenAI API by @ishaan-jaff in #11934
  • [Feat] UI QA: Pass through endpoints by @ishaan-jaff in #11939

New Contributors

Full Changelog: v1.72.7-nightly...v1.72.9-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.72.9-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 240.0 259.91299670392254 6.2187422072270495 0.0 1861 0 210.9276310000041 1676.9406920000165
Aggregated Passed ✅ 240.0 259.91299670392254 6.2187422072270495 0.0 1861 0 210.9276310000041 1676.9406920000165