Skip to content

Releases: BerriAI/litellm

v1.40.29

28 Jun 02:24
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.40.28...v1.40.29

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.29

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 150.0 169.90346260031845 6.295057404345822 0.0 1884 0 116.81983199997603 1212.0624549999661
Aggregated Passed ✅ 150.0 169.90346260031845 6.295057404345822 0.0 1884 0 116.81983199997603 1212.0624549999661

v1.40.28

27 Jun 02:48
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.40.27...v1.40.28

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.28

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 140.0 159.37771652368323 6.278392214516223 0.0 1879 0 117.58081900001116 1089.7057880000034
Aggregated Passed ✅ 140.0 159.37771652368323 6.278392214516223 0.0 1879 0 117.58081900001116 1089.7057880000034

v1.40.27

26 Jun 03:48
d772b8d
Compare
Choose a tag to compare

✨ Thrilled to launch support for @NVIDIA NIM LLM API on @LiteLLM 1.40.27 👉 Start here: https://docs.litellm.ai/docs/providers/nvidia_nim

🔥 Proxy 100+ LLMS & set budgets

🔑 [Enterprise] Add secret detection pre call hook https://docs.litellm.ai/docs/proxy/enterprise#content-moderation

🛠️ [Fix] - use n in mock completion response on litellm mock responses

⚡️ [Feat] add endpoint to debug memory utilization

🔑 enterprise - allow verifying license in air gapped vpc
Group 5865

What's Changed

New Contributors

Full Changelog: v1.40.26...v1.40.27

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.27

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 130.0 156.61068343517005 6.372506185089714 0.0 1905 0 109.52021800000011 1799.9076889999515
Aggregated Passed ✅ 130.0 156.61068343517005 6.372506185089714 0.0 1905 0 109.52021800000011 1799.9076889999515

v1.40.26

25 Jun 05:48
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.40.25...v1.40.26

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.26

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 130.0 156.53979613674755 6.329682565511454 0.0 1894 0 109.70386399998233 2175.4312479999953
Aggregated Passed ✅ 130.0 156.53979613674755 6.329682565511454 0.0 1894 0 109.70386399998233 2175.4312479999953

v1.40.25

23 Jun 08:35
Compare
Choose a tag to compare

What's Changed

  • feat(dynamic_rate_limiter.py): Dynamic tpm quota (multiple projects) by @krrishdholakia in #4349
  • fix(router.py): Content Policy Fallbacks for Azure 'content_filter' responses by @krrishdholakia in #4365
  • Disable message redaction in logs via request header by @msabramo in #4352

Full Changelog: v1.40.24...v1.40.25

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.25

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 120.0 137.1328113917474 6.561515229902106 0.0 1963 0 98.4713450000072 1831.7410280000104
Aggregated Passed ✅ 120.0 137.1328113917474 6.561515229902106 0.0 1963 0 98.4713450000072 1831.7410280000104

v1.40.24

23 Jun 01:43
Compare
Choose a tag to compare

What's Changed

  • refactor(litellm_logging.py): refactors how slack_alerting generates langfuse trace url by @krrishdholakia in #4344
  • [Security Fix - Proxy Server ADMIN UI] - Store credentials in cookies + use strong JWT signing secret by @ishaan-jaff in #4357
  • [Test] Test routes on LiteLLM Proxy always includes OpenAI Routes by @ishaan-jaff in #4356
  • fix - Can't access /v1/audio/speech with some user key by @ishaan-jaff in #4360

Full Changelog: v1.40.22...v1.40.24

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.24

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 130.0 154.96801309401323 6.361794835434345 0.0 1904 0 115.54615199997897 1161.0779019999882
Aggregated Passed ✅ 130.0 154.96801309401323 6.361794835434345 0.0 1904 0 115.54615199997897 1161.0779019999882

v1.40.22

22 Jun 04:25
Compare
Choose a tag to compare

What's Changed

  • fix: use per-token costs for claude via vertex_ai by @spdustin in #4337
  • [Feat] Admin UI - Show Cache hit stats by @ishaan-jaff in #4340
  • fix - liteLLM proxy /moderations endpoint returns 500 error when model is not specified by @ishaan-jaff in #4342
  • [Fix + Test] - Spend tags not getting stored on 1.40.9 by @ishaan-jaff in #4345
  • Print content window fallbacks on startup to help verify configuration by @lolsborn in #4350

New Contributors

Full Changelog: v1.40.21...v1.40.22

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.22

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 120.0 147.06905652027004 6.431081863451109 0.0 1924 0 100.04098199999589 1834.3141159999732
Aggregated Passed ✅ 120.0 147.06905652027004 6.431081863451109 0.0 1924 0 100.04098199999589 1834.3141159999732

v1.40.21

21 Jun 19:26
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.40.20...v1.40.21

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.21

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 140.0 163.37791068962042 6.407997421786419 0.0 1917 0 114.2956310000045 1749.6762069999932
Aggregated Passed ✅ 140.0 163.37791068962042 6.407997421786419 0.0 1917 0 114.2956310000045 1749.6762069999932

v1.40.20

21 Jun 05:39
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.40.19...v1.40.20

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.20

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 79 105.64488981589672 6.515118790194818 0.0 1950 0 67.60507300003837 3342.9461570000285
Aggregated Passed ✅ 79 105.64488981589672 6.515118790194818 0.0 1950 0 67.60507300003837 3342.9461570000285

v1.40.19

20 Jun 22:06
Compare
Choose a tag to compare

🚨🚨🚨 Known Bug on LiteLLM Proxy Server on this Release- we do not recommend upgrading until issue is fixed

You can use claude 3.5 sonnet on older versions - no upgrade is required

What's Changed

New Contributors

Full Changelog: v1.40.17...v1.40.19