Skip to content

Releases: BerriAI/litellm

v1.74.15.rc.2

05 Aug 05:33
Compare
Choose a tag to compare

Full Changelog: 1.74.15.rc.1...v1.74.15.rc.2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.15.rc.2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 140.0 181.79772817994646 6.340033721452323 0.0 1895 0 111.80489700001317 1370.5411799999752
Aggregated Passed ✅ 140.0 181.79772817994646 6.340033721452323 0.0 1895 0 111.80489700001317 1370.5411799999752

v1.74.9-stable.patch.1

03 Aug 06:51
Compare
Choose a tag to compare

Full Changelog: v1.74.9-stable...v1.74.9-stable.patch.1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.9-stable.patch.1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 160.0 195.02149162499995 6.2583306558386465 0.0 1872 0 124.22932999999148 1133.3790610000278
Aggregated Passed ✅ 160.0 195.02149162499995 6.2583306558386465 0.0 1872 0 124.22932999999148 1133.3790610000278

v1.74.15-nightly

02 Aug 22:35
Compare
Choose a tag to compare

What's Changed

  • Add advanced date picker to all the tabs on the usage page by @NANDINI-star in #13221
  • Add Perplexity citation annotations support by @kankute-sameer in #13225
  • fix: role chaining and session name with webauthentication for aws bedrock by @RichardoC in #13205
  • Fix missing extra_headers support for vLLM/openai_like embeddings by @colesmcintosh in #13198
  • litellm/proxy: preserve model order of /v1/models and /model_group/info by @AlexanderYastrebov in #13178
  • Prompt Management - abstract prompt templates away from model list (enables permission management on prompt templates) by @krrishdholakia in #13219
  • [QA Fixes for MCP] - Ensure MCPs load + don't run a health check everytime we load MCPs on UI by @ishaan-jaff in #13228
  • Revert "fix: role chaining and session name with webauthentication for aws bedrock" by @krrishdholakia in #13230
  • Prompt Management (2/2) - New /prompt/list endpoint + key-based access to prompt templates by @krrishdholakia in #13218
  • [UI QA Fixes] Stable release by @ishaan-jaff in #13231

New Contributors

Full Changelog: v1.74.14-nightly...v1.74.15-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.15-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 160.0 201.90947012894696 6.350644454200707 0.0 1900 0 128.45073899995896 1495.8115740000153
Aggregated Passed ✅ 160.0 201.90947012894696 6.350644454200707 0.0 1900 0 128.45073899995896 1495.8115740000153

1.74.15.rc.1

02 Aug 22:50
Compare
Choose a tag to compare

What's Changed

  • Add advanced date picker to all the tabs on the usage page by @NANDINI-star in #13221
  • Add Perplexity citation annotations support by @kankute-sameer in #13225
  • fix: role chaining and session name with webauthentication for aws bedrock by @RichardoC in #13205
  • Fix missing extra_headers support for vLLM/openai_like embeddings by @colesmcintosh in #13198
  • litellm/proxy: preserve model order of /v1/models and /model_group/info by @AlexanderYastrebov in #13178
  • Prompt Management - abstract prompt templates away from model list (enables permission management on prompt templates) by @krrishdholakia in #13219
  • [QA Fixes for MCP] - Ensure MCPs load + don't run a health check everytime we load MCPs on UI by @ishaan-jaff in #13228
  • Revert "fix: role chaining and session name with webauthentication for aws bedrock" by @krrishdholakia in #13230
  • Prompt Management (2/2) - New /prompt/list endpoint + key-based access to prompt templates by @krrishdholakia in #13218
  • [UI QA Fixes] Stable release by @ishaan-jaff in #13231

New Contributors

Full Changelog: v1.74.14-nightly...1.74.15.rc.1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-1.74.15.rc.1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 160.0 194.82821616116973 6.285152111114945 0.0 1880 0 124.01886500003911 1206.761389999997
Aggregated Passed ✅ 160.0 194.82821616116973 6.285152111114945 0.0 1880 0 124.01886500003911 1206.761389999997

v1.74.14.dev1

02 Aug 16:47
Compare
Choose a tag to compare

Full Changelog: v1.74.14-nightly...v1.74.14.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.14.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.14.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 160.0 199.19181043322646 6.258277007293016 0.0 1872 0 130.81277300000238 1170.560404000014
Aggregated Passed ✅ 160.0 199.19181043322646 6.258277007293016 0.0 1872 0 130.81277300000238 1170.560404000014

v1.74.14-nightly

02 Aug 05:55
Compare
Choose a tag to compare

What's Changed

Read more

v1.74.12-nightly

31 Jul 20:53
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.74.9.rc.1...v1.74.12-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.12-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 120.0 157.82532362996275 6.312836133958831 0.0 1889 0 83.03327700002683 1049.7224119999373
Aggregated Passed ✅ 120.0 157.82532362996275 6.312836133958831 0.0 1889 0 83.03327700002683 1049.7224119999373

v1.74.9.rc.1

29 Jul 01:37
Compare
Choose a tag to compare

What's Changed

  • Litellm release notes 07 27 2025 p1 by @krrishdholakia in #13027
  • VertexAI - camelcase optional params for image generation + Anthropic - streaming, always ensure assistant role set on only first chunk by @krrishdholakia in #12889
  • Bulk User Edit - additional improvements - edit all users + set 'no-default-models' on all users by @krrishdholakia in #12925
  • add X-Initiator header for GitHub Copilot to reduce premium requests by @ckoehler in #13016
  • docs - openwebui show how to include reasoning content for gemini models by @ishaan-jaff in #13060

New Contributors

Full Changelog: v1.74.9.rc-draft...v1.74.9.rc.1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.9.rc.1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.9.rc.1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 130.0 168.0418430074381 6.293403780045943 0.0 1882 0 97.96855799999094 1049.3644050000057
Aggregated Passed ✅ 130.0 168.0418430074381 6.293403780045943 0.0 1882 0 97.96855799999094 1049.3644050000057

v1.74.9-stable

01 Aug 21:40
Compare
Choose a tag to compare

What's Changed

  • Litellm release notes 07 27 2025 p1 by @krrishdholakia in #13027
  • VertexAI - camelcase optional params for image generation + Anthropic - streaming, always ensure assistant role set on only first chunk by @krrishdholakia in #12889
  • Bulk User Edit - additional improvements - edit all users + set 'no-default-models' on all users by @krrishdholakia in #12925
  • add X-Initiator header for GitHub Copilot to reduce premium requests by @ckoehler in #13016
  • docs - openwebui show how to include reasoning content for gemini models by @ishaan-jaff in #13060

New Contributors

Full Changelog: v1.74.9.rc-draft...v1.74.9-stable

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.9-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 110.0 151.02663461606545 6.447232352830848 0.0 1930 0 81.53728299998875 1408.0881720000207
Aggregated Passed ✅ 110.0 151.02663461606545 6.447232352830848 0.0 1930 0 81.53728299998875 1408.0881720000207

litellm_v1.65.4-dev_fix

28 Jul 23:32
Compare
Choose a tag to compare

Full Changelog: v1.65.4-stable...litellm_v1.65.4-dev_fix

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.65.4-dev_fix

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 150.0 182.3086370464627 6.335489722473191 0.0 1894 0 112.72804899999755 1087.5013490000072
Aggregated Passed ✅ 150.0 182.3086370464627 6.335489722473191 0.0 1894 0 112.72804899999755 1087.5013490000072