Releases: BerriAI/litellm
v1.74.15.rc.2
Full Changelog: 1.74.15.rc.1...v1.74.15.rc.2
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.15.rc.2
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 140.0 | 181.79772817994646 | 6.340033721452323 | 0.0 | 1895 | 0 | 111.80489700001317 | 1370.5411799999752 |
Aggregated | Passed ✅ | 140.0 | 181.79772817994646 | 6.340033721452323 | 0.0 | 1895 | 0 | 111.80489700001317 | 1370.5411799999752 |
v1.74.9-stable.patch.1
Full Changelog: v1.74.9-stable...v1.74.9-stable.patch.1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.9-stable.patch.1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 160.0 | 195.02149162499995 | 6.2583306558386465 | 0.0 | 1872 | 0 | 124.22932999999148 | 1133.3790610000278 |
Aggregated | Passed ✅ | 160.0 | 195.02149162499995 | 6.2583306558386465 | 0.0 | 1872 | 0 | 124.22932999999148 | 1133.3790610000278 |
v1.74.15-nightly
What's Changed
- Add advanced date picker to all the tabs on the usage page by @NANDINI-star in #13221
- Add Perplexity citation annotations support by @kankute-sameer in #13225
- fix: role chaining and session name with webauthentication for aws bedrock by @RichardoC in #13205
- Fix missing extra_headers support for vLLM/openai_like embeddings by @colesmcintosh in #13198
- litellm/proxy: preserve model order of /v1/models and /model_group/info by @AlexanderYastrebov in #13178
- Prompt Management - abstract prompt templates away from model list (enables permission management on prompt templates) by @krrishdholakia in #13219
- [QA Fixes for MCP] - Ensure MCPs load + don't run a health check everytime we load MCPs on UI by @ishaan-jaff in #13228
- Revert "fix: role chaining and session name with webauthentication for aws bedrock" by @krrishdholakia in #13230
- Prompt Management (2/2) - New
/prompt/list
endpoint + key-based access to prompt templates by @krrishdholakia in #13218 - [UI QA Fixes] Stable release by @ishaan-jaff in #13231
New Contributors
- @kankute-sameer made their first contribution in #13225
- @AlexanderYastrebov made their first contribution in #13178
Full Changelog: v1.74.14-nightly...v1.74.15-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.15-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 160.0 | 201.90947012894696 | 6.350644454200707 | 0.0 | 1900 | 0 | 128.45073899995896 | 1495.8115740000153 |
Aggregated | Passed ✅ | 160.0 | 201.90947012894696 | 6.350644454200707 | 0.0 | 1900 | 0 | 128.45073899995896 | 1495.8115740000153 |
1.74.15.rc.1
What's Changed
- Add advanced date picker to all the tabs on the usage page by @NANDINI-star in #13221
- Add Perplexity citation annotations support by @kankute-sameer in #13225
- fix: role chaining and session name with webauthentication for aws bedrock by @RichardoC in #13205
- Fix missing extra_headers support for vLLM/openai_like embeddings by @colesmcintosh in #13198
- litellm/proxy: preserve model order of /v1/models and /model_group/info by @AlexanderYastrebov in #13178
- Prompt Management - abstract prompt templates away from model list (enables permission management on prompt templates) by @krrishdholakia in #13219
- [QA Fixes for MCP] - Ensure MCPs load + don't run a health check everytime we load MCPs on UI by @ishaan-jaff in #13228
- Revert "fix: role chaining and session name with webauthentication for aws bedrock" by @krrishdholakia in #13230
- Prompt Management (2/2) - New
/prompt/list
endpoint + key-based access to prompt templates by @krrishdholakia in #13218 - [UI QA Fixes] Stable release by @ishaan-jaff in #13231
New Contributors
- @kankute-sameer made their first contribution in #13225
- @AlexanderYastrebov made their first contribution in #13178
Full Changelog: v1.74.14-nightly...1.74.15.rc.1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-1.74.15.rc.1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 160.0 | 194.82821616116973 | 6.285152111114945 | 0.0 | 1880 | 0 | 124.01886500003911 | 1206.761389999997 |
Aggregated | Passed ✅ | 160.0 | 194.82821616116973 | 6.285152111114945 | 0.0 | 1880 | 0 | 124.01886500003911 | 1206.761389999997 |
v1.74.14.dev1
Full Changelog: v1.74.14-nightly...v1.74.14.dev1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.14.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.14.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 160.0 | 199.19181043322646 | 6.258277007293016 | 0.0 | 1872 | 0 | 130.81277300000238 | 1170.560404000014 |
Aggregated | Passed ✅ | 160.0 | 199.19181043322646 | 6.258277007293016 | 0.0 | 1872 | 0 | 130.81277300000238 | 1170.560404000014 |
v1.74.14-nightly
What's Changed
- [Bug Fix] Pass through logging handler VertexAI - ensure multimodal embedding responses are logged by @ishaan-jaff in #13050
- Remove duplicate test case verifying field filtering logic by @emmanuel-ferdman in #13023
- Properly parse json options for key generation in the UI by @5731la in #12989
- fix: correct CompletionRequest messages type to match OpenAI API spec by @restato in #12980
- chore: Improve docs for cost tracking by @RichardoC in #12976
- feat(langfuse-otel): Add comprehensive metadata support to Langfuse OpenTelemetry integration by @strickvl in #12956
- [MCP Gateway] MCP tools fix scrolling issue by @jugaldb in #13015
- [MCP gateway] add url namespacing docs by @jugaldb in #13063
- [Bug Fix] The model gemini-2.5-flash with the merge_reasoning_content_in_choices parameter does not work by @ishaan-jaff in #13066
- [Feat] Add Google AI Studio Imagen4 model family by @ishaan-jaff in #13065
- Default Usage Chart Date Range: Last 7 Days by @NANDINI-star in #12917
- Added handling for pwd protected cert files in AOAI CertificateCreden… by @Ne0-1 in #12995
- docs: add Qwen Code CLI tutorial by @colesmcintosh in #12915
- Azure
api_version="preview"
support + Bedrock cost tracking via Anthropic/v1/messages
by @krrishdholakia in #13072 - Remove extraneous
s
in docs by @maxrabin in #13079 - Fix list team v2 security check by @tofarr in #13094
- [MCP gateway] add pre and during call hooks init by @jugaldb in #13067
- fix: improve MCP server URL validation to support internal/Kubernetes URLs by @colesmcintosh in #13099
- [Bug Fix] Gemini-CLI - The Gemini Custom API request has an incorrect authorization format by @ishaan-jaff in #13098
- set default value for mcp namespace tool name in spend table to prevent duplicate entry in table by @lvuna in #12894
- [Feat] Allow using query_params for setting API Key for generateContent routes by @ishaan-jaff in #13100
- [LLM translation] add openrouter grok4 by @jugaldb in #13018
- [Infra] Looses MCP python version restrictions by @ishaan-jaff in #13102
- fix: always use choice index=0 for Anthropic streaming responses by @Maximgitman in #12666
- BUGFIX: Jitter should be added not multiplied (#12877) by @pathikrit in #12901
- fix: helm migration job not running schema update by @huetterma in #12809
- [Bug]: Set user from token user_id for OpenMeter integration by @betterthanbreakfast in #13029
- Revert "[Bug]: Set user from token user_id for OpenMeter integration" by @krrishdholakia in #13107
- Fix fallback delete by @phosae in #12606
- Fix/gemini api key environment variable support by @sahusiddharth in #12507
- [MCP Gateway] Add protocol headers by @jugaldb in #13062
- Fix token counter to ignore unsupported keys like prefix (#11791) by @Amit-kr26 in #11954
- Custom Auth - bubble up custom exceptions by @krrishdholakia in #13093
- feat: Add dot notation support for all JWT fields by @colesmcintosh in #13013
- [Feat] MLFlow Logging - Allow adding tags for ML Flow logging requests by @ishaan-jaff in #13108
- [LLM translation] Add support for bedrock computer use by @jugaldb in #12948
- After selecting date range show loader on usage cost charts by @NANDINI-star in #13113
- Revert "[LLM translation] Add support for bedrock computer use" by @krrishdholakia in #13118
- fix(model_checks.py): handle custom values in wildcard model name (e.g. genai/test/*) by @krrishdholakia in #13116
- move to use_prisma_migrate by default + resolve team-only models on auth checks + UI - add sagemaker on UI by @krrishdholakia in #13117
- fix tool aws bedrock call index when the function only have optional arg by @jnhyperion in #13115
- [MCP Gateway] add health check endpoints for MCP by @jugaldb in #13106
- [MCP Protocol header] fix issue with clients protocol header by @jugaldb in #13112
- Added Voyage, Jinai, Deepinfra and VolcEngine providers on the UI by @NANDINI-star in #13131
- [Proxy UI] fix object permission for orgs by @jugaldb in #13142
- New Advanced Date Range Picker Component by @NANDINI-star in #13141
- [Feat] UI + Backend add a tab for use agent activity by @ishaan-jaff in #13146
- [LLM translation] Fix bedrock computer use by @jugaldb in #13143
- [MCP Guardrails] move pre and during hooks to ProxyLoggin by @jugaldb in #13109
- [Feat] v2 updates - tracking DAU, WAU, MAU for coding tool usage + show Daily Usage per User by @ishaan-jaff in #13147
- Litellm explore postgres db ci cd by @krrishdholakia in #13156
- [MCP Gateway] Litellm mcp client list fail by @jugaldb in #13114
- [MCP Gateway] fix migrations by @jugaldb in #13157
- build(config.yml): migrate build_and_test to ci/cd pg db by @krrishdholakia in #13166
- add framework name to UserAgent header in AWS Bedrock API call by @0x-fang in #13159
- fix: remove obsolete attribute
version
in docker compose by @kowyo in #13172 - [Feat] Background Health Checks - Allow disabling background health checks for a specific by @ishaan-jaff in #13186
- [Proxy Startup]fix db config through envs by @jugaldb in #13111
- fix: support negative indexes in cache_control_injection_points for Anthropic Claude (#10226) by @AnandKhinvasara in #13187
- [Bug Fix] Gemini-CLI Integration - ensure tool calling works as expected on generateContent by @ishaan-jaff in #13189
- [Bug Fix] Infra - ensure that stale Prisma clients disconnect DB connection by @ishaan-jaff in #13140
- [Feat] Allow redacting message / response content for specific logging integrations - DD LLM Observability by @ishaan-jaff in #13158
- Litellm fix fallbacks UI by @ishaan-jaff in #13191
- Fix langfuse test patch path causing CI failures by @colesmcintosh in #13192
- [Docs] Add details on when to use specific health endpoints by @ishaan-jaff in #13193
- Fix - using managed files w/ OTEL + UI - add model group alias on UI by @krrishdholakia in #13171
- Anthropic - working mid-stream fallbacks by @krrishdholakia in #13149
- Anthropic - mid stream fallbacks p2 (add token usage across both calls) by @krrishdholakia in #13170
- [UI QA] QA - Agent Activity Tab by @ishaan-jaff in #13203
- Fix/panw prisma airs post call hook by @jroberts2600 in #13185
- feat(helm): allow helm hooks for migrations job by @unique-jakub in #13174
- add openssl in apk install in runtime stage in dockerfile.non_root by @mdiloreto in #13168
- Allow to redifine LLM base api URL in the pass trough endpoints by @tyumentsev4 in #13134
- Fix API Key Being Logged by @aayush-malviya-acquia in #12978
- Index.md - cleanup docs by @krrishdholakia in #13215
- [LLM translation] Fix bedrock computer use #13143 by @jugaldb in #13150
- [QA] Viewing Agent Activity Headers on UI Usage Page by @ishaan-jaff in #13212
- [Separate Health App] Update Helm Deployment.yaml by @jugaldb in #13162
- [Proxy]fix key mgmt by @jugaldb in #13148
- [LLM] fix model reload on model update by @jugaldb in #13216
- [LLM] - suppress httpx logging...
v1.74.12-nightly
What's Changed
- [Bug Fix] Pass through logging handler VertexAI - ensure multimodal embedding responses are logged by @ishaan-jaff in #13050
- Remove duplicate test case verifying field filtering logic by @emmanuel-ferdman in #13023
- Properly parse json options for key generation in the UI by @5731la in #12989
- fix: correct CompletionRequest messages type to match OpenAI API spec by @restato in #12980
- chore: Improve docs for cost tracking by @RichardoC in #12976
- feat(langfuse-otel): Add comprehensive metadata support to Langfuse OpenTelemetry integration by @strickvl in #12956
- [MCP Gateway] MCP tools fix scrolling issue by @jugaldb in #13015
- [MCP gateway] add url namespacing docs by @jugaldb in #13063
- [Bug Fix] The model gemini-2.5-flash with the merge_reasoning_content_in_choices parameter does not work by @ishaan-jaff in #13066
- [Feat] Add Google AI Studio Imagen4 model family by @ishaan-jaff in #13065
- Default Usage Chart Date Range: Last 7 Days by @NANDINI-star in #12917
- Added handling for pwd protected cert files in AOAI CertificateCreden… by @Ne0-1 in #12995
- docs: add Qwen Code CLI tutorial by @colesmcintosh in #12915
- Azure
api_version="preview"
support + Bedrock cost tracking via Anthropic/v1/messages
by @krrishdholakia in #13072 - Remove extraneous
s
in docs by @maxrabin in #13079 - Fix list team v2 security check by @tofarr in #13094
- [MCP gateway] add pre and during call hooks init by @jugaldb in #13067
- fix: improve MCP server URL validation to support internal/Kubernetes URLs by @colesmcintosh in #13099
- [Bug Fix] Gemini-CLI - The Gemini Custom API request has an incorrect authorization format by @ishaan-jaff in #13098
- set default value for mcp namespace tool name in spend table to prevent duplicate entry in table by @lvuna in #12894
- [Feat] Allow using query_params for setting API Key for generateContent routes by @ishaan-jaff in #13100
- [LLM translation] add openrouter grok4 by @jugaldb in #13018
- [Infra] Looses MCP python version restrictions by @ishaan-jaff in #13102
- fix: always use choice index=0 for Anthropic streaming responses by @Maximgitman in #12666
- BUGFIX: Jitter should be added not multiplied (#12877) by @pathikrit in #12901
- fix: helm migration job not running schema update by @huetterma in #12809
- [Bug]: Set user from token user_id for OpenMeter integration by @betterthanbreakfast in #13029
- Revert "[Bug]: Set user from token user_id for OpenMeter integration" by @krrishdholakia in #13107
- Fix fallback delete by @phosae in #12606
- Fix/gemini api key environment variable support by @sahusiddharth in #12507
- [MCP Gateway] Add protocol headers by @jugaldb in #13062
- Fix token counter to ignore unsupported keys like prefix (#11791) by @Amit-kr26 in #11954
- Custom Auth - bubble up custom exceptions by @krrishdholakia in #13093
- feat: Add dot notation support for all JWT fields by @colesmcintosh in #13013
- [Feat] MLFlow Logging - Allow adding tags for ML Flow logging requests by @ishaan-jaff in #13108
- [LLM translation] Add support for bedrock computer use by @jugaldb in #12948
- After selecting date range show loader on usage cost charts by @NANDINI-star in #13113
- Revert "[LLM translation] Add support for bedrock computer use" by @krrishdholakia in #13118
- fix(model_checks.py): handle custom values in wildcard model name (e.g. genai/test/*) by @krrishdholakia in #13116
- move to use_prisma_migrate by default + resolve team-only models on auth checks + UI - add sagemaker on UI by @krrishdholakia in #13117
- fix tool aws bedrock call index when the function only have optional arg by @jnhyperion in #13115
- [MCP Gateway] add health check endpoints for MCP by @jugaldb in #13106
- [MCP Protocol header] fix issue with clients protocol header by @jugaldb in #13112
- Added Voyage, Jinai, Deepinfra and VolcEngine providers on the UI by @NANDINI-star in #13131
- [Proxy UI] fix object permission for orgs by @jugaldb in #13142
- New Advanced Date Range Picker Component by @NANDINI-star in #13141
- [Feat] UI + Backend add a tab for use agent activity by @ishaan-jaff in #13146
- [LLM translation] Fix bedrock computer use by @jugaldb in #13143
- [MCP Guardrails] move pre and during hooks to ProxyLoggin by @jugaldb in #13109
- [Feat] v2 updates - tracking DAU, WAU, MAU for coding tool usage + show Daily Usage per User by @ishaan-jaff in #13147
- Litellm explore postgres db ci cd by @krrishdholakia in #13156
- [MCP Gateway] Litellm mcp client list fail by @jugaldb in #13114
- [MCP Gateway] fix migrations by @jugaldb in #13157
- build(config.yml): migrate build_and_test to ci/cd pg db by @krrishdholakia in #13166
- add framework name to UserAgent header in AWS Bedrock API call by @0x-fang in #13159
- fix: remove obsolete attribute
version
in docker compose by @kowyo in #13172
New Contributors
- @5731la made their first contribution in #12989
- @restato made their first contribution in #12980
- @strickvl made their first contribution in #12956
- @Ne0-1 made their first contribution in #12995
- @maxrabin made their first contribution in #13079
- @lvuna made their first contribution in #12894
- @Maximgitman made their first contribution in #12666
- @pathikrit made their first contribution in #12901
- @huetterma made their first contribution in #12809
- @betterthanbreakfast made their first contribution in #13029
- @phosae made their first contribution in #12606
- @sahusiddharth made their first contribution in #12507
- @Amit-kr26 made their first contribution in #11954
- @kowyo made their first contribution in #13172
Full Changelog: v1.74.9.rc.1...v1.74.12-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.12-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 120.0 | 157.82532362996275 | 6.312836133958831 | 0.0 | 1889 | 0 | 83.03327700002683 | 1049.7224119999373 |
Aggregated | Passed ✅ | 120.0 | 157.82532362996275 | 6.312836133958831 | 0.0 | 1889 | 0 | 83.03327700002683 | 1049.7224119999373 |
v1.74.9.rc.1
What's Changed
- Litellm release notes 07 27 2025 p1 by @krrishdholakia in #13027
- VertexAI - camelcase optional params for image generation + Anthropic - streaming, always ensure assistant role set on only first chunk by @krrishdholakia in #12889
- Bulk User Edit - additional improvements - edit all users + set 'no-default-models' on all users by @krrishdholakia in #12925
- add X-Initiator header for GitHub Copilot to reduce premium requests by @ckoehler in #13016
- docs - openwebui show how to include reasoning content for gemini models by @ishaan-jaff in #13060
New Contributors
Full Changelog: v1.74.9.rc-draft...v1.74.9.rc.1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.9.rc.1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.9.rc.1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 130.0 | 168.0418430074381 | 6.293403780045943 | 0.0 | 1882 | 0 | 97.96855799999094 | 1049.3644050000057 |
Aggregated | Passed ✅ | 130.0 | 168.0418430074381 | 6.293403780045943 | 0.0 | 1882 | 0 | 97.96855799999094 | 1049.3644050000057 |
v1.74.9-stable
What's Changed
- Litellm release notes 07 27 2025 p1 by @krrishdholakia in #13027
- VertexAI - camelcase optional params for image generation + Anthropic - streaming, always ensure assistant role set on only first chunk by @krrishdholakia in #12889
- Bulk User Edit - additional improvements - edit all users + set 'no-default-models' on all users by @krrishdholakia in #12925
- add X-Initiator header for GitHub Copilot to reduce premium requests by @ckoehler in #13016
- docs - openwebui show how to include reasoning content for gemini models by @ishaan-jaff in #13060
New Contributors
Full Changelog: v1.74.9.rc-draft...v1.74.9-stable
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.74.9-stable
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 110.0 | 151.02663461606545 | 6.447232352830848 | 0.0 | 1930 | 0 | 81.53728299998875 | 1408.0881720000207 |
Aggregated | Passed ✅ | 110.0 | 151.02663461606545 | 6.447232352830848 | 0.0 | 1930 | 0 | 81.53728299998875 | 1408.0881720000207 |
litellm_v1.65.4-dev_fix
Full Changelog: v1.65.4-stable...litellm_v1.65.4-dev_fix
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.65.4-dev_fix
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 150.0 | 182.3086370464627 | 6.335489722473191 | 0.0 | 1894 | 0 | 112.72804899999755 | 1087.5013490000072 |
Aggregated | Passed ✅ | 150.0 | 182.3086370464627 | 6.335489722473191 | 0.0 | 1894 | 0 | 112.72804899999755 | 1087.5013490000072 |