Name | Modified | Size | Downloads / Week |
---|---|---|---|
Parent folder | |||
load_test.html | 2025-07-24 | 1.7 MB | |
load_test_stats.csv | 2025-07-24 | 534 Bytes | |
README.md | 2025-07-23 | 7.1 kB | |
v1.74.8-nightly source code.tar.gz | 2025-07-23 | 188.0 MB | |
v1.74.8-nightly source code.zip | 2025-07-23 | 190.0 MB | |
Totals: 5 Items | 379.8 MB | 2 |
What's Changed
- Fix SSO Logout | Create Unified Login Page with SSO and Username/Password Options by @NANDINI-star in https://github.com/BerriAI/litellm/pull/12703
- docs(moonshot): correct base url and document CN-specific endpoint by @magicalne in https://github.com/BerriAI/litellm/pull/12804
- Adding HolmesGPT to projects using LiteLLM by @pavangudiwada in https://github.com/BerriAI/litellm/pull/12798
- feat: Add Lambda AI provider support by @colesmcintosh in https://github.com/BerriAI/litellm/pull/12817
- [UI Bug Fix] Show correct guardrails when editing a team by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12823
- [Docs] Show correct list of vertex ai mistral models by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12828
- [LLM Translation - GH Copilot] added dynamic endpoint support by @jugaldb in https://github.com/BerriAI/litellm/pull/12827
- feat: add Morph provider support by @colesmcintosh in https://github.com/BerriAI/litellm/pull/12821
- fix: remove deprecated groq/qwen-qwq-32b and add qwen/qwen3-32b by @aholmberg in https://github.com/BerriAI/litellm/pull/12831
- Add Google Cloud Model Armor guardrail documentation by @colesmcintosh in https://github.com/BerriAI/litellm/pull/12814
- [Feat] Add Recraft Image Generation API Support - New LLM Provider by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12832
- [Bug Fix] - gemini leaking FD for sync calls with litellm.completion by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12824
- [Feat] Add fireworks -
fireworks/models/kimi-k2-instruct
by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12837 - [LLM Translation] add qwen-vl-plus by @jugaldb in https://github.com/BerriAI/litellm/pull/12829
- [Azure OpenAI Feature] - Support DefaultAzureCredential without hard-coded environment variables by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12841
- Docs - litellm benchmarks by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12842
- fix(watsonx): IBM Watsonx - use correct parameter name for tool choice by @Tomas2D in https://github.com/BerriAI/litellm/pull/9980
- feat: add Hyperbolic provider support by @colesmcintosh in https://github.com/BerriAI/litellm/pull/12826
- Litellm batch cost tracking debug by @krrishdholakia in https://github.com/BerriAI/litellm/pull/12782
- Passthrough Auth - make Auth checks OSS + Anthropic - only show 'reasoning_effort' for supported models by @krrishdholakia in https://github.com/BerriAI/litellm/pull/12847
- build(deps): bump form-data from 4.0.0 to 4.0.4 in /ui/litellm-dashboard by @dependabot[bot] in https://github.com/BerriAI/litellm/pull/12851
- Fix team_member_budget update logic by @krrishdholakia in https://github.com/BerriAI/litellm/pull/12843
- Fix async retryer on
.acompletion()
+ forward clientside headers - filter outcontent-type
from clientside request (causes llm api call to hang) by @krrishdholakia in https://github.com/BerriAI/litellm/pull/12848 - Openrouter - filter out cache_control flag for non-anthropic models (allows usage with claude code) by @krrishdholakia in https://github.com/BerriAI/litellm/pull/12850
- Improvements on the Regenerate Key Flow by @NANDINI-star in https://github.com/BerriAI/litellm/pull/12788
- Replace non-root Dockerfile base with Alpine multi-stage build; by @mdiloreto in https://github.com/BerriAI/litellm/pull/12707
- build(deps): bump form-data from 4.0.3 to 4.0.4 in /docs/my-website by @dependabot[bot] in https://github.com/BerriAI/litellm/pull/12867
- Bug fix - Azure KeyVault not in image, add
azure-keyvault==4.2.0
to Docker img by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12873 - [Separate Health App] Pass through cmd args via supervisord by @jugaldb in https://github.com/BerriAI/litellm/pull/12871
- [Feat] Add Recraft API - Image Edits Support by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12874
- [LLM Translation] Litellm gemini 2.0 live support by @jugaldb in https://github.com/BerriAI/litellm/pull/12839
- [Feat] Add cost tracking for new vertex_ai/llama-3 API models by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12878
- [Feat] - Track cost + add tags for health checks done by LiteLLM Proxy by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12880
- Fix async retryer on .acompletion() + forward clientside headers - filter out content-type from clientside request (causes llm api call to hang) by @krrishdholakia in https://github.com/BerriAI/litellm/pull/12886
- Request Headers - support
x-litellm-num-retries
+ Usage - support usage by model group by @krrishdholakia in https://github.com/BerriAI/litellm/pull/12890 - [Docs ]Litellm mcp access group doc by @jugaldb in https://github.com/BerriAI/litellm/pull/12883
- [Add health check] add architecture diagram by @jugaldb in https://github.com/BerriAI/litellm/pull/12879
- rm retired anthropic models from model_prices_and_context_window.json by @staeiou in https://github.com/BerriAI/litellm/pull/12864
- feat: extended
/v1/models
endpoint, now it returns with fallbacks on demand by @murad-khafizov in https://github.com/BerriAI/litellm/pull/12811 - feat: Add Pillar Security guardrail integration by @eagle-p in https://github.com/BerriAI/litellm/pull/12791
- Swagger - bump DOMPurify version + add back Support local swagger bundle (good for airgapped environments) + support local swagger on custom root paths by @krrishdholakia in https://github.com/BerriAI/litellm/pull/12911
- [Feat] LiteLLM x Cloudzero integration - Allow exporting spend to cloudzero by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/12908
New Contributors
- @magicalne made their first contribution in https://github.com/BerriAI/litellm/pull/12804
- @pavangudiwada made their first contribution in https://github.com/BerriAI/litellm/pull/12798
- @mdiloreto made their first contribution in https://github.com/BerriAI/litellm/pull/12707
- @murad-khafizov made their first contribution in https://github.com/BerriAI/litellm/pull/12811
- @eagle-p made their first contribution in https://github.com/BerriAI/litellm/pull/12791
Full Changelog: https://github.com/BerriAI/litellm/compare/v1.74.7.rc.1...v1.74.8-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.8-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 170.0 | 208.0958998087373 | 6.271169991247407 | 0.0 | 1877 | 0 | 130.56155500004252 | 1168.738483000027 |
Aggregated | Passed ✅ | 170.0 | 208.0958998087373 | 6.271169991247407 | 0.0 | 1877 | 0 | 130.56155500004252 | 1168.738483000027 |