Name | Modified | Size | Downloads / Week |
---|---|---|---|
Parent folder | |||
load_test.html | 2025-08-20 | 1.7 MB | |
load_test_stats.csv | 2025-08-20 | 524 Bytes | |
README.md | 2025-08-20 | 4.9 kB | |
v1.75.9-nightly source code.tar.gz | 2025-08-20 | 193.7 MB | |
v1.75.9-nightly source code.zip | 2025-08-20 | 195.8 MB | |
Totals: 5 Items | 391.2 MB | 0 |
What's Changed
- Litellm stable release fixes by @krrishdholakia in https://github.com/BerriAI/litellm/pull/13682
- [UI QA] Aug 16th Fixes by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/13684
- Litellm dev 08 16 2025 p3 by @krrishdholakia in https://github.com/BerriAI/litellm/pull/13694
- GPT-5-chat does not support function by @superpoussin22 in https://github.com/BerriAI/litellm/pull/13612
- fix(vertexai-batch): fix vertexai batch file format by @thiagosalvatore in https://github.com/BerriAI/litellm/pull/13576
- [Feat] Datadog LLM Observability - Add support for Failure Logging by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/13726
- [Feat] DD LLM Observability - Add time to first token, litellm overhead, guardrail overhead latency metrics by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/13734
- [Bug Fix] litellm incompatible with newest release of openAI v1.100.0 by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/13728
- [Bug Fix] image_edit() function returns APIConnectionError with
litellm_proxy
- Support for both image edits and image generations by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/13735 - [Fix] Cooldowns - don't return raw Azure Exceptions to client by @krrishdholakia in https://github.com/BerriAI/litellm/pull/13529
- Responses API - add default api version for openai responses api calls + Openrouter - fix claude-sonnet-4 on openrouter + Azure - Handle
openai/v1/responses
by @krrishdholakia in https://github.com/BerriAI/litellm/pull/13526 - Use namespace as prefix for s3 cache by @michal-otmianowski in https://github.com/BerriAI/litellm/pull/13704
- Add Search Functionality for Public Model Names in Model Dashboard by @NANDINI-star in https://github.com/BerriAI/litellm/pull/13687
- Add Azure Deployment Name Support in UI by @NANDINI-star in https://github.com/BerriAI/litellm/pull/13685
- Fix - gemini prompt caching cost calculation by @krrishdholakia in https://github.com/BerriAI/litellm/pull/13742
- Refactor - forward model group headers - reuse same logic as global header forwarding by @krrishdholakia in https://github.com/BerriAI/litellm/pull/13741
- Fix Groq streaming ASCII encoding issue by @colesmcintosh in https://github.com/BerriAI/litellm/pull/13675
- Add possibility to configure resources for migrations-job in Helm chart by @moandersson in https://github.com/BerriAI/litellm/pull/13617
- [Feat] Datadog LLM Observability - Add support for tracing guardrail input/output by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/13767
- Models page row UI restructure by @NANDINI-star in https://github.com/BerriAI/litellm/pull/13771
- [Bug Fix] Bedrock KB - Using LiteLLM Managed Credentials for Query by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/13787
- [Bug Fix] Fixes for using Auto Router with LiteLLM Docker Image by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/13788
- [Feat] - UI Allow using Key/Team Based Logging for Langfuse OTEL by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/13791
- Add long context support for claude-4-sonnet by @kankute-sameer in https://github.com/BerriAI/litellm/pull/13759
- Migrate to aim new firewall api by @hxdror in https://github.com/BerriAI/litellm/pull/13748
- [LLM Translation] Adjust max_input_tokens for azure/gpt-5-chat models in JSON configuration by @jugaldb in https://github.com/BerriAI/litellm/pull/13660
- Added Qwen3, Deepseek R1 0528 Throughput, GLM 4.5 and GPT-OSS models for Together AI by @Tasmay-Tibrewal in https://github.com/BerriAI/litellm/pull/13637
- Fix query passthrough deletion by @NANDINI-star in https://github.com/BerriAI/litellm/pull/13622
New Contributors
- @michal-otmianowski made their first contribution in https://github.com/BerriAI/litellm/pull/13704
- @moandersson made their first contribution in https://github.com/BerriAI/litellm/pull/13617
- @Tasmay-Tibrewal made their first contribution in https://github.com/BerriAI/litellm/pull/13637
Full Changelog: https://github.com/BerriAI/litellm/compare/v1.75.8-nightly...v1.75.9-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.9-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 98 | 143.26229013447838 | 6.435162941463474 | 0.0 | 1926 | 0 | 70.0977230000035 | 1988.133740999956 |
Aggregated | Passed ✅ | 98 | 143.26229013447838 | 6.435162941463474 | 0.0 | 1926 | 0 | 70.0977230000035 | 1988.133740999956 |