v1.75.5.dev3
·
0 commits
to 828e3e3deb8eb8120172022f8311d2c464a392eb
since this release
What's Changed
- [Bug Fix] - Allow using
reasoning_effort
for gpt-5 model family andreasoning
for Responses API by @ishaan-jaff in #13475 - [Bug Fix]: Azure OpenAI GPT-5 max_tokens +
reasoning
param support by @ishaan-jaff in #13510 - [Draft] [LLM Translation] Add model id check by @jugaldb in #13507
- [Docs] - Document how to sending tags with Litellm Python SDK Calls to LiteLM Proxy by @ishaan-jaff in #13517
- Fix OCI streaming by @breno-aumo in #13437
- feat: add CometAPI provider support with chat completions and streaming by @TensorNull in #13458
- Allow unsetting TPM and RPM - Teams Settings by @NANDINI-star in #13430
- [Feat] - Add key/team logging for Langfuse OTEL Logger by @ishaan-jaff in #13512
- [Feat] Add Streaming support + Docs for bedrock gpt-oss model family by @ishaan-jaff in #13346
New Contributors
- @TensorNull made their first contribution in #13458
Full Changelog: v1.75.5.rc.1...v1.75.5.dev3
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.5.dev3
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 170.0 | 209.8803997534473 | 6.313294138003075 | 0.0 | 1886 | 0 | 126.72262299997783 | 1268.0979020000223 |
Aggregated | Passed ✅ | 170.0 | 209.8803997534473 | 6.313294138003075 | 0.0 | 1886 | 0 | 126.72262299997783 | 1268.0979020000223 |