Skip to content

Releases: BerriAI/litellm

v1.75.7-nightly

15 Aug 06:02
Compare
Choose a tag to compare

What's Changed

  • [Proxy] LiteLLM mock test fix by @jugaldb in #13635
  • [Proxy] Litellm add DB metrics to prometheus by @jugaldb in #13626
  • [LLM Translation] Fix Realtime API endpoint for no intent by @jugaldb in #13476
  • [MCP Gateway] LiteLLM Fix MCP gateway key auth by @jugaldb in #13630
  • [Fix] Ensure /messages works when using `bedrock/converse/ with LiteLLM by @ishaan-jaff in #13627
  • UI - Fix image overflow in LiteLLM model by @ishaan-jaff in #13639
  • [Bug Fix] /messages endpoint - ensure tool use arguments are returned for non-anthropic models by @ishaan-jaff in #13638

Full Changelog: v1.75.6-nightly...v1.75.7-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.7-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 140.0 182.49004836898376 6.247441314306712 0.0 1870 0 114.26430999995318 2038.3160259999613
Aggregated Passed ✅ 140.0 182.49004836898376 6.247441314306712 0.0 1870 0 114.26430999995318 2038.3160259999613

v1.75.6-nightly

14 Aug 23:48
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.75.5.rc.1...v1.75.6-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.6-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 110.0 152.45377470174589 6.50695562099742 0.0 1948 0 86.13810599996441 2202.5806519999946
Aggregated Passed ✅ 110.0 152.45377470174589 6.50695562099742 0.0 1948 0 86.13810599996441 2202.5806519999946

litellm_v1.75.5-dev_memory_fix_2

13 Aug 00:59
Compare
Choose a tag to compare

Full Changelog: litellm_v1.75.5-dev_memory_fix...litellm_v1.75.5-dev_memory_fix_2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.75.5-dev_memory_fix_2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 110.0 148.3727584527906 6.403502625001942 0.0 1917 0 82.2168479999732 1009.9184229999878
Aggregated Passed ✅ 110.0 148.3727584527906 6.403502625001942 0.0 1917 0 82.2168479999732 1009.9184229999878

litellm_v1.73.0-dev_memory_fix_2

14 Aug 00:09
Compare
Choose a tag to compare

Full Changelog: litellm_v1.73.0-dev_memory_fix_1...litellm_v1.73.0-dev_memory_fix_2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.73.0-dev_memory_fix_2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 160.0 197.89495319271438 6.333208452490925 0.0 1894 0 120.76194100001203 1698.8860899999736
Aggregated Passed ✅ 160.0 197.89495319271438 6.333208452490925 0.0 1894 0 120.76194100001203 1698.8860899999736

litellm_v1.73.0-dev_memory_fix_1

13 Aug 22:49
Compare
Choose a tag to compare

Full Changelog: litellm_v1.73.0-dev_memory_fix...litellm_v1.73.0-dev_memory_fix_1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.73.0-dev_memory_fix_1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 130.0 168.52018052761707 6.353625704455274 0.0 1901 0 102.81331199996657 955.3499159999888
Aggregated Passed ✅ 130.0 168.52018052761707 6.353625704455274 0.0 1901 0 102.81331199996657 955.3499159999888

litellm_v1.73.0-dev_memory_fix

13 Aug 05:01
Compare
Choose a tag to compare

Full Changelog: v1.73.0.rc.1...litellm_v1.73.0-dev_memory_fix

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.73.0-dev_memory_fix

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 170.0 201.69670727064485 6.271320946816614 0.0 1877 0 124.88003300001083 1273.2718879999823
Aggregated Passed ✅ 170.0 201.69670727064485 6.271320946816614 0.0 1877 0 124.88003300001083 1273.2718879999823

v1.75.5.dev3

12 Aug 23:07
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.75.5.rc.1...v1.75.5.dev3

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.5.dev3

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 170.0 209.8803997534473 6.313294138003075 0.0 1886 0 126.72262299997783 1268.0979020000223
Aggregated Passed ✅ 170.0 209.8803997534473 6.313294138003075 0.0 1886 0 126.72262299997783 1268.0979020000223

litellm_v1.75.5-dev_memory_fix_1

12 Aug 22:25
Compare
Choose a tag to compare

What's Changed

  • [Bug Fix] - Allow using reasoning_effort for gpt-5 model family and reasoning for Responses API by @ishaan-jaff in #13475
  • [Bug Fix]: Azure OpenAI GPT-5 max_tokens + reasoning param support by @ishaan-jaff in #13510
  • [Draft] [LLM Translation] Add model id check by @jugaldb in #13507

Full Changelog: v1.75.5.rc.1...litellm_v1.75.5-dev_memory_fix_1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.75.5-dev_memory_fix_1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 180.0 211.1796445664915 6.382692480357281 0.0 1910 0 132.33785100001683 1892.3347159999935
Aggregated Passed ✅ 180.0 211.1796445664915 6.382692480357281 0.0 1910 0 132.33785100001683 1892.3347159999935

litellm_v1.75.5-dev_memory_fix

12 Aug 01:51
Compare
Choose a tag to compare

What's Changed

  • [Bug Fix] - Allow using reasoning_effort for gpt-5 model family and reasoning for Responses API by @ishaan-jaff in #13475
  • [Bug Fix]: Azure OpenAI GPT-5 max_tokens + reasoning param support by @ishaan-jaff in #13510
  • [Draft] [LLM Translation] Add model id check by @jugaldb in #13507

Full Changelog: v1.75.5.rc.1...litellm_v1.75.5-dev_memory_fix

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.75.5-dev_memory_fix

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.75.5-dev_memory_fix

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 120.0 153.71541160805808 6.384305831136473 0.0 1911 0 80.34984599999007 1251.9617030000063
Aggregated Passed ✅ 120.0 153.71541160805808 6.384305831136473 0.0 1911 0 80.34984599999007 1251.9617030000063

v1.75.5.rc.1

10 Aug 17:18
0aeb4f1
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.75.5-stable.rc-draft...v1.75.5.rc.1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.5.rc.1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 140.0 177.87900723772194 6.396151577527124 0.0 1914 0 110.98393499997883 1386.0049980000042
Aggregated Passed ✅ 140.0 177.87900723772194 6.396151577527124 0.0 1914 0 110.98393499997883 1386.0049980000042