Skip to content

Releases: BerriAI/litellm

v1.55.4

18 Dec 17:30
7a5dd29
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.55.3...v1.55.4

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.55.4

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.55.4

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 277.54264076081057 6.1096008757758025 0.0 1827 0 217.2655560000294 1776.9381030000204
Aggregated Passed ✅ 250.0 277.54264076081057 6.1096008757758025 0.0 1827 0 217.2655560000294 1776.9381030000204

v1.55.3

16 Dec 17:16
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.55.2...v1.55.3

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.55.3

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 281.1265187306242 6.1657490001280255 0.0033418693767631575 1845 1 119.36488499998177 3755.8482019999815
Aggregated Passed ✅ 250.0 281.1265187306242 6.1657490001280255 0.0033418693767631575 1845 1 119.36488499998177 3755.8482019999815

v1.55.1-stable

15 Dec 00:36
Compare
Choose a tag to compare

Full Changelog: v1.55.1...v1.55.1-stable

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_dec_14-stable

litellm-database image

ghcr.io/berriai/litellm-database:litellm_stable_dec_14-stable

litellm-non-root image

ghcr.io/berriai/litellm-non_root:litellm_stable_dec_14-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 200.0 217.72878062246997 6.2754597178458145 0.0033415653449658223 1878 1 76.6410740000083 1257.3869729999956
Aggregated Passed ✅ 200.0 217.72878062246997 6.2754597178458145 0.0033415653449658223 1878 1 76.6410740000083 1257.3869729999956

v1.55.2

13 Dec 22:04
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.55.1...v1.55.2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.55.2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.55.2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.55.2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 282.51255728779716 6.192691226975396 0.0 1852 0 223.9336790000266 3178.0424589999257
Aggregated Passed ✅ 250.0 282.51255728779716 6.192691226975396 0.0 1852 0 223.9336790000266 3178.0424589999257

v1.55.1

13 Dec 17:02
Compare
Choose a tag to compare

What's Changed

  • (feat) add response_time to StandardLoggingPayload - logged on datadog, gcs_bucket, s3_bucket etc by @ishaan-jaff in #7199
  • build(deps): bump nanoid from 3.3.7 to 3.3.8 in /ui by @dependabot in #7198
  • (Feat) DataDog Logger - Add HOSTNAME and POD_NAME to DataDog logs by @ishaan-jaff in #7189
  • (feat) add error_code, error_class, llm_provider to StandardLoggingPayload by @ishaan-jaff in #7200
  • (docs) Document StandardLoggingPayload Spec by @ishaan-jaff in #7201
  • fix: Support WebP image format and avoid token calculation error by @ishaan-jaff in #7182
  • (feat) UI - Disable Usage Tab once SpendLogs is 1M+ Rows by @ishaan-jaff in #7208
  • (minor fix proxy) Clarify Proxy Rate limit errors are showing hash of litellm virtual key by @ishaan-jaff in #7210
  • (fix) latency fix - revert prompt caching check on litellm router by @ishaan-jaff in #7211

Full Changelog: v1.55.0...v1.55.1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.55.1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 274.17864765330575 6.170501674094568 0.0 1846 0 212.15181599995958 2203.3609819999356
Aggregated Passed ✅ 250.0 274.17864765330575 6.170501674094568 0.0 1846 0 212.15181599995958 2203.3609819999356

v1.55.0.dev2

13 Dec 03:00
Compare
Choose a tag to compare

What's Changed

  • (feat) add response_time to StandardLoggingPayload - logged on datadog, gcs_bucket, s3_bucket etc by @ishaan-jaff in #7199
  • build(deps): bump nanoid from 3.3.7 to 3.3.8 in /ui by @dependabot in #7198
  • (Feat) DataDog Logger - Add HOSTNAME and POD_NAME to DataDog logs by @ishaan-jaff in #7189
  • (feat) add error_code, error_class, llm_provider to StandardLoggingPayload by @ishaan-jaff in #7200
  • (docs) Document StandardLoggingPayload Spec by @ishaan-jaff in #7201
  • fix: Support WebP image format and avoid token calculation error by @ishaan-jaff in #7182
  • (feat) UI - Disable Usage Tab once SpendLogs is 1M+ Rows by @ishaan-jaff in #7208

Full Changelog: v1.55.0...v1.55.0.dev2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.55.0.dev2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 210.0 236.69042419128075 6.133942906309277 0.0 1835 0 175.69668400000182 4096.7015589999955
Aggregated Passed ✅ 210.0 236.69042419128075 6.133942906309277 0.0 1835 0 175.69668400000182 4096.7015589999955

v1.55.0.dev1

12 Dec 23:09
Compare
Choose a tag to compare

Full Changelog: v1.55.0...v1.55.0.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.55.0.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 220.0 243.6248328955295 6.208881714875033 0.0 1857 0 195.87834699996165 1936.7717839999727
Aggregated Passed ✅ 220.0 243.6248328955295 6.208881714875033 0.0 1857 0 195.87834699996165 1936.7717839999727

v1.55.0

12 Dec 19:40
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.54.1...v1.55.0

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.55.0

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 286.19507948581224 5.886697197840291 0.0033409178194326278 1762 1 211.68456200001629 3578.4067740000296
Aggregated Passed ✅ 250.0 286.19507948581224 5.886697197840291 0.0033409178194326278 1762 1 211.68456200001629 3578.4067740000296

v1.54.1

09 Dec 17:31
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.54.0...v1.54.1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.54.1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 280.0 340.7890831504466 5.986291177372485 0.0 1788 0 236.28402200000664 4047.592437999981
Aggregated Failed ❌ 280.0 340.7890831504466 5.986291177372485 0.0 1788 0 236.28402200000664 4047.592437999981

v1.54.0

08 Dec 04:50
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.53.9...v1.54.0

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.54.0

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 200.0 220.2003271503722 6.29832230581454 0.0 1882 0 179.34225999999853 1827.969679000006
Aggregated Passed ✅ 200.0 220.2003271503722 6.29832230581454 0.0 1882 0 179.34225999999853 1827.969679000006