v1.52.16
版本发布时间: 2024-11-26 18:22:49
BerriAI/litellm最新发布版本:v1.54.0(2024-12-08 12:50:00)
What's Changed
- feat - allow sending
tags
on vertex pass through requests by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/6876 - (feat) Add support for using @google/generative-ai JS with LiteLLM Proxy by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/6899
- (UI fix) UI does not reload when you login / open a new tab by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/6909
- (fix) pass through endpoints - run logging async + use thread pool executor for sync logging callbacks by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/6907
- (redis fix) - fix
AbstractConnection.__init__() got an unexpected keyword argument 'ssl'
by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/6908 - (docs) Simplify
/vertex_ai/
pass through docs by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/6910
Full Changelog: https://github.com/BerriAI/litellm/compare/v1.52.15...v1.52.16
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.16
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 230.0 | 261.12838541230485 | 6.194778256251999 | 0.0 | 1853 | 0 | 206.21302299997524 | 2167.8605710000056 |
Aggregated | Passed ✅ | 230.0 | 261.12838541230485 | 6.194778256251999 | 0.0 | 1853 | 0 | 206.21302299997524 | 2167.8605710000056 |
1、 load_test.html 1.59MB
2、 load_test_stats.csv 540B