v1.54.0
版本发布时间: 2024-12-08 12:50:00
BerriAI/litellm最新发布版本:v1.54.0(2024-12-08 12:50:00)
What's Changed
- (feat) Track
custom_llm_provider
in LiteLLMSpendLogs by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/7081 - Add MLflow to the side bar by @B-Step62 in https://github.com/BerriAI/litellm/pull/7031
- (bug fix) SpendLogs update DB catch all possible DB errors for retrying by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/7082
- (Feat) Add StructuredOutputs support for Fireworks.AI by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/7085
- added deepinfra/Meta-Llama-3.1-405B-Instruct to the Model json by @AliSayyah in https://github.com/BerriAI/litellm/pull/7084
- (feat) Add created_at and updated_at for LiteLLM_UserTable by @ishaan-jaff in https://github.com/BerriAI/litellm/pull/7089
New Contributors
- @AliSayyah made their first contribution in https://github.com/BerriAI/litellm/pull/7084
Full Changelog: https://github.com/BerriAI/litellm/compare/v1.53.9...v1.54.0
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.54.0
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 200.0 | 220.2003271503722 | 6.29832230581454 | 0.0 | 1882 | 0 | 179.34225999999853 | 1827.969679000006 |
Aggregated | Passed ✅ | 200.0 | 220.2003271503722 | 6.29832230581454 | 0.0 | 1882 | 0 | 179.34225999999853 | 1827.969679000006 |
1、 load_test.html 1.59MB
2、 load_test_stats.csv 532B