v0.4.1
版本发布时间: 2024-05-07 16:20:47
InternLM/lmdeploy最新发布版本:v0.6.0a0(2024-08-26 17:12:19)
What's Changed
🚀 Features
- Add colab demo by @AllentDan in https://github.com/InternLM/lmdeploy/pull/1428
- support starcoder2 by @grimoire in https://github.com/InternLM/lmdeploy/pull/1468
- support OpenGVLab/InternVL-Chat-V1-5 by @irexyc in https://github.com/InternLM/lmdeploy/pull/1490
💥 Improvements
- variable
CTA_H
& fix qkv bias by @lzhangzz in https://github.com/InternLM/lmdeploy/pull/1491 - refactor vision model loading by @irexyc in https://github.com/InternLM/lmdeploy/pull/1482
- fix installation requirements for windows by @irexyc in https://github.com/InternLM/lmdeploy/pull/1531
- Remove split batch inside pipline inference function by @AllentDan in https://github.com/InternLM/lmdeploy/pull/1507
- Remove first empty chunck for api_server by @AllentDan in https://github.com/InternLM/lmdeploy/pull/1527
- add benchmark script to profile pipeline APIs by @lvhan028 in https://github.com/InternLM/lmdeploy/pull/1528
- Add input validation by @AllentDan in https://github.com/InternLM/lmdeploy/pull/1525
🐞 Bug fixes
- fix local variable 'response' referenced before assignment in async_engine.generate by @irexyc in https://github.com/InternLM/lmdeploy/pull/1513
- Fix turbomind import in windows by @irexyc in https://github.com/InternLM/lmdeploy/pull/1533
- Fix convert qwen2 to turbomind by @AllentDan in https://github.com/InternLM/lmdeploy/pull/1546
- Adding api_key and model_name parameters to the restful benchmark by @NiuBlibing in https://github.com/InternLM/lmdeploy/pull/1478
📚 Documentations
- update supported models for Baichuan by @zhyncs in https://github.com/InternLM/lmdeploy/pull/1485
- Fix typo in w8a8.md by @Infinity4B in https://github.com/InternLM/lmdeploy/pull/1523
- complete build.md by @YanxingLiu in https://github.com/InternLM/lmdeploy/pull/1508
- update readme wechat qrcode by @vansin in https://github.com/InternLM/lmdeploy/pull/1529
- Update docker docs for VL api by @vody-am in https://github.com/InternLM/lmdeploy/pull/1534
- Format supported model table using html syntax by @lvhan028 in https://github.com/InternLM/lmdeploy/pull/1493
- doc: add example of deploying api server to Kubernetes by @uzuku in https://github.com/InternLM/lmdeploy/pull/1488
🌐 Other
- add modelscope and lora testcase by @zhulinJulia24 in https://github.com/InternLM/lmdeploy/pull/1506
- bump version to v0.4.1 by @lvhan028 in https://github.com/InternLM/lmdeploy/pull/1544
New Contributors
- @NiuBlibing made their first contribution in https://github.com/InternLM/lmdeploy/pull/1478
- @Infinity4B made their first contribution in https://github.com/InternLM/lmdeploy/pull/1523
- @YanxingLiu made their first contribution in https://github.com/InternLM/lmdeploy/pull/1508
- @vody-am made their first contribution in https://github.com/InternLM/lmdeploy/pull/1534
- @uzuku made their first contribution in https://github.com/InternLM/lmdeploy/pull/1488
Full Changelog: https://github.com/InternLM/lmdeploy/compare/v0.4.0...v0.4.1
1、 lmdeploy-0.4.1+cu118-cp310-cp310-manylinux2014_x86_64.whl 69.78MB
2、 lmdeploy-0.4.1+cu118-cp310-cp310-win_amd64.whl 48.01MB
3、 lmdeploy-0.4.1+cu118-cp311-cp311-manylinux2014_x86_64.whl 69.8MB
4、 lmdeploy-0.4.1+cu118-cp311-cp311-win_amd64.whl 48.01MB
5、 lmdeploy-0.4.1+cu118-cp38-cp38-manylinux2014_x86_64.whl 69.8MB
6、 lmdeploy-0.4.1+cu118-cp38-cp38-win_amd64.whl 48.01MB
7、 lmdeploy-0.4.1+cu118-cp39-cp39-manylinux2014_x86_64.whl 69.79MB