v1.2.1
版本发布时间: 2023-11-08 17:14:45
intel/intel-extension-for-transformers最新发布版本:v1.4.2(2024-05-24 20:23:38)
- Examples
- Bug Fixing & Improvements
Examples
- Add docker for code-generation (dd3829 )
- Enable Qwen-7B-Chat for NeuralChat (698e58 )
- Enable Baichuan & Baichuan2 CPP inference (98e5f9 )
- Add sidebyside UI for NeuralChat (dbbcc2 )
- Support Falcon-180B CPP inference (900ebf )
- Support starcoder finetuning example (073bdd )
- Enable text-generation using qwen (8f41d4 )
- Add docker for neuralchat (a17d952 )
Bug Fixing & Improvements
- Fix bug for woq with AWQ due to not set calib_iters if calib_dataloader is not None.( 565ab4)
- Fix init issue of langchain chroma (fdefe2)
- Fix NeuralChat starcoder mha fusion issue (ce3d24)
- Fix setuptools version limitation for build (2cae32)
- Fix post process with topk topp of python api (7b4730)
- Fix msvc compile issues (87b00d)
- Refine notebook and fix restful api issues (d8cc11)
- Upgrade qbits backend (45e03b )
- Fix starcoder issues for IPEX int8 and Weight Only int4 (e88c7b )
- Fix ChatGLM2 model loading issue (4f2169 )
- Remove OneDNN graph env setting for BF16 inference (59ab03 )
- Improve database by escape sql string (be6790 )
- fix qbits backend get wrong workspace malloc size (6dbd0b )
Validated Configurations
- Python 3.9, 3.10
- Centos 8.4 & Ubuntu 22.04
- Intel® Extension for TensorFlow 2.13.0
- PyTorch 2.1.0+cpu
- Intel® Extension for PyTorch 2.1.0+cpu
- Transformers 4.34.1