Skip to content

Intel® Extension for Transformers v1.2.1 Release

Compare
Choose a tag to compare
@kevinintel kevinintel released this 08 Nov 09:14
· 933 commits to main since this release
  • Examples
  • Bug Fixing & Improvements

Examples

  • Add docker for code-generation (dd3829 )
  • Enable Qwen-7B-Chat for NeuralChat (698e58 )
  • Enable Baichuan & Baichuan2 CPP inference (98e5f9 )
  • Add sidebyside UI for NeuralChat (dbbcc2 )
  • Support Falcon-180B CPP inference (900ebf )
  • Support starcoder finetuning example (073bdd )
  • Enable text-generation using qwen (8f41d4 )
  • Add docker for neuralchat (a17d952 )

Bug Fixing & Improvements

  • Fix bug for woq with AWQ due to not set calib_iters if calib_dataloader is not None.( 565ab4)
  • Fix init issue of langchain chroma (fdefe2)
  • Fix NeuralChat starcoder mha fusion issue (ce3d24)
  • Fix setuptools version limitation for build (2cae32)
  • Fix post process with topk topp of python api (7b4730)
  • Fix msvc compile issues (87b00d)
  • Refine notebook and fix restful api issues (d8cc11)
  • Upgrade qbits backend (45e03b )
  • Fix starcoder issues for IPEX int8 and Weight Only int4 (e88c7b )
  • Fix ChatGLM2 model loading issue (4f2169 )
  • Remove OneDNN graph env setting for BF16 inference (59ab03 )
  • Improve database by escape sql string (be6790 )
  • fix qbits backend get wrong workspace malloc size (6dbd0b )

Validated Configurations

  • Python 3.9, 3.10
  • Centos 8.4 & Ubuntu 22.04
  • Intel® Extension for TensorFlow 2.13.0
  • PyTorch 2.1.0+cpu
  • Intel® Extension for PyTorch 2.1.0+cpu
  • Transformers 4.34.1