-
Notifications
You must be signed in to change notification settings - Fork 117
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support weight-only quantization with quantized operators in intel-extension-for-transformers. #455
Support weight-only quantization with quantized operators in intel-extension-for-transformers. #455
Conversation
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. |
fbf9ddf
to
9d03415
Compare
…tension-for-transformers
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
…tension-for-transformers
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
ecaac6e
to
ed873c9
Compare
Hi, @echarlaix , I rebased code with main branch. Please review it, thanks very much! |
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
bee33e1
to
de190fd
Compare
Hi, @echarlaix , it seems to build intel-extension-for-transformers failed in the pre-CI test. Before executing the "python setup.py install" command, we should pip install dependency packages with "pip install -r requirements.txt" in intel-extension-for-transformers. Could you review it? thanks! |
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks a lot for your work @PenghuiCheng
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
* [OV]: Fixed inferece after 4 bit weight compression * Fixed issue * Update optimum/intel/openvino/modeling_decoder.py Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com> * Applied comments * Fixed issue when request is None --------- Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
* Updated docs with load_in_4bit * Update documentation * Update documentation * typo --------- Co-authored-by: Ella Charlaix <ella@huggingface.co>
* fix compatibility for latest transformers release * update setup * update setup * fix test input size * fix prepare generation for llama models
* deprecate compression options * style * fix configuration * Update CLI argument * update documentation * deprecate torch nn modules for ov quantizer * fix ov config for fp32 models * fix format * update documentation * Add check for configuration * fix ratio default value for SD models * add quantization_config argument for OVModel * remove commented line * Update docs/source/inference.mdx Co-authored-by: Alexander Kozlov <alexander.kozlov@intel.com> * add default config for causal LM * fix warning message --------- Co-authored-by: Alexander Kozlov <alexander.kozlov@intel.com>
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
2a905cc
to
5ddd360
Compare
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks great thanks @PenghuiCheng
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
What does this PR do?
Intel-extension-for-transformers package implements weight-only quantization operators with jblas kernel. So we integrated weight-only quantization with intel-extension-for-transformers in this PR.
Before submitting