Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

skip private mdoel loading test for external contributors #1082

Merged
merged 1 commit into from
Dec 19, 2024

Conversation

echarlaix
Copy link
Collaborator

@echarlaix echarlaix commented Dec 19, 2024

@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

@echarlaix
Copy link
Collaborator Author

cc @helena-intel fixing the issue from #1072 (comment)

@echarlaix echarlaix merged commit 106a5b7 into main Dec 19, 2024
23 of 25 checks passed
@echarlaix echarlaix deleted the external-contrib-private branch December 19, 2024 16:28
AlexKoff88 pushed a commit that referenced this pull request Dec 23, 2024
* Support AWQ models

* Add tests

* Add dependencies

* Fix tests

* enable awq export only if ov support it

* fix style (#2)

* disable awq and gptq install for old torch (#3)

* fix style

* disable autogptq and autoawq install for old transformers testing

* separate common quant models patching and gptq (#4)

* disable windows install (#5)

* separate common quant models patching and gptq

* disable awq windows

* skip logits check for quantized models (#6)

* fix test after rebase

* fix testing condition for 2024.6 and unpatch in case if failed

* Fix qwen2-vl tests (#1084)

* Skip private mdoel loading test for external contributors (#1082)

* Fix reshaping unet if timestep is 0d tensor (#1083)

* Disable kv cache compression for fp vlm (#1080)

* Support AWQ models

* Add tests

* Add dependencies

* Fix tests

* enable awq export only if ov support it

* fix style (#2)

* disable awq and gptq install for old torch (#3)

* fix style

* disable autogptq and autoawq install for old transformers testing

* separate common quant models patching and gptq (#4)

* disable windows install (#5)

* separate common quant models patching and gptq

* disable awq windows

* skip logits check for quantized models (#6)

* fix test after rebase

* fix testing condition for 2024.6 and unpatch in case if failed

* add necessary packages in test_openvino_full

* fix code style after rebase (#7)

---------

Co-authored-by: eaidova <ekaterina.aidova@intel.com>
Co-authored-by: Nikita Savelyev <nikita.savelyev@intel.com>
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants