[OpenVINO] Support Qwen3.5#1634
[OpenVINO] Support Qwen3.5#1634rkazants wants to merge 9 commits intohuggingface:transformers-v5from
Conversation
Add conversion rule for the RecurrentAttentionCellOp operation used for GatedDeltaNet patching in OpenVINO PyTorch frontend. Co-authored-by: Copilot <223556219+Copilot@users.noreply.github.com>
Co-authored-by: rkazants <35459624+rkazants@users.noreply.github.com>
Co-authored-by: rkazants <35459624+rkazants@users.noreply.github.com>
Co-authored-by: rkazants <35459624+rkazants@users.noreply.github.com>
|
Thank you!! 🙏 |
Signed-off-by: Kazantsev, Roman <roman.kazantsev@intel.com>
Signed-off-by: Kazantsev, Roman <roman.kazantsev@intel.com>
|
Can we ensure this PR includes a hardware compatibility check for the Core Ultra 200 series (245 through 285) and other Xe platforms? Previous OpenVino MoE optimizations have caused kernel-level failures on these platforms without any documented warnings. We need to verify that this PR either provides full support or—at a minimum—documented and implements a graceful exit/error message rather than a system crash. See this issue:
and related issues: |
|
Testing this branch on Linux (Python 3.12) with an Intel Arc A770. I've hit a roadblock with the qwen3_5 implementation and wanted to check if anyone has successfully completed an INT4 export yet. What I've tried: The Error: It seems the library recognizes the qwen3_5 model type but fails to find the registered class. For those who have this working: what specific versions of transformers and huggingface-hub are you using, and did you have to manually source a specific modeling_qwen3_5.py file? |
Signed-off-by: Kazantsev, Roman <roman.kazantsev@intel.com>
What does this PR do?
Fixes # (issue)
Before submitting