-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add Qwen2.5, Gemma-1, Phi-2 #369
base: main
Are you sure you want to change the base?
Conversation
fyi @uaydonat |
For Qwen2 /2.5 test, it returns many warnings of this type. I also attach full log. [2025-03-18 21:20:13,936] torch._dynamo.convert_frame: [WARNING] torch._dynamo.exc.BackendCompilerFailed: backend=‘ttnn_backend’ raised: To reproduce:
---------------------------- Captured log teardown ----------------------------- |
After more detailed investigation - it seems that for both Qwen2 and Qwen2.5 there is some problem with torch.compile itself, as even in case when TTNN backend is not used for compilation, the same error occurs. Looks like torch.compile does not behave properly with either the qwen model or newer versions of transformers. transformers>=4.45.0 . |
For this particular error concerning
You can try to investigate why the newer implementation is causing problems with |
I'm adding support for a few more language models which are high priority for the models team.
Note that the Gemma model requires a huggingface-cli login. In addition, Qwen2-VL is marked to be skipped since the transformers version is < 4.45.
@kevinwuTT could you help me understand why the Qwen and Gemma tests fail like this? It doesn't look like these tests are any different from Phi-2 and GPT-Neo, which pass.