-
Notifications
You must be signed in to change notification settings - Fork 157
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[HuggingFace] Compilation of vision models fail for batch_size > 1 #1080
Comments
Thanks for filing the issue. We will take a look and get back to you soon. |
The model can be compiled successfully (for BS>1) by including the Sample code is attached here && can be used to successfully compile the model.
This was tested on artifacts from latest Neuron release container: |
@0x6b64 thank you for looking into this issue. The reason why |
Just to +1, I have another issue with Swin models where for Neuron to compile larger than 32x32 it needs a special compiler arg passed in. (In my case I use
(and apparently the triple quotes are important)) An "append to compiler_args" option would be nice. |
@0x6b64 I think there are two issues, and model-type=transformers only fixes one. The 224 model is a higher resolution and seems to need the flag above. 224 is the one I used on the internal ticket. I tested the code there (that doesn't use optimum-neuron) with model-type=transformers and replicated the OTHER error message.
|
@0x6b64 In Optimum-Neuron, we only allow users to configure the arguments that directly impact model performance. |
Thanks @dacorvo @JingyaHuang for all of the context! To answer your questions
Without revealing too many internals, short answer is that it shouldn't matter. But you're right, a unet based architecture should be using
We should continue to use |
Hi team! We also observed some issues when compiling vision models(swin and donut) with
batch_size > 1
after bumping to Neuron SDK 2.21.0:To reproduce:
(The compilation is successful with
batch_size=1
though)The text was updated successfully, but these errors were encountered: