-
Notifications
You must be signed in to change notification settings - Fork 81
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ImportError: This modeling file requires... flash_attn
#8
Comments
From the error message, it looks like the CUDA drivers are not installed. Also, we already list |
is the section I read? If |
(an m1 mac has no nvidia card so I don't think I can install nvcc? Too bad, but I get that some stuff can't run without an nvidia card) |
now I see that you listed it in the model description, but it appears to be necessary for inference as well, so it should be included in that list of required python packages is what I mean |
You don't need flash attention for inference -- it's a "nice to have" that makes inference faster, but to my knowledge it works only on NVIDIA GPUs (as you need CUDA).
Hope this helps. Also, make sure to run on the latest version of the Transformers library! |
That's exactly what I did that caused the error to occur! |
Can you run |
I will do so tomorrow (I have to re-download the model now), but I was working in a clean virtualenv |
(which I assume means pip will download the newest version of a lib? But maybe that assumption is false if there's a previously cached version?) |
I'm unable to reproduce. Sincere apologies for the noise and wasting your time, and thanks for the model |
No problem! Glad it worked in the end :) |
I have mac m2 max 32 GB. |
for me it doesn't work with |
Yes, it worked. Thank you. :) |
Trying to follow the instructions on an m1 mac, I get the above error.
Unfortunately, attempting to install
flash_attn
does not succeed, due to:RuntimeError: flash_attn was requested, but nvcc was not found.
, which may be just an unfortunate aspect of not having an nvidia card.Anyway, the point is probably you should add
flash_attn
to your list of required modules?The text was updated successfully, but these errors were encountered: