-
Notifications
You must be signed in to change notification settings - Fork 564
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Memory efficient attention not gain speedups on A10 and V100 #762
Comments
Hi, |
|
Oh yes - good catch! We have kernels for f32 but they are not really efficient. You should use f16 or bf16 if possible to get the best speed. In fact, it's very likely that xFormers induces a slow-down when training in f32. |
Okay, Thank you! I'll try using fp16 |
Using diffuers and enable enable_xformers_memory_efficient_attention
But the speed didn't get any changed. Why?
The text was updated successfully, but these errors were encountered: