We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Per: Dao-AILab/flash-attention#203
if you're using pytorch 2.0 then FlashAttention is already available through torch.nn.functional.scaled_dot_product_attention.
The flash-attn project has build problems for many people. Is it possible to consider using PyTorch 2.0 equivalent Flash Attention?
The text was updated successfully, but these errors were encountered:
For MPT we need to be able to use causal=True and we'd need to use attn_mask (aka attn_bias) to have ALiBi.
The variant exposed in scaled_dot_product_attention docs does not allow both. From docs:
Sorry, something went wrong.
So this issues needs to propagate to PyTorch. Until then, I'll work with Flash Attn then as is. Thanks.
Update README.md (#103)
59891cb
Single-line typo fix
vchiley
No branches or pull requests
Per: Dao-AILab/flash-attention#203
The flash-attn project has build problems for many people.
Is it possible to consider using PyTorch 2.0 equivalent Flash Attention?
The text was updated successfully, but these errors were encountered: