Skip to content

Flash Attention v2 achieved 44% faster than xformers/pytorch_sdp_attention on large image #11884

WuSiYu started this conversation in Optimization
Discussion options

You must be logged in to vote

Replies: 6 comments 13 replies

Comment options

You must be logged in to vote
5 replies
@AshtakaOOf
Comment options

@cheald
Comment options

@WuSiYu
Comment options

@Enferlain
Comment options

@cavit99
Comment options

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
1 reply
@AshtakaOOf
Comment options

Comment options

You must be logged in to vote
1 reply
@YUANMU227
Comment options

Comment options

You must be logged in to vote
2 replies
@WuSiYu
Comment options

@slayermaster
Comment options

Comment options

You must be logged in to vote
4 replies
@WuSiYu
Comment options

@MadProbe
Comment options

@AshtakaOOf
Comment options

@MadProbe
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
10 participants