2.2.0+ regresses SDPA performance on Windows #125070
Labels
module: multi-headed-attention
module: windows
Windows support for PyTorch
triaged
This issue has been looked at a team member, and triaged and prioritized into an appropriate module
馃悰 Describe the bug
In PyTorch <2.2.0, SDPA (scaled_dot_product_attention) supports Flash Attention v1 on Windows. In PyTorch 2.2.0>=, it does not support any Flash Attention on Windows.
Versions
This is a report of a regression between 2.1.2 and 2.2.0+
cc @peterjc123 @mszhanyi @skyline75489 @nbcsm @vladimir-aubrecht @iremyux @Blackhex @cristianPanaite
The text was updated successfully, but these errors were encountered: