Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
PR #11444: [XLA:GPU] disable mask in cuDNN attention
Imported from GitHub PR openxla/xla#11444 1. cuDNN attention mask is not doing masking with -inf but multiply which is not correct. Hence disable patterns with mask. 2. Follow up PR to clean up the remaining mask related logic. Copybara import of the project: -- acf95b6cc7e1084026eaf87c0119ba3801ba8f8c by cjkkkk <ske@nvidia.com>: disable mask Merging this change closes #11444 FUTURE_COPYBARA_INTEGRATE_REVIEW=openxla/xla#11444 from Cjkkkk:remove_mask acf95b6cc7e1084026eaf87c0119ba3801ba8f8c PiperOrigin-RevId: 624057479
- Loading branch information