From c577d03b6e56f2ff9192fd31c8222f7efc0c2cab Mon Sep 17 00:00:00 2001 From: notlober Date: Thu, 6 Mar 2025 01:39:13 +0300 Subject: [PATCH] Update flex_attention.py --- unsloth/kernels/flex_attention.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/unsloth/kernels/flex_attention.py b/unsloth/kernels/flex_attention.py index 6f82394228..11a84cdfb8 100644 --- a/unsloth/kernels/flex_attention.py +++ b/unsloth/kernels/flex_attention.py @@ -32,7 +32,7 @@ create_block_mask as _create_block_mask, ) _flex_attention = torch.compile(_flex_attention, dynamic = True, options = torch_compile_options) - HAS_FLEX_ATTENTION = False + HAS_FLEX_ATTENTION = True except: HAS_FLEX_ATTENTION = False pass