diff --git a/llama/addon/AddonContext.cpp b/llama/addon/AddonContext.cpp index 2d02326e..d1b1052b 100644 --- a/llama/addon/AddonContext.cpp +++ b/llama/addon/AddonContext.cpp @@ -420,7 +420,8 @@ AddonContext::AddonContext(const Napi::CallbackInfo& info) : Napi::ObjectWrap().Value(); + bool flashAttention = options.Get("flashAttention").As().Value(); + context_params.flash_attn_type = flashAttention ? LLAMA_FLASH_ATTN_TYPE_ENABLED : LLAMA_FLASH_ATTN_TYPE_DISABLED; } if (options.Has("threads")) {