From 56940ddc81a89f218a2ae5dd777c9ce2a3bf6686 Mon Sep 17 00:00:00 2001 From: hiworldwzj <30762946+hiworldwzj@users.noreply.github.com> Date: Tue, 19 Mar 2024 14:51:16 +0800 Subject: [PATCH] Update context_flashattention_nopad.py To fix triton 2.0.0 import error. (#364) --- .../models/llama/triton_kernel/context_flashattention_nopad.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/lightllm/models/llama/triton_kernel/context_flashattention_nopad.py b/lightllm/models/llama/triton_kernel/context_flashattention_nopad.py index 8086cd45..5493c0b9 100644 --- a/lightllm/models/llama/triton_kernel/context_flashattention_nopad.py +++ b/lightllm/models/llama/triton_kernel/context_flashattention_nopad.py @@ -506,6 +506,8 @@ def context_attention_fwd(q, k, v, o, b_start_loc, b_seq_len, max_input_len): num_stages=1, ) return + + context_attention_fwd_no_prompt_cache = None else: raise Exception("error triton version!")