Skip to content

Commit 855a2a6

Browse files
authored
fix attn_params (#4787)
1 parent 9887025 commit 855a2a6

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

fastdeploy/model_executor/layers/attention/append_attn_backend.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -365,7 +365,7 @@ def forward_mixed(
365365
getattr(layer, "cache_v_zp", None),
366366
layer.linear_shift,
367367
layer.linear_smooth,
368-
None,
368+
forward_meta.attn_mask_offsets,
369369
metadata.kv_signal_data_list[layer.layer_id],
370370
getattr(layer, "q_norm_weight", None),
371371
getattr(layer, "k_norm_weight", None),
@@ -384,7 +384,7 @@ def forward_mixed(
384384
metadata.max_partition_size,
385385
metadata.encoder_max_partition_size,
386386
self.speculate_max_draft_token_num + 1,
387-
True,
387+
self.causal,
388388
self.speculative_method is not None,
389389
sliding_window,
390390
)

0 commit comments

Comments
 (0)