Skip to content

Commit 4e73bb8

Browse files
mgoincharlifu
authored andcommitted
[Bugfix] Remove contiguous output req for context parallel MLA (vllm-project#25414)
Signed-off-by: Michael Goin <mgoin64@gmail.com> Signed-off-by: charlifu <charlifu@amd.com>
1 parent 27ac795 commit 4e73bb8

File tree

1 file changed

+0
-1
lines changed

1 file changed

+0
-1
lines changed

vllm/attention/ops/common.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -134,6 +134,5 @@ def cp_lse_ag_out_rs(cp_attn_out: torch.Tensor,
134134
cp_attn_lse = cp_attn_lse.contiguous()
135135
lses = cp_group.all_gather(cp_attn_lse, dim=0).view_as(lses)
136136
out, _ = correct_attn_out(cp_attn_out, lses, cp_group.rank_in_group, ctx)
137-
assert out.is_contiguous()
138137
out = cp_group.reduce_scatter(out, dim=1)
139138
return out

0 commit comments

Comments
 (0)