We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent d4840a4 commit 70fd41cCopy full SHA for 70fd41c
lmdeploy/pytorch/kernels/ascend/paged_attention_fwd.py
@@ -33,7 +33,7 @@ def flash_context_attention(
33
key=key_states,
34
value=value_states,
35
q_start_loc=q_start_loc[i:i + 1],
36
- seq_len_list=q_seq_len[i:i + 1],
+ seq_len_list=q_seq_len_list[i:i + 1],
37
num_q_heads=num_q_heads,
38
num_kv_heads=num_kv_heads,
39
attn_mask=context.attention_mask[i:i + 1],
0 commit comments