We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4c68412 commit 506fc72Copy full SHA for 506fc72
QEfficient/transformers/models/llama/modeling_llama.py
@@ -174,7 +174,7 @@ def forward(
174
)
175
176
attn_output = attn_output.reshape(*input_shape, -1).contiguous()
177
- attn_output = self.o_proj(attn_output)
+ attn_output = self.o_proj(attn_output, **kwargs)
178
return attn_output, attn_weights, past_key_value
179
180
0 commit comments