We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent cabda24 commit e07fa9dCopy full SHA for e07fa9d
tensorrt_llm/models/gemma/model.py
@@ -157,10 +157,13 @@ def forward(self,
157
if default_net().plugin_config.reduce_fusion else
158
AllReduceFusionOp.NONE,
159
residual=residual,
160
- norm_weight=self.pre_feedforward_layernorm.weight.value,
161
- norm_pre_residual_weight=self.post_layernorm.weight.value
+ norm_weight=self.pre_feedforward_layernorm.weight.value
162
if self.config.inter_layernorms else None,
163
- eps=self.pre_feedforward_layernorm.eps))
+ norm_pre_residual_weight=self.post_layernorm.weight.value,
+ eps=self.pre_feedforward_layernorm.eps
164
+ if self.config.inter_layernorms else 1e-06,
165
+ ),
166
+ )
167
168
if use_cache:
169
attention_output, presents = attention_output
0 commit comments