We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 75d951b commit ef2cd47Copy full SHA for ef2cd47
QEfficient/transformers/models/t5/modeling_t5.py
@@ -177,7 +177,7 @@ def forward(
177
output_attentions=output_attentions,
178
cache_position=cache_position,
179
)
180
- hidden_states = hidden_states * self.scaling_factor + self.dropout(attention_output[0]) # Modified by patch
+ hidden_states = hidden_states * 1.0 + self.dropout(attention_output[0]) # Modified by patch
181
outputs = (hidden_states,) + attention_output[1:] # add attentions if we output them
182
return outputs
183
0 commit comments