We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 75d951b commit 6a9f14bCopy full SHA for 6a9f14b
QEfficient/transformers/models/t5/modeling_t5.py
@@ -155,6 +155,9 @@ def forward(
155
156
157
class QEffT5LayerSelfAttention(T5LayerSelfAttention):
158
+ def __qeff_init__(self):
159
+ self.scaling_factor = 1.0
160
+
161
def forward(
162
self,
163
hidden_states,
@@ -177,7 +180,7 @@ def forward(
177
180
output_attentions=output_attentions,
178
181
cache_position=cache_position,
179
182
)
- hidden_states = hidden_states * self.scaling_factor + self.dropout(attention_output[0]) # Modified by patch
183
+ hidden_states = hidden_states * 1.0 + self.dropout(attention_output[0]) # Modified by patch
184
outputs = (hidden_states,) + attention_output[1:] # add attentions if we output them
185
return outputs
186
0 commit comments