@@ -93,7 +93,7 @@ def __init__(self,
93
93
out_features = self .encoder_dim ,
94
94
bias = True )
95
95
self .pos_encode = AddPositionalEmbedding (embedding_dim = self .encoder_dim )
96
- self .dropout = nn .Dropout (p = self .input_dropout_rate )
96
+ self .dropout = nn .Dropout (p = self .input_dropout_rate , inplace = True )
97
97
98
98
def forward (self , inputs , input_paddings ):
99
99
output_paddings = input_paddings
@@ -195,7 +195,7 @@ def __init__(self, config: ConformerConfig):
195
195
in_features = config .encoder_dim ,
196
196
out_features = config .encoder_dim * config .feed_forward_expansion_factor ,
197
197
bias = True )
198
- self .dropout1 = nn .Dropout (p = config .feed_forward_dropout_rate )
198
+ self .dropout1 = nn .Dropout (p = config .feed_forward_dropout_rate , inplace = True )
199
199
self .linear2 = nn .Linear (
200
200
in_features = config .encoder_dim * config .feed_forward_expansion_factor ,
201
201
out_features = config .encoder_dim ,
@@ -206,7 +206,8 @@ def __init__(self, config: ConformerConfig):
206
206
else :
207
207
feed_forward_residual_dropout_rate = (
208
208
config .feed_forward_residual_dropout_rate )
209
- self .dropout2 = nn .Dropout (p = feed_forward_residual_dropout_rate )
209
+ self .dropout2 = nn .Dropout (
210
+ p = feed_forward_residual_dropout_rate , inplace = True )
210
211
211
212
def forward (self , inputs , padding_mask ):
212
213
inputs = self .ln (inputs )
@@ -316,7 +317,7 @@ def __init__(self, config: ConformerConfig):
316
317
attention_residual_dropout_rate = 0.1
317
318
else :
318
319
attention_residual_dropout_rate = config .attention_residual_dropout_rate
319
- self .dropout = nn .Dropout (p = attention_residual_dropout_rate )
320
+ self .dropout = nn .Dropout (p = attention_residual_dropout_rate , inplace = True )
320
321
321
322
def forward (self , outputs , paddings ):
322
323
outputs = self .ln (outputs )
@@ -407,7 +408,7 @@ def __init__(self, config):
407
408
conv_residual_dropout_rate = 0.0
408
409
else :
409
410
conv_residual_dropout_rate = config .conv_residual_dropout_rate
410
- self .dropout = nn .Dropout (p = conv_residual_dropout_rate )
411
+ self .dropout = nn .Dropout (p = conv_residual_dropout_rate , inplace = True )
411
412
412
413
def forward (self , inputs , input_paddings ):
413
414
inputs = self .ln (inputs )
0 commit comments