Skip to content

Commit 0c3c237

Browse files
AdamRajferjan-golda
authored andcommitted
[ConvNets/EffNetV2][PyT] Fix grad scaling and dropout
1 parent c903326 commit 0c3c237

2 files changed

Lines changed: 3 additions & 9 deletions

File tree

PyTorch/Classification/ConvNets/image_classification/models/common.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -289,13 +289,13 @@ def forward(self, residual: torch.Tensor, x: torch.Tensor) -> torch.Tensor:
289289
return torch.add(residual, other=x)
290290
else:
291291
with torch.no_grad():
292-
F.dropout(
292+
mask = F.dropout(
293293
self.mask,
294294
p=1 - self.survival_prob,
295295
training=self.training,
296296
inplace=False,
297297
)
298-
return torch.addcmul(residual, self.mask, x)
298+
return torch.addcmul(residual, mask, x)
299299

300300
class Flatten(nn.Module):
301301
def forward(self, x: torch.Tensor) -> torch.Tensor:

PyTorch/Classification/ConvNets/main.py

Lines changed: 1 addition & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -265,12 +265,6 @@ def add_parser_arguments(parser, skip_arch=False):
265265
default=1,
266266
help="Static loss scale, positive power of 2 values can improve amp convergence.",
267267
)
268-
parser.add_argument(
269-
"--dynamic-loss-scale",
270-
action="store_true",
271-
help="Use dynamic loss scaling. If supplied, this argument supersedes "
272-
+ "--static-loss-scale.",
273-
)
274268
parser.add_argument(
275269
"--prof", type=int, default=-1, metavar="N", help="Run only N iterations"
276270
)
@@ -483,7 +477,7 @@ def _worker_init_fn(id):
483477
init_scale=args.static_loss_scale,
484478
growth_factor=2,
485479
backoff_factor=0.5,
486-
growth_interval=100 if args.dynamic_loss_scale else 1000000000,
480+
growth_interval=100,
487481
enabled=args.amp,
488482
)
489483

0 commit comments

Comments
 (0)