diff --git a/timm/optim/optim_factory.py b/timm/optim/optim_factory.py index 80bac373..d3592e80 100644 --- a/timm/optim/optim_factory.py +++ b/timm/optim/optim_factory.py @@ -61,8 +61,10 @@ def create_optimizer(args, model, filter_bias_and_bn=True): opt_split = opt_lower.split('_') opt_lower = opt_split[-1] if opt_lower == 'sgd' or opt_lower == 'nesterov': + del opt_args['eps'] optimizer = optim.SGD(parameters, momentum=args.momentum, nesterov=True, **opt_args) elif opt_lower == 'momentum': + del opt_args['eps'] optimizer = optim.SGD(parameters, momentum=args.momentum, nesterov=False, **opt_args) elif opt_lower == 'adam': optimizer = optim.Adam(parameters, **opt_args) @@ -93,8 +95,10 @@ def create_optimizer(args, model, filter_bias_and_bn=True): elif opt_lower == 'nvnovograd': optimizer = NvNovoGrad(parameters, **opt_args) elif opt_lower == 'fusedsgd': + del opt_args['eps'] optimizer = FusedSGD(parameters, momentum=args.momentum, nesterov=True, **opt_args) elif opt_lower == 'fusedmomentum': + del opt_args['eps'] optimizer = FusedSGD(parameters, momentum=args.momentum, nesterov=False, **opt_args) elif opt_lower == 'fusedadam': optimizer = FusedAdam(parameters, adam_w_mode=False, **opt_args)