File tree Expand file tree Collapse file tree 2 files changed +5
-4
lines changed
Expand file tree Collapse file tree 2 files changed +5
-4
lines changed Original file line number Diff line number Diff line change 1- from .adamp import AdamP
2- from .adamw import AdamW
1+ from .adabelief import AdaBelief
32from .adafactor import Adafactor
43from .adahessian import Adahessian
4+ from .adamp import AdamP
5+ from .adamw import AdamW
6+ from .lamb import Lamb
7+ from .lars import Lars
58from .lookahead import Lookahead
69from .madgrad import MADGRAD
710from .nadam import Nadam
811from .nvnovograd import NvNovoGrad
912from .radam import RAdam
1013from .rmsprop_tf import RMSpropTF
1114from .sgdp import SGDP
12- from .adabelief import AdaBelief
1315from .optim_factory import create_optimizer , create_optimizer_v2 , optimizer_kwargs
Original file line number Diff line number Diff line change @@ -87,7 +87,6 @@ def step(self, closure=None):
8787 device = self .param_groups [0 ]['params' ][0 ].device
8888 one_tensor = torch .tensor (1.0 , device = device ) # because torch.where doesn't handle scalars correctly
8989
90- # exclude scaling for params with 0 weight decay
9190 for group in self .param_groups :
9291 weight_decay = group ['weight_decay' ]
9392 momentum = group ['momentum' ]
You can’t perform that action at this time.
0 commit comments