Commit Graph

16 Commits (a3c6685e201e1845b523603b548619419901c229)

Author SHA1 Message Date
Ross Wightman b1b024dfed Scheduler update, add v2 factory method, support scheduling on updates instead of just epochs. Add LR to summary csv. Add lr_base scaling calculations to train script. Fix #1168
2 years ago
han a16171335b fix: change milestones to decay-milestones
3 years ago
han 57a988df30 fix: multistep lr decay epoch bugs
3 years ago
Ross Wightman 29a37e23ee LR scheduler update:
3 years ago
Sepehr Sameni abf3e044bb
Update scheduler_factory.py
3 years ago
samarth daab57a6d9 1. Added a simple multi step LR scheduler
3 years ago
Ross Wightman 6c17d57a2c Fix some attributions, add copyrights to some file docstrings
4 years ago
Ross Wightman d72ac0db25 Fix #173, lr cycle default 0 vs 1. Fix #177, mirror resnest weights for future stability.
4 years ago
Ross Wightman f225ae8e59 Update README with model results and attribution. Make scheduler factory bit more robust to arg differences, add noise to plateau lr and fix min/max.
4 years ago
Ross Wightman 6441e9cc1b Fix memory_efficient mode for DenseNets. Add AntiAliasing (Blur) support for DenseNets and create one test model. Add lr cycle/mul params to train args.
5 years ago
Ross Wightman c16f25ced2 Add MobileNetV3 Large weights, results, update README and sotabench for merge
5 years ago
Ross Wightman 27b3680d49 Revamp LR noise, move logic to scheduler base. Fixup PlateauLRScheduler and add it as an option.
5 years ago
Ross Wightman 514b0938c4 Experimenting with per-epoch learning rate noise w/ step scheduler
5 years ago
Ross Wightman fac58f609a Add RAdam, NovoGrad, Lookahead, and AdamW optimizers, a few ResNet tweaks and scheduler factory tweak.
5 years ago
Ross Wightman e7c8a37334 Make min-lr and cooldown-epochs cmdline args, change dash in color_jitter arg for consistency
5 years ago
Ross Wightman aa4354f466 Big re-org, working towards making pip/module as 'timm'
5 years ago