Commit Graph

1292 Commits (7eb7e73216c69b87f32973965ae61dd43f8ac7ec)
 

Author SHA1 Message Date
Ross Wightman 484e61648d Adding the attn series weights, tweaking model names, comments...
3 years ago
Ross Wightman 0639d9a591 Fix updated validation_batch_size fallback
3 years ago
Ross Wightman 5db057dca0 Fix misnamed arg, tweak other train script args for better defaults.
3 years ago
Ross Wightman fb94350896 Update training script and loader factory to allow use of scheduler updates, repeat augment, and bce loss
3 years ago
Ross Wightman f262137ff2 Add RepeatAugSampler as per DeiT RASampler impl, showing promise for current (distributed) training experiments.
3 years ago
Ross Wightman ba9c1108a1 Add a BCE loss impl that converts dense targets to sparse /w smoothing as an alternate to CE w/ smoothing. For training experiments.
3 years ago
Ross Wightman 29a37e23ee LR scheduler update:
3 years ago
nateraw 28d2841acf 💄 apply isort
3 years ago
Ross Wightman 492c0a4e20 Update HaloAttn comment
3 years ago
nateraw e72c989973 add ability to push to hf hub
3 years ago
Richard Chen 7ab9d4555c add crossvit
3 years ago
Ross Wightman 3b9032ea48 Use Tensor.unfold().unfold() for HaloAttn, fast like as_strided but more clarity
3 years ago
Ross Wightman fc894c375c Another attempt at sgd momentum test passing...
3 years ago
Ross Wightman 78933122c9 Fix silly typo
3 years ago
Ross Wightman 2568ffc5ef Merge branch 'master' into attn_update
3 years ago
Ross Wightman 708d87a813 Fix ViT SAM weight compat as weights at URL changed to not use repr layer. Fix #825. Tweak optim test.
3 years ago
Ross Wightman 8449ba210c Improve performance of HaloAttn, change default dim calc. Some cleanup / fixes for byoanet. Rename resnet26ts to tfs to distinguish (extra fc).
3 years ago
Ross Wightman f2e14685a8 Add force-cpu flag for train/validate, fix CPU fallback for device init, remove old force cpu flag for EMA model weights
3 years ago
Ross Wightman a8b65695f1 Add resnet26ts and resnext26ts models for non-attn baselines
3 years ago
Ross Wightman a5a542f17d Fix typo
3 years ago
Ross Wightman 925e102982 Update attention / self-attn based models from a series of experiments:
3 years ago
Ross Wightman acd6c687fd git push origin masterMerge branch 'yohann84L-fix_accuracy'
3 years ago
Ross Wightman d667351eac Tweak accuracy topk safety. Fix #807
3 years ago
Ross Wightman 2ee398d501 Merge branch 'master' into bits_and_tpu
3 years ago
Yohann Lereclus 35c9740826 Fix accuracy when topk > num_classes
3 years ago
Ross Wightman a16a753852 Add lamb/lars to optim init imports, remove stray comment
3 years ago
Ross Wightman c207e02782 MOAR optimizer changes. Woo!
4 years ago
Ross Wightman 42c1f0cf6c Fix lars tests
4 years ago
Ross Wightman f4fb068b11 Merge branch 'master' into bits_and_tpu
4 years ago
Ross Wightman a426511c95 More optimizer cleanup. Change all to no longer use .data. Improve (b)float16 use with adabelief. Add XLA compatible Lars.
4 years ago
Ross Wightman b0265ef8a6 Merge branch 'master' into bits_and_tpu
4 years ago
Ross Wightman 9541f4963b One more scalar -> tensor fix for lamb optimizer
4 years ago
Ross Wightman 0d82876132 Add comment for reference re PyTorch XLA 'race' issue
4 years ago
Ross Wightman b76b48e8e9 Update optimizer creation for master optimizer changes
4 years ago
Ross Wightman f98662b9c9 Merge branch 'master' into bits_and_tpu
4 years ago
Ross Wightman 8f68193c91
Update lamp.py comment
4 years ago
Ross Wightman 4d284017b8
Merge pull request #813 from rwightman/opt_cleanup
4 years ago
Ross Wightman a6af48be64 add madgradw optimizer
4 years ago
Ross Wightman 55fb5eedf6 Remove experiment from lamb impl
4 years ago
Ross Wightman 8a9eca5157 A few optimizer comments, dead import, missing import
4 years ago
Ross Wightman 959eaff121 Add optimizer tests and update testing to pytorch 1.9
4 years ago
Ross Wightman ac469b50da Optimizer improvements, additions, cleanup
4 years ago
Ross Wightman 368211d19a
Merge pull request #805 from Separius/patch-1
4 years ago
Sepehr Sameni abf3e044bb
Update scheduler_factory.py
4 years ago
Ross Wightman cb621e0f00 Remove print, arg order
4 years ago
Ross Wightman b974d85026 Merge branch 'bits_and_tpu' of github.com:rwightman/pytorch-image-models into bits_and_tpu
4 years ago
Ross Wightman c06c739901 Merge branch 'master' into bits_and_tpu
4 years ago
Ross Wightman 40457e5691 Transforms, augmentation work for bits, add RandomErasing support for XLA (pushing into transforms), revamp of transform/preproc config, etc ongoing...
4 years ago
Ross Wightman 3cdaf5ed56 Add `mmax` config key to auto_augment for increasing upper bound of RandAugment magnitude beyond 10. Make AugMix uniform sampling default not override config setting.
4 years ago
Ross Wightman 1042b8a146 Add non fused LAMB optimizer option
4 years ago