f2e14685a8Add force-cpu flag for train/validate, fix CPU fallback for device init, remove old force cpu flag for EMA model weights
Ross Wightman
2021-08-22 13:15:27 -0700
a8b65695f1Add resnet26ts and resnext26ts models for non-attn baselines
Ross Wightman
2021-08-21 12:42:10 -0700
a5a542f17dFix typo
Ross Wightman
2021-08-20 17:47:23 -0700
925e102982Update attention / self-attn based models from a series of experiments: * remove dud attention, involution + my swin attention adaptation don't seem worth keeping * add or update several new 26/50 layer ResNe(X)t variants that were used in experiments * remove models associated with dead-end or uninteresting experiment results * weights coming soon...
Ross Wightman
2021-08-20 16:13:11 -0700
d54cdedc86add two last outputs to feature_info
#816
Yohann Lereclus
2021-08-19 10:56:53 +0200
a16a753852Add lamb/lars to optim init imports, remove stray comment
Ross Wightman
2021-08-18 22:55:02 -0700
c207e02782MOAR optimizer changes. Woo!
Ross Wightman
2021-08-18 22:20:35 -0700
42c1f0cf6cFix lars tests
Ross Wightman
2021-08-18 21:05:34 -0700
f4fb068b11Merge branch 'master' into bits_and_tpu
Ross Wightman
2021-08-18 17:25:00 -0700
a426511c95More optimizer cleanup. Change all to no longer use .data. Improve (b)float16 use with adabelief. Add XLA compatible Lars.
Ross Wightman
2021-08-18 17:21:56 -0700
b0265ef8a6Merge branch 'master' into bits_and_tpu
Ross Wightman
2021-08-18 11:21:21 -0700
9541f4963bOne more scalar -> tensor fix for lamb optimizer
Ross Wightman
2021-08-18 11:20:25 -0700
0d82876132Add comment for reference re PyTorch XLA 'race' issue
Ross Wightman
2021-08-18 11:15:32 -0700
b76b48e8e9Update optimizer creation for master optimizer changes
Ross Wightman
2021-08-18 11:14:51 -0700
f98662b9c9Merge branch 'master' into bits_and_tpu
Ross Wightman
2021-08-18 09:31:10 -0700
Merge pull request #813 from rwightman/opt_cleanup
Ross Wightman
2021-08-18 09:12:00 -0700
a6af48be64add madgradw optimizer
#813
Ross Wightman
2021-08-17 22:19:27 -0700
55fb5eedf6Remove experiment from lamb impl
Ross Wightman
2021-08-17 21:48:26 -0700
8a9eca5157A few optimizer comments, dead import, missing import
Ross Wightman
2021-08-17 18:01:33 -0700
959eaff121Add optimizer tests and update testing to pytorch 1.9
Ross Wightman
2021-08-17 17:59:15 -0700
ac469b50daOptimizer improvements, additions, cleanup * Add MADGRAD code * Fix Lamb (non-fused variant) to work w/ PyTorch XLA * Tweak optimizer factory args (lr/learning_rate and opt/optimizer_name), may break compat * Use newer fn signatures for all add,addcdiv, addcmul in optimizers * Use upcoming PyTorch native Nadam if it's available * Cleanup lookahead opt * Add optimizer tests * Remove novograd.py impl as it was messy, keep nvnovograd * Make AdamP/SGDP work in channels_last layout * Add rectified adablief mode (radabelief) * Support a few more PyTorch optim, adamax, adagrad
Ross Wightman
2021-08-17 17:51:20 -0700
93b27d69e7Fix RuntimeError: Integer division of tensors using div or / is no longer supported, and in a future release div will perform true division as in Python 3. Use true_divide or floor_divide (// in Python) instead.
#812
danil328
2021-08-17 14:19:59 +0300
b1d570b0adFix RuntimeError: Integer division of tensors using div or / is no longer supported, and in a future release div will perform true division as in Python 3. Use true_divide or floor_divide (// in Python) instead.
danil328
2021-08-17 13:57:44 +0300
cb621e0f00Remove print, arg order
Ross Wightman
2021-08-13 12:54:12 -0700
b974d85026Merge branch 'bits_and_tpu' of github.com:rwightman/pytorch-image-models into bits_and_tpu
Ross Wightman
2021-08-13 12:49:16 -0700
c06c739901Merge branch 'master' into bits_and_tpu
Ross Wightman
2021-08-13 12:49:04 -0700
40457e5691Transforms, augmentation work for bits, add RandomErasing support for XLA (pushing into transforms), revamp of transform/preproc config, etc ongoing...
Ross Wightman
2021-08-13 12:45:43 -0700
3cdaf5ed56Add `mmax` config key to auto_augment for increasing upper bound of RandAugment magnitude beyond 10. Make AugMix uniform sampling default not override config setting.
Ross Wightman
2021-08-12 15:39:05 -0700
fix a typo in ### Select specific feature levels or limit the stride
#801
Yonghye Kwon
2021-08-13 01:36:48 +0900
1042b8a146Add non fused LAMB optimizer option
Ross Wightman
2021-08-09 13:13:43 -0700
01cb46a9a5Add gc_efficientnetv2_rw_t weights (global context instead of SE attn). Add TF XL weights even though the fine-tuned ones don't validate that well. Change default arg for GlobalContext to use scal (mul) mode.
Ross Wightman
2021-08-07 16:45:29 -0700
f8039c7492Fix gc effv2 model cfg name
Ross Wightman
2021-07-11 12:14:31 -0700
3a55a30ed1add notes from author
#747
Alexander Soare
2021-07-11 14:25:58 +0100
899cf84cccbug fix - missing _dist postfix for many of the 224_dist models
Alexander Soare
2021-07-11 12:41:51 +0100
623e8b8eb8wip xcit
Alexander Soare
2021-07-08 21:05:44 +0100
392368e210Add efficientnetv2_rw_t defs w/ weights, and gc variant, as well as gcresnet26ts for experiments. Version 0.4.13
Ross Wightman
2021-07-09 16:46:52 -0700
daab57a6d91. Added a simple multi step LR scheduler
#746
samarth
2021-07-09 16:18:27 -0400
6d8272e92cAdd SAM pretrained model defs/weights for ViT B16 and B32 models.
Ross Wightman
2021-07-08 11:23:55 -0700
ee4d8fc69aRemove unecessary line from nest post refactor
Ross Wightman
2021-07-05 21:22:46 -0700
c8ec1ffcb9Merge branch 'alexander-soare-nested_transformer'
Ross Wightman
2021-07-05 18:22:50 -0700
8165cacd82Realized LayerNorm2d won't work in all cases as is, fixed.
Ross Wightman
2021-07-05 18:21:34 -0700
81cd6863c8Move aggregation (convpool) for nest into NestLevel, cleanup and enable features_only use. Finalize weight url.
Ross Wightman
2021-07-05 18:20:49 -0700
Add color highlighting to BibTeX entry in README
#729
Santiago Castro
2021-06-30 14:07:16 -0700
7096b52a61Remove sotabench files, no longer working / maintained
v0.4.12
Ross Wightman
2021-06-30 09:25:58 -0700
d10b071a28Update results csvs w/ latest ViT, ResMLP, and NfNet-L2 weights present
Ross Wightman
2021-06-30 09:10:35 -0700
766b4d3262Fix features for resnetv2_50t
Ross Wightman
2021-06-28 15:56:24 -0700
e8045e712fFix BatchNorm for ResNetV2 non GN models, add more ResNetV2 model defs for future experimentation, fix zero_init of last residual for pre-act.
Ross Wightman
2021-06-28 10:52:45 -0700
85f894e03dFix ViT in21k representation (pre_logits) layer handling across old and new npz checkpoints
Ross Wightman
2021-06-23 10:38:34 -0700
b41cffaa93Fix a few issues loading pretrained vit/bit npz weights w/ num_classes=0 __init__ arg. Missed a few other small classifier handling detail on Mlp, GhostNet, Levit. Should fix#713
Ross Wightman
2021-06-22 23:16:05 -0700
381b279785Add hybrid model fwds back
Ross Wightman
2021-06-19 22:28:44 -0700
26f04a8e3eFix a weight link
Ross Wightman
2021-06-19 16:39:36 -0700
8f4a0222edAdd GMixer-24 MLP model weights, trained w/ TPU + PyTorch XLA
Ross Wightman
2021-06-18 16:49:28 -0700
4c09a2f169Bump version 0.4.12
Ross Wightman
2021-06-18 16:17:34 -0700
b319eb5b5dUpdate ViT weights, more details to be added before merge.
Ross Wightman
2021-06-18 16:16:49 -0700
8257b86550Fix up resnetv2 bit/bitm model default res
Ross Wightman
2021-06-18 16:16:06 -0700
1228f5a3d8Add BiT distilled 50x1 and teacher 152x2 models from 'A good teacher is patient and consistent' paper.
Ross Wightman
2021-06-18 11:40:33 -0700
511a8e8c96Add official ResMLP weights.
Ross Wightman
2021-06-14 17:01:12 -0700
b9cfb64412Support npz custom load for vision transformer hybrid models. Add posembed rescale for npz load.
Ross Wightman
2021-06-14 12:31:44 -0700
8319e0c373Add file docstring to std_conv.py
Ross Wightman
2021-06-13 12:31:06 -0700
0020268d9bTry lower max size for non_std default_cfg test
Ross Wightman
2021-06-12 23:31:24 -0700
4d96165989Merge branch 'master' into cleanup_xla_model_fixes
Ross Wightman
2021-06-12 23:19:25 -0700
8880f696b6Refactoring, cleanup, improved test coverage. * Add eca_nfnet_l2 weights, 84.7 @ 384x384 * All 'non-std' (ie transformer / mlp) models have classifier / default_cfg test added * Fix#694 reset_classifer / num_features / forward_features / num_classes=0 consistency for transformer / mlp models * Add direct loading of npz to vision transformer (pure transformer so far, hybrid to come) * Rename vit_deit* to deit_* * Remove some deprecated vit hybrid model defs * Clean up classifier flatten for conv classifiers and unusual cases (mobilenetv3/ghostnet) * Remove explicit model fns for levit conv, just pass in arg
Ross Wightman
2021-06-12 16:40:02 -0700