Commit Graph

25 Commits (11704cc721197be84ed8a0a816dc54c1468f7cc0)

Author SHA1 Message Date
Ross Wightman c3be79a8b7 Rename _pretrained.py -> pretrained.py, not feasible to change the other files to same scheme without breaking uses
1 year ago
Ross Wightman 23b357f1df Add ported Tensorflow MaxVit weights. Add a few more CLIP ViT fine-tunes. Tweak some model tag names. Improve model tag name sorting. Update HF hub push config layout.
1 year ago
Ross Wightman 2eb825c014 Add missing model config
2 years ago
Ross Wightman 0761ce7a1b Updating vit model defs for mult-weight support trial (vit first). Prepping for CLIP (laion2b and openai) fine-tuned weights.
2 years ago
Ross Wightman ebb99a1f8d Initial multi-weight support, handled so old pretraing config handling co-exists with new tags.
2 years ago
Ross Wightman 9709dbaaa9 Adding support for fine-tune CLIP LAION-2B image tower weights for B/32, L/14, H/14 and g/14. Still WIP
2 years ago
Ross Wightman 41dc49a337 Vision Transformer refactoring and Rel Pos impl
2 years ago
Ross Wightman b049a5c5c6 Merge remote-tracking branch 'origin/master' into norm_norm_norm
2 years ago
Ross Wightman b20665d379
Merge pull request #1007 from qwertyforce/patch-1
2 years ago
Ross Wightman 95cfc9b3e8 Merge remote-tracking branch 'origin/master' into norm_norm_norm
2 years ago
Ross Wightman abc9ba2544 Transitioning default_cfg -> pretrained_cfg. Improving handling of pretrained_cfg source (HF-Hub, files, timm config, etc). Checkpoint handling tweaks.
2 years ago
Ross Wightman 656757d26b Fix MobileNetV2 head conv size for multiplier < 1.0. Add some missing modification copyrights, fix starting date of some old ones.
2 years ago
qwertyforce ccb3815360
update arxiv link
3 years ago
Ross Wightman 9c9755a808 AugReg release
3 years ago
Ross Wightman 381b279785 Add hybrid model fwds back
3 years ago
Ross Wightman 26f04a8e3e Fix a weight link
3 years ago
Ross Wightman b319eb5b5d Update ViT weights, more details to be added before merge.
3 years ago
Ross Wightman b9cfb64412 Support npz custom load for vision transformer hybrid models. Add posembed rescale for npz load.
3 years ago
Ross Wightman 8880f696b6 Refactoring, cleanup, improved test coverage.
3 years ago
Ross Wightman ba2ca4b464 One codepath for stdconv, switch layernorm to batchnorm so gain included. Tweak epsilon values for nfnet, resnetv2, vit hybrid.
3 years ago
Ross Wightman 8e4ac3549f All ScaledStdConv and StdConv uses default to using F.layernorm so that they work with PyTorch XLA. eps value tweaking is a WIP.
3 years ago
Ross Wightman bfc72f75d3 Expand scope of testing for non-std vision transformer / mlp models. Some related cleanup and create fn cleanup for all vision transformer and mlp models. More CoaT weights.
3 years ago
Ross Wightman 288682796f Update benchmark script to add precision arg. Fix some downstream (DeiT) compat issues with latest changes. Bump version to 0.4.7
3 years ago
Ross Wightman ea9c9550b2 Fully move ViT hybrids to their own file, including embedding module. Remove some extra DeiT models that were for benchmarking only.
3 years ago
Ross Wightman a5310a3451 Merge remote-tracking branch 'origin/benchmark-fixes-vit_hybrids' into pit_and_vit_update
3 years ago