Ross Wightman
30b9880d06
Minor adjustment, mutable default arg, extra check of valid len...
4 years ago
Alexander Soare
8086943b6f
allow resize positional embeddings to non-square grid
4 years ago
Ross Wightman
b2c305c2aa
Move Mlp and PatchEmbed modules into layers. Being used in lots of models now...
4 years ago
Ross Wightman
a0492e3b48
A few miil weights naming tweaks to improve compat with model registry and filtering wildcards.
4 years ago
talrid
19e1b67a84
old spaces
4 years ago
talrid
a443865876
update naming and scores
4 years ago
talrid
cf0e371594
84_0
4 years ago
talrid
0968bdeca3
vit, tresnet and mobilenetV3 ImageNet-21K-P weights
4 years ago
Ross Wightman
f606c45c38
Add Swin Transformer models from https://github.com/microsoft/Swin-Transformer
4 years ago
Ross Wightman
bf2ca6bdf4
Merge jax and original weight init
4 years ago
Ross Wightman
acbd698c83
Update README.md with updates. Small tweak to head_dist handling.
4 years ago
Ross Wightman
288682796f
Update benchmark script to add precision arg. Fix some downstream (DeiT) compat issues with latest changes. Bump version to 0.4.7
4 years ago
Ross Wightman
ea9c9550b2
Fully move ViT hybrids to their own file, including embedding module. Remove some extra DeiT models that were for benchmarking only.
4 years ago
Ross Wightman
a5310a3451
Merge remote-tracking branch 'origin/benchmark-fixes-vit_hybrids' into pit_and_vit_update
4 years ago
Ross Wightman
7953e5d11a
Fix pos_embed scaling for ViT and num_classes != 1000 for pretrained distilled deit and pit models. Fix #426 and fix #433
4 years ago
Ross Wightman
a760a4c3f4
Some ViT cleanup, merge distilled model with main, fixup torchscript support for distilled models
4 years ago
Ross Wightman
cf5fec5047
Cleanup experimental vit weight init a bit
4 years ago
Ross Wightman
cbcb76d72c
Should have included Conv2d layers in original weight init. Lets see what the impact is...
4 years ago
Ross Wightman
4de57ccf01
Add weight init scheme that's closer to JAX impl
4 years ago
Ross Wightman
45c048ba13
A few minor fixes and bit more cleanup on the huggingface hub integration.
4 years ago
Ross Wightman
d584e7f617
Support for huggingface hub via create_model and default_cfgs.
...
* improve consistency of model creation helper fns
* add comments to some of the model helpers
* support passing external default_cfgs so they can be sourced from hub
4 years ago
Ross Wightman
17cdee7354
Fix C&P patch_size error, and order of op patch_size arg resolution bug. Remove a test vit model.
4 years ago
Ross Wightman
0706d05d52
Benchmark models listed in txt file. Add more hybrid vit variants for testing
4 years ago
Ross Wightman
de97be9146
Spell out diff between my small and deit small vit models.
4 years ago
Ross Wightman
f0ffdf89b3
Add numerous experimental ViT Hybrid models w/ ResNetV2 base. Update the ViT naming for hybrids. Fix #426 for pretrained vit resizing.
4 years ago
Ross Wightman
5a8e1e643e
Initial Normalizer-Free Reg/ResNet impl. A bit of related layer refactoring.
4 years ago
Ross Wightman
bb50ac4708
Add DeiT distilled weights and distilled model def. Remove some redudant ViT model args.
4 years ago
Ross Wightman
c16e965037
Add some ViT comments and fix a few minor issues.
4 years ago
Ross Wightman
55f7dfa9ea
Refactor vision_transformer entrpy fns, add pos embedding resize support for fine tuning, add some deit models for testing
4 years ago
Ross Wightman
855d6cc217
More dataset work including factories and a tensorflow datasets (TFDS) wrapper
...
* Add parser/dataset factory methods for more flexible dataset & parser creation
* Add dataset parser that wraps TFDS image classification datasets
* Tweak num_classes handling bug for 21k models
* Add initial deit models so they can be benchmarked in next csv results runs
4 years ago
Ross Wightman
ce69de70d3
Add 21k weight urls to vision_transformer. Cleanup feature_info for preact ResNetV2 (BiT) models
4 years ago
Ross Wightman
231d04e91a
ResNetV2 pre-act and non-preact model, w/ BiT pretrained weights and support for ViT R50 model. Tweaks for in21k num_classes passing. More to do... tests failing.
4 years ago
Ross Wightman
b401952caf
Add newly added vision transformer large/base 224x224 weights ported from JAX official repo
4 years ago
Ross Wightman
61200db0ab
in_chans=1 working w/ pretrained weights for vision_transformer
4 years ago
Ross Wightman
f591e90b0d
Make sure num_features attr is present in vit models as with others
4 years ago
Ross Wightman
f944242cb0
Fix #262 , num_classes arg mixup. Make vision_transformers a bit closer to other models wrt get/reset classfier/forward_features. Fix torchscript for ViT.
4 years ago
Ross Wightman
736f209e7d
Update vision transformers to be compatible with official code. Port official ViT weights from jax impl.
4 years ago
Ross Wightman
27a93e9de7
Improve test crop for ViT models. Small now 77.85, added base weights at 79.35 top-1.
4 years ago
Ross Wightman
d4db9e7977
Add small vision transformer weights. 77.42 top-1.
4 years ago
Ross Wightman
f31933cb37
Initial Vision Transformer impl w/ patch and hybrid variants. Refactor tuple helpers.
4 years ago