d0b45c9b4dMake safetensor import option for now. Improve avg/clean checkpoints ext handling a bit (more consistent).
Ross Wightman
2023-02-18 16:06:42 -0800
022403ce0aUpdate README
Ross Wightman
2023-02-16 17:20:27 -0800
2e38d53dcaRemove dead line
Ross Wightman
2023-02-15 15:38:41 -0800
f77c04ff36Torchscript fixes/hacks for rms_norm, refactor ParallelScalingBlock with manual combination of input projections, closer paper match
Ross Wightman
2023-02-15 14:53:55 -0800
122621daefAdd Final annotation to attn_fas to avoid symbol lookup of new scaled_dot_product_attn fn on old PyTorch in jit
Ross Wightman
2023-02-15 08:53:50 -0800
621e1b2182Add ideas from 'Scaling ViT to 22-B Params', testing PyTorch 2.0 fused F.scaled_dot_product_attention impl in vit, vit_relpos, maxxvit / coatnet.
Ross Wightman
2023-02-14 23:32:04 -0800
a3d528524aVersion 0.8.12dev0
Ross Wightman
2023-02-16 16:27:29 -0800
a09d403c24changed warning to info
testbot
2023-02-16 11:27:27 +0100
8470e29541Add support to load safetensors weights
testbot
2023-02-16 10:47:04 +0100
f35d6ea57bAdd multi-tensor (foreach) version of Lion in style of upcoming PyTorch 2.0 optimizers
Ross Wightman
2023-02-16 15:48:00 -0800
c53cf76fa3Torchscript fixes/hacks for rms_norm, refactor ParallelScalingBlock with manual combination of input projections, closer paper match
Ross Wightman
2023-02-15 14:53:55 -0800
49f248fa18Merge branch 'main' into scaling_vit
Ross Wightman
2023-02-15 09:26:28 -0800
b6eb652924Add Final annotation to attn_fas to avoid symbol lookup of new scaled_dot_product_attn fn on old PyTorch in jit
Ross Wightman
2023-02-15 08:53:50 -0800
00001f91d0fix bugs when tracing of swin_transformer
#1673
root
2023-02-15 21:42:24 +0800
709d5e0d9dAdd Lion optimizer
Ross Wightman
2023-02-14 23:55:05 -0800
a9739258f4Add ideas from 'Scaling ViT to 22-B Params', testing PyTorch 2.0 fused F.scaled_dot_product_attention impl in vit, vit_relpos, maxxvit / coatnet.
Ross Wightman
2023-02-14 23:32:04 -0800
0f2803de7aMove ImageNet metadata (aka info) files to timm/data/_info. Add helper classes to make info available for labelling. Update inference.py for first use.
Ross Wightman
2023-02-06 17:41:07 -0800
8682528096Add first conv metadata for efficientformer_v2
Ross Wightman
2023-02-04 23:02:02 -0800
72fba669a8is_scripting() guard on checkpoint_seq
Ross Wightman
2023-02-04 14:21:49 -0800
95ec255f7fFinish timm mode api for efficientformer_v2, add grad checkpointing support to both efficientformers
Ross Wightman
2023-02-03 21:21:23 -0800
9d03c6f526Merge remote-tracking branch 'origin/main' into levit_efficientformer_redux
Ross Wightman
2023-02-03 14:47:01 -0800
086bd55a94Add EfficientFormer-V2, refactor EfficientFormer and Levit for more uniformity across the 3 related arch. Add features_out support to levit conv models and efficientformer_v2. All weights on hub.
Ross Wightman
2023-02-03 14:12:29 -0800
2cb2699dc8Apply fix from #1649 to main
Ross Wightman
2023-02-03 11:28:57 -0800
6f28b562c6Factor NormMlpClassifierHead from MaxxViT and use across MaxxViT / ConvNeXt / DaViT, refactor some type hints & comments
Ross Wightman
2023-01-27 14:57:01 -0800
29fda20e6dMerge branch 'fffffgggg54-main'
Ross Wightman
2023-01-27 13:55:17 -0800
9a53c3f727Finalize DaViT, some formatting and modelling simplifications (separate PatchEmbed to Stem + Downsample, weights on HF hub.
Ross Wightman
2023-01-27 13:54:04 -0800