Chris Ha
acd1b6cccd
Implement Functional Blur on resnet.py
...
1. add ResNet argument blur=''
2. implement blur for maxpool and strided convs in downsampling blocks
5 years ago
Chris Ha
ce3d82b58b
Update blurpool.py
...
clean up code for PR
5 years ago
Chris Ha
3a287a6e76
Create blurpool.py
...
Initial implementation of blur layer.
currently tests as correct against Downsample of original github
5 years ago
Ross Wightman
c99a5abed4
Merge pull request #115 from rwightman/mobilenetv2-experiment
...
MobileNet-V2 experiments
5 years ago
Ross Wightman
e34074b4da
Add final weights for MobileNet-V2 experiments
5 years ago
Ross Wightman
a6a5565de7
Fiddling...
5 years ago
Ross Wightman
5a16c533ff
Add better resnext50_32x4d weights trained by andravin
5 years ago
Ross Wightman
bc998cad91
Experimenting with some MobileNetV2 variations to compare against EfficientNet-Lite
5 years ago
Ross Wightman
3406e582cf
Add EfficientNet-Lite results, update README
5 years ago
Ross Wightman
bd05258f7b
EfficientNet-Lite model added w/ converted checkpoints, validation in progress...
5 years ago
Andrew Lavin
d92cc4da54
Modified save_checkpoint to always save last checkpoint. Fixes #98 .
5 years ago
Ross Wightman
56e2ac3a6d
Merge pull request #94 from rwightman/lr_noise
...
Learning rate noise, MobileNetV3 weights, and activate MobileNetV3/EfficientNet weight init change
5 years ago
Ross Wightman
c16f25ced2
Add MobileNetV3 Large weights, results, update README and sotabench for merge
5 years ago
Ross Wightman
c60069c1eb
Annotate types on drop fns to avoid torchscript error
5 years ago
Ross Wightman
9fee316752
Enable fixed fanout calc in EfficientNet/MobileNetV3 weight init by default. Fix #84
5 years ago
Ross Wightman
27b3680d49
Revamp LR noise, move logic to scheduler base. Fixup PlateauLRScheduler and add it as an option.
5 years ago
Ross Wightman
cc5a11abba
version bump for PyPi update
5 years ago
Ross Wightman
514b0938c4
Experimenting with per-epoch learning rate noise w/ step scheduler
5 years ago
Ross Wightman
43225d110c
Unify drop connect vs drop path under 'drop path' name, switch all EfficientNet/MobilenetV3 refs to 'drop_path'. Update factory to handle new drop args.
5 years ago
Ross Wightman
f1d5f8a6c4
Update comments for Selective Kernel and DropBlock/Path impl, add skresnet34 weights
5 years ago
Ross Wightman
569419b38d
Tweak some comments, add SKNet models with weights to sotabench, remove an unused branch
5 years ago
Ross Wightman
53c47479c4
Batch validation batch size adjustment, tweak L2 crop pct
5 years ago
Ross Wightman
08553e16b3
Merge branch 'master' into attention
5 years ago
Ross Wightman
fa38f24967
Update SK network configs, add weights for skresnet8 and skresnext50
5 years ago
Ross Wightman
f098fda2ca
Add map_location='cpu' to ModelEma resume, should improve #72
5 years ago
Ross Wightman
ba15ca47e8
Add ported EfficientNet-L2, B0-B7 NoisyStudent weights from TF TPU
5 years ago
Ross Wightman
5e6dbbaf30
Add CBAM for experimentation
5 years ago
Ross Wightman
d725991870
Remove debug print from ECA module
5 years ago
Ross Wightman
2a7d256fd5
Re-enable mem-efficient/jit activations after torchscript tests
5 years ago
Ross Wightman
f902bcd54c
Layer refactoring continues, ResNet downsample rewrite for proper dilation in 3x3 and avg_pool cases
...
* select_conv2d -> create_conv2d
* added create_attn to create attention module from string/bool/module
* factor padding helpers into own file, use in both conv2d_same and avg_pool2d_same
* add some more test eca resnet variants
* minor tweaks, naming, comments, consistency
5 years ago
Ross Wightman
a99ec4e7d1
A bunch more layer reorg, splitting many layers into own files. Improve torchscript compatibility.
5 years ago
Ross Wightman
13746a33fc
Big move, layer modules and fn to timm/models/layers
5 years ago
Ross Wightman
f54612f648
Merge branch 'select_kernel' into attention
5 years ago
Ross Wightman
4defbbbaa8
Fix module name mistake, start layers sub-package
5 years ago
Ross Wightman
7011cd0902
A little bit of ECA cleanup
5 years ago
Ross Wightman
46471df7b2
Merge pull request #82 from VRandme/eca
...
ECA-Net Efficient Channel Attention
5 years ago
Ross Wightman
d0eb59ef46
Remove unused default_init for EfficientNets, experimenting with fanout calc for #84
5 years ago
Chris Ha
e6a762346a
Implement Adaptive Kernel selection
...
When channel size is given,
calculate adaptive kernel size according to original paper.
Otherwise use the given kernel size(k_size), which defaults to 3
5 years ago
Ross Wightman
13e8da2b46
SelectKernel split_input works best when input channels split like grouped conv, but output is full width. Disable zero_init for SK nets, seems a bad combo.
5 years ago
Chris Ha
6db087a1ff
Merge remote-tracking branch 'upstream/master' into eca
5 years ago
Chris Ha
904c618040
Update EcaModule.py
...
Make pylint happy
(commas, unused imports, missed imports)
5 years ago
Chris Ha
db91ba053b
EcaModule(CamelCase)
...
CamelCased EcaModule.
Renamed all instances of ecalayer to EcaModule.
eca_module.py->EcaModule.py
5 years ago
Ross Wightman
5c4991a088
Add PyTorch trained EfficientNet-ES weights from Andrew Lavin
5 years ago
Chris Ha
d04ff95eda
Merge branch 'master' into eca
5 years ago
Chris Ha
d63ae121d5
Clean up eca_module code
...
functionally similar
adjusted rwightman's version of reshaping and viewing.
Use F.pad for circular eca version for cleaner code
5 years ago
Ross Wightman
d66819d1f3
Indentation mistake. Fixes #81
5 years ago
Chris Ha
f87fcd7e88
Implement Eca modules
...
implement ECA module by
1. adopting original eca_module.py into models folder
2. adding use_eca layer besides every instance of SE layer
5 years ago
Ross Wightman
4808b3c32f
Bump version for PyPi update, fix few out of date README items/mistakes, add README updates for TF EfficientNet-B8 (RandAugment)
5 years ago
Ross Wightman
7d07ebb660
Adding some configs to sknet, incl ResNet50 variants from 'Compounding ... Assembled Techniques' paper and original SKNet50
5 years ago
Ross Wightman
a9d2424fd1
Add separate zero_init_last_bn function to support more block variety without a mess
5 years ago
Ross Wightman
355aa152d5
Just leave it float for now, will look at fp16 later. Remove unused reference code.
5 years ago
Ross Wightman
ef457555d3
BlockDrop working on GPU
5 years ago
Ross Wightman
3ff19079f9
Missed nn_ops.py from last commit
5 years ago
Ross Wightman
9f11b4e8a2
Add ConvBnAct layer to parallel integrated SelectKernelConv, add support for DropPath and DropBlock to ResNet base and SK blocks
5 years ago
Ross Wightman
cefc9b7761
Move SelectKernelConv to conv2d_layers and more
...
* always apply attention in SelectKernelConv, leave MixedConv for no attention alternative
* make MixedConv torchscript compatible
* refactor first/previous dilation name to make more sense in ResNet* networks
5 years ago
Ross Wightman
9abe610931
Used wrong channel var for split
5 years ago
Ross Wightman
58e28dc7e7
Move Selective Kernel blocks/convs to their own sknet.py file
5 years ago
Ross Wightman
a93bae6dc5
A SelectiveKernelBasicBlock for more experiments
5 years ago
Ross Wightman
ad087b4b17
Missed bias=False in selection conv
5 years ago
Ross Wightman
c8b3d6b81a
Initial impl of Selective Kernel Networks. Very much a WIP.
5 years ago
Ross Wightman
1daa303744
Add support to Dataset for class id mapping file, clean up a bit of old logic. Add results file arg for validation and update script.
5 years ago
Ross Wightman
91534522f9
Add newly added TF ported EfficientNet-B8 weights (RandAugment)
5 years ago
Ross Wightman
12dbc74742
New ResNet50 JSD + RandAugment weights
5 years ago
Ross Wightman
2f41905ba5
Update ResNet50 weights to AuxMix trained 78.994 top-1. A few commentes re 'tiered_narrow' tn variant.
5 years ago
Ross Wightman
d9a6a9d0af
Merge pull request #74 from rwightman/augmix-jsd
...
AugMix, JSD loss, SplitBatchNorm (Auxiliary BN), and more
5 years ago
Ross Wightman
3eb4a96eda
Update AugMix, JSD, etc comments and references
5 years ago
Ross Wightman
a28117ea46
Add tiered narrow ResNet (tn) and weights for seresnext26tn_32x4d
5 years ago
Ross Wightman
833066b540
A few minor things in SplitBN
5 years ago
Ross Wightman
7547119891
Add SplitBatchNorm. AugMix, Rand/AutoAugment, Split (Aux) BatchNorm, Jensen-Shannon Divergence, RandomErasing all working together
5 years ago
Ross Wightman
2e955cfd0c
Update RandomErasing with some improved arg names, tweak to aspect range
5 years ago
Ross Wightman
3cc0f91e23
Fix augmix variable name scope overlap, default non-blended mode
5 years ago
Ross Wightman
ec0dd4053a
Add updated RandAugment trained EfficientNet-B0 trained weights from @michaelklachko
5 years ago
Ross Wightman
40fea63ebe
Add checkpoint averaging script. Add headers, shebangs, exec perms to all scripts
5 years ago
Ross Wightman
4666cc9aed
Add --pin-mem arg to enable dataloader pin_memory (showing more benefit in some scenarios now), also add --torchscript arg to validate.py for testing models with jit.script
5 years ago
Ross Wightman
53001dd292
ResNet / Res2Net additions:
...
* ResNet torchscript compat
* output_stride arg supported to limit network stride via dilations (support for dilation added to Res2Net)
* allow activation layer to be changed via act_layer arg
5 years ago
Ross Wightman
f96b3e5e92
InceptionResNetV2 torchscript compatible
5 years ago
Ross Wightman
19d93fe454
Add selecsls60 weights
5 years ago
Ross Wightman
0062c15fb0
Update checkpoint url with modelzoo compatible ones.
5 years ago
Ross Wightman
b5315e66b5
Streamline SelecSLS model without breaking checkpoint compat. Move cfg handling out of model class. Update feature/pooling behaviour to match current.
5 years ago
Ross Wightman
d59a756c16
Run PyCharm autoformat on selecsls and change mix cap variables and model names to all lower
5 years ago
Ross Wightman
fb3a0f4bb8
Merge pull request #65 from mehtadushy/selecsls
...
Incorporate SelecSLS Models
5 years ago
Ross Wightman
19fc205a4d
Update comments on the new SE-ResNeXt26 models
5 years ago
Ross Wightman
acc3ed2b8c
Add EfficientNet-B3 weights, trained from scratch with RA.
5 years ago
Dushyant Mehta
2404361f62
correct asset paths
5 years ago
Dushyant Mehta
31939311f6
Added SelecSLS models
5 years ago
rwightman
1f4498f217
Add ResNet deep tiered stem and model weights for seresnext26t_32x4d and seresnext26d_32x4d
5 years ago
Dushyant Mehta
32012a44fd
Added SelecSLS model
5 years ago
Ross Wightman
73b78459dc
Add update RandAugment MixNet-XL weights
5 years ago
Ross Wightman
3afc2a4dc0
Some cleanup/improvements to AugMix impl:
...
* make 'increasing' levels for Contrast, Color, Brightness, Saturation ops
* remove recursion from faster blending mix
* add config striing parsing for AugMix
5 years ago
Ross Wightman
232ab7fb12
Working on an implementation of AugMix with JensenShannonDivergence loss that's compatible with my AutoAugment and RandAugment impl
5 years ago
Ross Wightman
a435ea1327
Change reduce_bn to distribute_bn, add ability to choose between broadcast and reduce (mean). Add crop_pct arg to allow selecting validation crop while training.
5 years ago
Ross Wightman
3bff2b21dc
Add support for keeping running bn stats the same across distributed training nodes before eval/save
5 years ago
Ross Wightman
0161de0127
Switch RandoErasing back to on GPU normal sampling
5 years ago
Ross Wightman
ff421e5e09
New PyTorch trained EfficientNet-B2 weights with my RandAugment impl
5 years ago
Ross Wightman
3bef524f9c
Finish with HRNet, weights and models updated. Improve consistency in model classifier/global pool treatment.
5 years ago
Ross Wightman
6ca0828166
Update EfficientNet comments, MobileNetV3 non-TF create fns, fix factory arg checks, bump PyTorch version req to 1.2
5 years ago
Ross Wightman
eccbadca74
Update EfficientNet comments
5 years ago
Ross Wightman
902d32fb16
Renamed gen_efficientnet.py -> efficientnet.py
5 years ago
Ross Wightman
5a0a8de7e3
ResNet updates:
...
* remove redundant GluonResNet model/blocks and use the code in ResNet for Gluon weights
* change SEModules back to using AdaptiveAvgPool instead of mean, PyTorch issue long fixed
5 years ago
Ross Wightman
a39cc43374
Bring EfficientNet and MobileNetV3 up to date with my gen-efficientnet repo
...
* Split MobileNetV3 and EfficientNet model files and put builder and blocks in own files (getting too large)
* Finalize CondConv EfficientNet variant
* Add the AdvProp weights files and B8 EfficientNet model
* Refine the feature extraction module for EfficientNet and MobileNetV3
5 years ago
Ross Wightman
ad93347548
Initial HRNet classification model commit
5 years ago
Ross Wightman
2393708650
Missed stashing of out_indices in model
5 years ago
Ross Wightman
35e8f0c5e7
Fixup a few comments, add PyTorch version aware Flatten and finish as_sequential for GenEfficientNet
5 years ago
Ross Wightman
7ac6db4543
Missed activations.py
5 years ago
Ross Wightman
506df0e3d0
Add CondConv support for EfficientNet into WIP for GenEfficientNet Feature extraction setup
5 years ago
Ross Wightman
576d360f20
Bring in JIT version of optimized swish activation from gen_efficientnet as default (while working on feature extraction functionality here).
5 years ago
Ross Wightman
7b83e67f77
Pass drop connect arg through to EfficientNet models
5 years ago
Ross Wightman
31453b039e
Update Auto/RandAugment comments, README, more.
...
* Add a weighted choice option for RandAugment
* Adjust magnitude noise/std naming, config
5 years ago
Ross Wightman
4243f076f1
Adding RandAugment to AutoAugment impl, some tweaks to AA included
5 years ago
Ross Wightman
0d58c50fb1
Add TF RandAug weights for B5/B7 EfficientNet models.
5 years ago
Ross Wightman
c099374771
Map pretrained checkpoint to cpu to avoid issue with some pretrained checkpoints still having CUDA tensors. Fixes #42
5 years ago
Ross Wightman
b93fcf0708
Add Facebook Research Semi-Supervised and Semi-Weakly Supervised ResNet model weights.
5 years ago
Ross Wightman
a9eb484835
Add memory efficient Swish impl
5 years ago
rwightman
d3ba34ee7e
Fix Mobilenet V3 model name for sotabench. Minor res2net cleanup.
5 years ago
Ross Wightman
2680ad14bb
Add Res2Net and DLA to README
5 years ago
rwightman
adbf770f16
Add Res2Net and DLA models w/ pretrained weights. Update sotabench.
5 years ago
Ross Wightman
4002c0d4ce
Fix AutoAugment abs translate calc
5 years ago
Ross Wightman
c06274e5a2
Add note on random selection of magnitude value
5 years ago
Ross Wightman
b750b76f67
More AutoAugment work. Ready to roll...
5 years ago
Ross Wightman
25d2088d9e
Working on auto-augment
5 years ago
Ross Wightman
aff194f42c
Merge pull request #32 from rwightman/opt
...
More optimizer work
5 years ago
Ross Wightman
64966f61f7
Add Nvidia's NovogGrad impl from Jasper (cleaner/faster than current) and Apex Fused optimizers
5 years ago
Ross Wightman
3d9c8a6489
Add support for new AMP checkpointing support w/ amp.state_dict
5 years ago
Ross Wightman
ba3c97c3ad
Some Lookahead cleanup and fixes
5 years ago
Ross Wightman
e9d2ec4d8e
Merge pull request #31 from rwightman/opt
...
Optimizers and more
5 years ago
Ross Wightman
fac58f609a
Add RAdam, NovoGrad, Lookahead, and AdamW optimizers, a few ResNet tweaks and scheduler factory tweak.
...
* Add some of the trendy new optimizers. Decent results but not clearly better than the standards.
* Can create a None scheduler for constant LR
* ResNet defaults to zero_init of last BN in residual
* add resnet50d config
5 years ago
Ross Wightman
81875d52a6
Update sotabench model list, add Mean-Max pooling DPN variants, disable download progress
5 years ago
Ross Wightman
f37e633e9b
Merge remote-tracking branch 'origin/re-exp' into opt
5 years ago
Ross Wightman
b06dce8d71
Bump version for next push to pypi
5 years ago
Ross Wightman
73fbd97ed4
Add weights for my MixNet-XL creation, include README updates for EdgeTPU models
5 years ago
Ross Wightman
51a2375b0c
Experimenting with a custom MixNet-XL and MixNet-XXL definition
5 years ago
Ross Wightman
9ec6824bab
Finally got around to adding EdgeTPU EfficientNet variant
5 years ago
Ross Wightman
daeaa113e2
Add initial sotabench attempt. Split create_transform out of create_loader. Update requirements.txt
5 years ago
Ross Wightman
66634d2200
Add support to split random erasing blocks into randomly selected number with --recount arg. Fix random selection of aspect ratios.
5 years ago
Ross Wightman
6946281fde
Experimenting with random erasing changes
5 years ago
Ross Wightman
aeaaad7304
Merge pull request #24 from rwightman/gluon_xception
...
Port Gluon Aligned Xception models
5 years ago
Ross Wightman
3b4868f6dc
A few more additions to Gluon Xception models to match interface of others.
5 years ago
Ross Wightman
4d505e0785
Add working Gluon Xception-65 model. Some cleanup still needed.
5 years ago
Minqin Chen
4e7a854dd0
Update helpers.py
...
Fixing out of memory error by loading the checkpoint onto the CPU.
5 years ago
Ross Wightman
0c874195db
Update results csv files, bump version for timm pip release
5 years ago
Ross Wightman
4fe2da558c
Add MixNet Small and Large PyTorch native weights (no same padding)
5 years ago
Ross Wightman
e879cf52fa
Update validation scores for new TF EfficientNet weights.
5 years ago
Ross Wightman
77e2e0c4e3
Add new auto-augmentation Tensorflow EfficientNet weights, incl B6 and B7 models. Validation scores still pending but looking good.
5 years ago
Ross Wightman
857f33015a
Add native PyTorch weights for MixNet-Medium with no SAME padding necessary. Remove unused block of code.
5 years ago
Ross Wightman
e7c8a37334
Make min-lr and cooldown-epochs cmdline args, change dash in color_jitter arg for consistency
5 years ago
Ross Wightman
d4debe6597
Update version, results csv files, and move remaining dropbox weights to github
5 years ago
Ross Wightman
dfa9298b4e
Add MixNet ( https://arxiv.org/abs/1907.09595 ) with pretrained weights converted from Tensorflow impl
...
* refactor 'same' convolution and add helper to use MixedConv2d when needed
* improve performance of 'same' padding for cases that can be handled statically
* add support for extra exp, pw, and dw kernel specs with grouping support to decoder/string defs for MixNet
* shuffle some args for a bit more consistency, a little less clutter overall in gen_efficientnet.py
5 years ago
Ross Wightman
7a92caa560
Add basic image folder style dataset to read directly out of tar files, example in validate.py
5 years ago
Ross Wightman
d6ac5bbc48
EfficientNet and related cleanup
...
* remove folded_bn support and corresponding untrainable tflite ported weights
* combine bn args into dict
* add inplace support to activations and use where possible for reduced mem on large models
5 years ago
Ross Wightman
3d9be78fc6
A bit more ResNet cleanup.
...
* add inplace=True back
* minor comment improvements
* few clarity changes
5 years ago