Deployed 80c3051 with MkDocs version: 1.1.2

gh-pages
Ross Wightman 4 years ago
parent 4f6e4af795
commit 7ec600e0f5

@ -549,15 +549,17 @@
<h3 id="feb-29-2020">Feb 29, 2020</h3>
<ul>
<li>New MobileNet-V3 Large weights trained from stratch with this code to 75.77% top-1</li>
<li>IMPORTANT CHANGE - default weight init changed for all MobilenetV3 / EfficientNet / related models</li>
<li>IMPORTANT CHANGE - default weight init changed for all MobilenetV3 / EfficientNet / related models<ul>
<li>overall results similar to a bit better training from scratch on a few smaller models tried</li>
<li>performance early in training seems consistently improved but less difference by end</li>
<li>set <code>fix_group_fanout=False</code> in <code>_init_weight_goog</code> fn if you need to reproducte past behaviour</li>
</ul>
</li>
<li>Experimental LR noise feature added applies a random perturbation to LR each epoch in specified range of training</li>
</ul>
<h3 id="feb-18-2020">Feb 18, 2020</h3>
<ul>
<li>Big refactor of model layers and addition of several attention mechanisms. Several additions motivated by 'Compounding the Performance Improvements...' (<a href="https://arxiv.org/abs/2001.06268">https://arxiv.org/abs/2001.06268</a>):</li>
<li>Big refactor of model layers and addition of several attention mechanisms. Several additions motivated by 'Compounding the Performance Improvements...' (<a href="https://arxiv.org/abs/2001.06268">https://arxiv.org/abs/2001.06268</a>):<ul>
<li>Move layer/module impl into <code>layers</code> subfolder/module of <code>models</code> and organize in a more granular fashion</li>
<li>ResNet downsample paths now properly support dilation (output stride != 32) for avg_pool ('D' variant) and 3x3 (SENets) networks</li>
<li>Add Selective Kernel Nets on top of ResNet base, pretrained weights<ul>
@ -570,6 +572,8 @@
<li>CBAM attention experiment (not the best results so far, may remove)</li>
<li>Attention factory to allow dynamically selecting one of SE, ECA, CBAM in the <code>.se</code> position for all ResNets</li>
<li>Add DropBlock and DropPath (formerly DropConnect for EfficientNet/MobileNetv3) support to all ResNet variants</li>
</ul>
</li>
<li>Full dataset results updated that incl NoisyStudent weights and 2 of the 3 SK weights</li>
</ul>
<h3 id="feb-12-2020">Feb 12, 2020</h3>
@ -609,7 +613,7 @@
</ul>
<h3 id="dec-28-2019">Dec 28, 2019</h3>
<ul>
<li>Add new model weights and training hparams (see Training Hparams section)</li>
<li>Add new model weights and training hparams (see Training Hparams section)<ul>
<li><code>efficientnet_b3</code> - 81.5 top-1, 95.7 top-5 at default res/crop, 81.9, 95.8 at 320x320 1.0 crop-pct<ul>
<li>trained with RandAugment, ended up with an interesting but less than perfect result (see training section)</li>
</ul>
@ -625,6 +629,8 @@
</ul>
</li>
</ul>
</li>
</ul>
<h3 id="dec-23-2019">Dec 23, 2019</h3>
<ul>
<li>Add RandAugment trained MixNet-XL weights with 80.48 top-1.</li>
@ -636,13 +642,17 @@
</ul>
<h3 id="nov-29-2019">Nov 29, 2019</h3>
<ul>
<li>Brought EfficientNet and MobileNetV3 up to date with my <a href="https://github.com/rwightman/gen-efficientnet-pytorch">https://github.com/rwightman/gen-efficientnet-pytorch</a> code. Torchscript and ONNX export compat excluded.</li>
<li>Brought EfficientNet and MobileNetV3 up to date with my <a href="https://github.com/rwightman/gen-efficientnet-pytorch">https://github.com/rwightman/gen-efficientnet-pytorch</a> code. Torchscript and ONNX export compat excluded.<ul>
<li>AdvProp weights added</li>
<li>Official TF MobileNetv3 weights added</li>
</ul>
</li>
<li>EfficientNet and MobileNetV3 hook based 'feature extraction' classes added. Will serve as basis for using models as backbones in obj detection/segmentation tasks. Lots more to be done here...</li>
<li>HRNet classification models and weights added from <a href="https://github.com/HRNet/HRNet-Image-Classification">https://github.com/HRNet/HRNet-Image-Classification</a></li>
<li>Consistency in global pooling, <code>reset_classifer</code>, and <code>forward_features</code> across models</li>
<li>Consistency in global pooling, <code>reset_classifer</code>, and <code>forward_features</code> across models<ul>
<li><code>forward_features</code> always returns unpooled feature maps now</li>
</ul>
</li>
<li>Reasonable chance I broke something... let me know</li>
</ul>
<h3 id="nov-22-2019">Nov 22, 2019</h3>

@ -449,28 +449,36 @@
<h1 id="recent-changes">Recent Changes</h1>
<h3 id="aug-1-2020">Aug 1, 2020</h3>
<p>Universal feature extraction, new models, new weights, new test sets.
* All models support the <code>features_only=True</code> argument for <code>create_model</code> call to return a network that extracts features from the deepest layer at each stride.
* New models
* CSPResNet, CSPResNeXt, CSPDarkNet, DarkNet
* ReXNet
* (Aligned) Xception41/65/71 (a proper port of TF models)
* New trained weights
* SEResNet50 - 80.3
* CSPDarkNet53 - 80.1 top-1
* CSPResNeXt50 - 80.0 to-1
* DPN68b - 79.2 top-1
* EfficientNet-Lite0 (non-TF ver) - 75.5 (submitted by @hal-314)
* Add 'real' labels for ImageNet and ImageNet-Renditions test set, see <a href="results/README.md"><code>results/README.md</code></a></p>
<p>Universal feature extraction, new models, new weights, new test sets.</p>
<ul>
<li>All models support the <code>features_only=True</code> argument for <code>create_model</code> call to return a network that extracts features from the deepest layer at each stride.</li>
<li>New models<ul>
<li>CSPResNet, CSPResNeXt, CSPDarkNet, DarkNet</li>
<li>ReXNet</li>
<li>(Aligned) Xception41/65/71 (a proper port of TF models)</li>
</ul>
</li>
<li>New trained weights<ul>
<li>SEResNet50 - 80.3</li>
<li>CSPDarkNet53 - 80.1 top-1</li>
<li>CSPResNeXt50 - 80.0 to-1</li>
<li>DPN68b - 79.2 top-1</li>
<li>EfficientNet-Lite0 (non-TF ver) - 75.5 (submitted by @hal-314)</li>
</ul>
</li>
<li>Add 'real' labels for ImageNet and ImageNet-Renditions test set, see <a href="results/README.md"><code>results/README.md</code></a></li>
</ul>
<h3 id="june-11-2020">June 11, 2020</h3>
<p>Bunch of changes:</p>
<ul>
<li>DenseNet models updated with memory efficient addition from torchvision (fixed a bug), blur pooling and deep stem additions</li>
<li>VoVNet V1 and V2 models added, 39 V2 variant (ese_vovnet_39b) trained to 79.3 top-1</li>
<li>Activation factory added along with new activations:</li>
<li>Activation factory added along with new activations:<ul>
<li>select act at model creation time for more flexibility in using activations compatible with scripting or tracing (ONNX export)</li>
<li>hard_mish (experimental) added with memory-efficient grad, along with ME hard_swish</li>
<li>context mgr for setting exportable/scriptable/no_jit states</li>
</ul>
</li>
<li>Norm + Activation combo layers added with initial trial support in DenseNet and VoVNet along with impl of EvoNorm and InplaceAbn wrapper that fit the interface</li>
<li>Torchscript works for all but two of the model types as long as using Pytorch 1.5+, tests added for this</li>
<li>Some import cleanup and classifier reset changes, all models will have classifier reset to nn.Identity on reset_classifer(0) call</li>
@ -486,20 +494,24 @@
</ul>
<h3 id="may-1-2020">May 1, 2020</h3>
<ul>
<li>Merged a number of execellent contributions in the ResNet model family over the past month</li>
<li>Merged a number of execellent contributions in the ResNet model family over the past month<ul>
<li>BlurPool2D and resnetblur models initiated by <a href="https://github.com/VRandme">Chris Ha</a>, I trained resnetblur50 to 79.3.</li>
<li>TResNet models and SpaceToDepth, AntiAliasDownsampleLayer layers by <a href="https://github.com/mrT23">mrT23</a></li>
<li>ecaresnet (50d, 101d, light) models and two pruned variants using pruning as per (<a href="https://arxiv.org/abs/2002.08258">https://arxiv.org/abs/2002.08258</a>) by <a href="https://github.com/yoniaflalo">Yonathan Aflalo</a></li>
</ul>
</li>
<li>200 pretrained models in total now with updated results csv in results folder</li>
</ul>
<h3 id="april-5-2020">April 5, 2020</h3>
<ul>
<li>Add some newly trained MobileNet-V2 models trained with latest h-params, rand augment. They compare quite favourably to EfficientNet-Lite</li>
<li>Add some newly trained MobileNet-V2 models trained with latest h-params, rand augment. They compare quite favourably to EfficientNet-Lite<ul>
<li>3.5M param MobileNet-V2 100 @ 73%</li>
<li>4.5M param MobileNet-V2 110d @ 75%</li>
<li>6.1M param MobileNet-V2 140 @ 76.5%</li>
<li>5.8M param MobileNet-V2 120d @ 77.3%</li>
</ul>
</li>
</ul>
<h3 id="march-18-2020">March 18, 2020</h3>
<ul>
<li>Add EfficientNet-Lite models w/ weights ported from <a href="https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet/lite">Tensorflow TPU</a></li>

@ -701,10 +701,12 @@
<h1 id="model-architectures">Model Architectures</h1>
<p>The model architectures included come from a wide variety of sources. Sources, including papers, original impl ("reference code") that I rewrote / adapted, and PyTorch impl that I leveraged directly ("code") are listed below.</p>
<p>Most included models have pretrained weights. The weights are either:
1. from their original sources
2. ported by myself from their original impl in a different framework (e.g. Tensorflow models)
3. trained from scratch using the included training script</p>
<p>Most included models have pretrained weights. The weights are either:</p>
<ol>
<li>from their original sources</li>
<li>ported by myself from their original impl in a different framework (e.g. Tensorflow models)</li>
<li>trained from scratch using the included training script</li>
</ol>
<p>The validation results for the pretrained weights can be found <a href="../results/">here</a></p>
<h2 id="cross-stage-partial-networks-cspnetpy">Cross-Stage Partial Networks [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/cspnet.py">cspnet.py</a>]</h2>
<ul>
@ -763,7 +765,7 @@
</ul>
<h2 id="efficientnet-efficientnetpy">EfficientNet [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/efficientnet.py">efficientnet.py</a>]</h2>
<ul>
<li>Papers<ul>
<li>Papers:<ul>
<li>EfficientNet NoisyStudent (B0-B7, L2) - <a href="https://arxiv.org/abs/1911.04252">https://arxiv.org/abs/1911.04252</a></li>
<li>EfficientNet AdvProp (B0-B8) - <a href="https://arxiv.org/abs/1911.09665">https://arxiv.org/abs/1911.09665</a></li>
<li>EfficientNet (B0-B7) - <a href="https://arxiv.org/abs/1905.11946">https://arxiv.org/abs/1905.11946</a></li>
@ -852,9 +854,11 @@
<li>Code: <a href="https://github.com/mehtadushy/SelecSLS-Pytorch">https://github.com/mehtadushy/SelecSLS-Pytorch</a></li>
</ul>
<h2 id="squeeze-and-excitation-networks-senetpy">Squeeze-and-Excitation Networks [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/senet.py">senet.py</a>]</h2>
<p>NOTE: I am deprecating this version of the networks, the new ones are part of <code>resnet.py</code>
* Paper: <code>Squeeze-and-Excitation Networks</code> - <a href="https://arxiv.org/abs/1709.01507">https://arxiv.org/abs/1709.01507</a>
* Code: <a href="https://github.com/Cadene/pretrained-models.pytorch">https://github.com/Cadene/pretrained-models.pytorch</a> </p>
<p>NOTE: I am deprecating this version of the networks, the new ones are part of <code>resnet.py</code></p>
<ul>
<li>Paper: <code>Squeeze-and-Excitation Networks</code> - <a href="https://arxiv.org/abs/1709.01507">https://arxiv.org/abs/1709.01507</a></li>
<li>Code: <a href="https://github.com/Cadene/pretrained-models.pytorch">https://github.com/Cadene/pretrained-models.pytorch</a> </li>
</ul>
<h2 id="tresnet-tresnetpy">TResNet [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/tresnet.py">tresnet.py</a>]</h2>
<ul>
<li>Paper: <code>TResNet: High Performance GPU-Dedicated Architecture</code> - <a href="https://arxiv.org/abs/2003.13630">https://arxiv.org/abs/2003.13630</a></li>

File diff suppressed because one or more lines are too long

Binary file not shown.
Loading…
Cancel
Save