Deployed d8e6920 with MkDocs version: 1.1.2

gh-pages
Ross Wightman 4 years ago
parent da8447c4ae
commit 550a478fa5

@ -297,6 +297,27 @@
</label> </label>
<ul class="md-nav__list" data-md-scrollfix> <ul class="md-nav__list" data-md-scrollfix>
<li class="md-nav__item">
<a href="#may-12-2020" class="md-nav__link">
May 12, 2020
</a>
</li>
<li class="md-nav__item">
<a href="#may-3-2020" class="md-nav__link">
May 3, 2020
</a>
</li>
<li class="md-nav__item">
<a href="#may-1-2020" class="md-nav__link">
May 1, 2020
</a>
</li>
<li class="md-nav__item"> <li class="md-nav__item">
<a href="#april-5-2020" class="md-nav__link"> <a href="#april-5-2020" class="md-nav__link">
April 5, 2020 April 5, 2020
@ -309,6 +330,20 @@
March 18, 2020 March 18, 2020
</a> </a>
</li>
<li class="md-nav__item">
<a href="#april-5-2020_1" class="md-nav__link">
April 5, 2020
</a>
</li>
<li class="md-nav__item">
<a href="#march-18-2020_1" class="md-nav__link">
March 18, 2020
</a>
</li> </li>
<li class="md-nav__item"> <li class="md-nav__item">
@ -441,6 +476,27 @@
</label> </label>
<ul class="md-nav__list" data-md-scrollfix> <ul class="md-nav__list" data-md-scrollfix>
<li class="md-nav__item">
<a href="#may-12-2020" class="md-nav__link">
May 12, 2020
</a>
</li>
<li class="md-nav__item">
<a href="#may-3-2020" class="md-nav__link">
May 3, 2020
</a>
</li>
<li class="md-nav__item">
<a href="#may-1-2020" class="md-nav__link">
May 1, 2020
</a>
</li>
<li class="md-nav__item"> <li class="md-nav__item">
<a href="#april-5-2020" class="md-nav__link"> <a href="#april-5-2020" class="md-nav__link">
April 5, 2020 April 5, 2020
@ -453,6 +509,20 @@
March 18, 2020 March 18, 2020
</a> </a>
</li>
<li class="md-nav__item">
<a href="#april-5-2020_1" class="md-nav__link">
April 5, 2020
</a>
</li>
<li class="md-nav__item">
<a href="#march-18-2020_1" class="md-nav__link">
March 18, 2020
</a>
</li> </li>
<li class="md-nav__item"> <li class="md-nav__item">
@ -574,6 +644,24 @@
<h1 id="archived-changes">Archived Changes</h1> <h1 id="archived-changes">Archived Changes</h1>
<h3 id="may-12-2020">May 12, 2020</h3>
<ul>
<li>Add ResNeSt models (code adapted from <a href="https://github.com/zhanghang1989/ResNeSt">https://github.com/zhanghang1989/ResNeSt</a>, paper <a href="https://arxiv.org/abs/2004.08955">https://arxiv.org/abs/2004.08955</a>))</li>
</ul>
<h3 id="may-3-2020">May 3, 2020</h3>
<ul>
<li>Pruned EfficientNet B1, B2, and B3 (<a href="https://arxiv.org/abs/2002.08258">https://arxiv.org/abs/2002.08258</a>) contributed by <a href="https://github.com/yoniaflalo">Yonathan Aflalo</a></li>
</ul>
<h3 id="may-1-2020">May 1, 2020</h3>
<ul>
<li>Merged a number of execellent contributions in the ResNet model family over the past month<ul>
<li>BlurPool2D and resnetblur models initiated by <a href="https://github.com/VRandme">Chris Ha</a>, I trained resnetblur50 to 79.3.</li>
<li>TResNet models and SpaceToDepth, AntiAliasDownsampleLayer layers by <a href="https://github.com/mrT23">mrT23</a></li>
<li>ecaresnet (50d, 101d, light) models and two pruned variants using pruning as per (<a href="https://arxiv.org/abs/2002.08258">https://arxiv.org/abs/2002.08258</a>) by <a href="https://github.com/yoniaflalo">Yonathan Aflalo</a></li>
</ul>
</li>
<li>200 pretrained models in total now with updated results csv in results folder</li>
</ul>
<h3 id="april-5-2020">April 5, 2020</h3> <h3 id="april-5-2020">April 5, 2020</h3>
<ul> <ul>
<li>Add some newly trained MobileNet-V2 models trained with latest h-params, rand augment. They compare quite favourably to EfficientNet-Lite<ul> <li>Add some newly trained MobileNet-V2 models trained with latest h-params, rand augment. They compare quite favourably to EfficientNet-Lite<ul>
@ -589,6 +677,21 @@
<li>Add EfficientNet-Lite models w/ weights ported from <a href="https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet/lite">Tensorflow TPU</a></li> <li>Add EfficientNet-Lite models w/ weights ported from <a href="https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet/lite">Tensorflow TPU</a></li>
<li>Add RandAugment trained ResNeXt-50 32x4d weights with 79.8 top-1. Trained by <a href="https://github.com/andravin">Andrew Lavin</a> (see Training section for hparams)</li> <li>Add RandAugment trained ResNeXt-50 32x4d weights with 79.8 top-1. Trained by <a href="https://github.com/andravin">Andrew Lavin</a> (see Training section for hparams)</li>
</ul> </ul>
<h3 id="april-5-2020_1">April 5, 2020</h3>
<ul>
<li>Add some newly trained MobileNet-V2 models trained with latest h-params, rand augment. They compare quite favourably to EfficientNet-Lite<ul>
<li>3.5M param MobileNet-V2 100 @ 73%</li>
<li>4.5M param MobileNet-V2 110d @ 75%</li>
<li>6.1M param MobileNet-V2 140 @ 76.5%</li>
<li>5.8M param MobileNet-V2 120d @ 77.3%</li>
</ul>
</li>
</ul>
<h3 id="march-18-2020_1">March 18, 2020</h3>
<ul>
<li>Add EfficientNet-Lite models w/ weights ported from <a href="https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet/lite">Tensorflow TPU</a></li>
<li>Add RandAugment trained ResNeXt-50 32x4d weights with 79.8 top-1. Trained by <a href="https://github.com/andravin">Andrew Lavin</a> (see Training section for hparams)</li>
</ul>
<h3 id="feb-29-2020">Feb 29, 2020</h3> <h3 id="feb-29-2020">Feb 29, 2020</h3>
<ul> <ul>
<li>New MobileNet-V3 Large weights trained from stratch with this code to 75.77% top-1</li> <li>New MobileNet-V3 Large weights trained from stratch with this code to 75.77% top-1</li>

@ -285,6 +285,55 @@
</label> </label>
<ul class="md-nav__list" data-md-scrollfix> <ul class="md-nav__list" data-md-scrollfix>
<li class="md-nav__item">
<a href="#feb-10-2021" class="md-nav__link">
Feb 10, 2021
</a>
</li>
<li class="md-nav__item">
<a href="#feb-8-2021" class="md-nav__link">
Feb 8, 2021
</a>
</li>
<li class="md-nav__item">
<a href="#jan-30-2021" class="md-nav__link">
Jan 30, 2021
</a>
</li>
<li class="md-nav__item">
<a href="#jan-25-2021" class="md-nav__link">
Jan 25, 2021
</a>
</li>
<li class="md-nav__item">
<a href="#jan-3-2021" class="md-nav__link">
Jan 3, 2021
</a>
</li>
<li class="md-nav__item">
<a href="#dec-18-2020" class="md-nav__link">
Dec 18, 2020
</a>
</li>
<li class="md-nav__item">
<a href="#dec-7-2020" class="md-nav__link">
Dec 7, 2020
</a>
</li>
<li class="md-nav__item"> <li class="md-nav__item">
<a href="#oct-30-2020" class="md-nav__link"> <a href="#oct-30-2020" class="md-nav__link">
Oct 30, 2020 Oct 30, 2020
@ -413,6 +462,55 @@
</label> </label>
<ul class="md-nav__list" data-md-scrollfix> <ul class="md-nav__list" data-md-scrollfix>
<li class="md-nav__item">
<a href="#feb-10-2021" class="md-nav__link">
Feb 10, 2021
</a>
</li>
<li class="md-nav__item">
<a href="#feb-8-2021" class="md-nav__link">
Feb 8, 2021
</a>
</li>
<li class="md-nav__item">
<a href="#jan-30-2021" class="md-nav__link">
Jan 30, 2021
</a>
</li>
<li class="md-nav__item">
<a href="#jan-25-2021" class="md-nav__link">
Jan 25, 2021
</a>
</li>
<li class="md-nav__item">
<a href="#jan-3-2021" class="md-nav__link">
Jan 3, 2021
</a>
</li>
<li class="md-nav__item">
<a href="#dec-18-2020" class="md-nav__link">
Dec 18, 2020
</a>
</li>
<li class="md-nav__item">
<a href="#dec-7-2020" class="md-nav__link">
Dec 7, 2020
</a>
</li>
<li class="md-nav__item"> <li class="md-nav__item">
<a href="#oct-30-2020" class="md-nav__link"> <a href="#oct-30-2020" class="md-nav__link">
Oct 30, 2020 Oct 30, 2020
@ -518,6 +616,77 @@
<h1 id="recent-changes">Recent Changes</h1> <h1 id="recent-changes">Recent Changes</h1>
<h3 id="feb-10-2021">Feb 10, 2021</h3>
<ul>
<li>More model archs, incl a flexible ByobNet backbone ('Bring-your-own-blocks')<ul>
<li>GPU-Efficient-Networks (<a href="https://github.com/idstcv/GPU-Efficient-Networks">https://github.com/idstcv/GPU-Efficient-Networks</a>), impl in <code>byobnet.py</code></li>
<li>RepVGG (<a href="https://github.com/DingXiaoH/RepVGG">https://github.com/DingXiaoH/RepVGG</a>), impl in <code>byobnet.py</code></li>
<li>classic VGG (from torchvision, impl in <code>vgg</code>)</li>
</ul>
</li>
<li>Refinements to normalizer layer arg handling and normalizer+act layer handling in some models</li>
<li>Default AMP mode changed to native PyTorch AMP instead of APEX. Issues not being fixed with APEX. Native works with <code>--channels-last</code> and <code>--torchscript</code> model training, APEX does not.</li>
<li>Fix a few bugs introduced since last pypi release</li>
</ul>
<h3 id="feb-8-2021">Feb 8, 2021</h3>
<ul>
<li>Add several ResNet weights with ECA attention. 26t &amp; 50t trained @ 256, test @ 320. 269d train @ 256, fine-tune @320, test @ 352.<ul>
<li><code>ecaresnet26t</code> - 79.88 top-1 @ 320x320, 79.08 @ 256x256</li>
<li><code>ecaresnet50t</code> - 82.35 top-1 @ 320x320, 81.52 @ 256x256</li>
<li><code>ecaresnet269d</code> - 84.93 top-1 @ 352x352, 84.87 @ 320x320</li>
</ul>
</li>
<li>Remove separate tiered (<code>t</code>) vs tiered_narrow (<code>tn</code>) ResNet model defs, all <code>tn</code> changed to <code>t</code> and <code>t</code> models removed (<code>seresnext26t_32x4d</code> only model w/ weights that was removed).</li>
<li>Support model default_cfgs with separate train vs test resolution <code>test_input_size</code> and remove extra <code>_320</code> suffix ResNet model defs that were just for test.</li>
</ul>
<h3 id="jan-30-2021">Jan 30, 2021</h3>
<ul>
<li>Add initial "Normalization Free" NF-RegNet-B* and NF-ResNet model definitions based on <a href="https://arxiv.org/abs/2101.08692">paper</a></li>
</ul>
<h3 id="jan-25-2021">Jan 25, 2021</h3>
<ul>
<li>Add ResNetV2 Big Transfer (BiT) models w/ ImageNet-1k and 21k weights from <a href="https://github.com/google-research/big_transfer">https://github.com/google-research/big_transfer</a></li>
<li>Add official R50+ViT-B/16 hybrid models + weights from <a href="https://github.com/google-research/vision_transformer">https://github.com/google-research/vision_transformer</a></li>
<li>ImageNet-21k ViT weights are added w/ model defs and representation layer (pre logits) support<ul>
<li>NOTE: ImageNet-21k classifier heads were zero'd in original weights, they are only useful for transfer learning</li>
</ul>
</li>
<li>Add model defs and weights for DeiT Vision Transformer models from <a href="https://github.com/facebookresearch/deit">https://github.com/facebookresearch/deit</a></li>
<li>Refactor dataset classes into ImageDataset/IterableImageDataset + dataset specific parser classes</li>
<li>Add Tensorflow-Datasets (TFDS) wrapper to allow use of TFDS image classification sets with train script<ul>
<li>Ex: <code>train.py /data/tfds --dataset tfds/oxford_iiit_pet --val-split test --model resnet50 -b 256 --amp --num-classes 37 --opt adamw --lr 3e-4 --weight-decay .001 --pretrained -j 2</code></li>
</ul>
</li>
<li>Add improved .tar dataset parser that reads images from .tar, folder of .tar files, or .tar within .tar<ul>
<li>Run validation on full ImageNet-21k directly from tar w/ BiT model: <code>validate.py /data/fall11_whole.tar --model resnetv2_50x1_bitm_in21k --amp</code></li>
</ul>
</li>
<li>Models in this update should be stable w/ possible exception of ViT/BiT, possibility of some regressions with train/val scripts and dataset handling</li>
</ul>
<h3 id="jan-3-2021">Jan 3, 2021</h3>
<ul>
<li>Add SE-ResNet-152D weights<ul>
<li>256x256 val, 0.94 crop top-1 - 83.75</li>
<li>320x320 val, 1.0 crop - 84.36</li>
</ul>
</li>
<li>Update results files</li>
</ul>
<h3 id="dec-18-2020">Dec 18, 2020</h3>
<ul>
<li>Add ResNet-101D, ResNet-152D, and ResNet-200D weights trained @ 256x256<ul>
<li>256x256 val, 0.94 crop (top-1) - 101D (82.33), 152D (83.08), 200D (83.25)</li>
<li>288x288 val, 1.0 crop - 101D (82.64), 152D (83.48), 200D (83.76)</li>
<li>320x320 val, 1.0 crop - 101D (83.00), 152D (83.66), 200D (84.01)</li>
</ul>
</li>
</ul>
<h3 id="dec-7-2020">Dec 7, 2020</h3>
<ul>
<li>Simplify EMA module (ModelEmaV2), compatible with fully torchscripted models</li>
<li>Misc fixes for SiLU ONNX export, default_cfg missing from Feature extraction models, Linear layer w/ AMP + torchscript</li>
<li>PyPi release @ 0.3.2 (needed by EfficientDet)</li>
</ul>
<h3 id="oct-30-2020">Oct 30, 2020</h3> <h3 id="oct-30-2020">Oct 30, 2020</h3>
<ul> <ul>
<li>Test with PyTorch 1.7 and fix a small top-n metric view vs reshape issue.</li> <li>Test with PyTorch 1.7 and fix a small top-n metric view vs reshape issue.</li>

@ -225,6 +225,13 @@
</label> </label>
<ul class="md-nav__list" data-md-scrollfix> <ul class="md-nav__list" data-md-scrollfix>
<li class="md-nav__item">
<a href="#big-transfer-resnetv2-bit-resnetv2py" class="md-nav__link">
Big Transfer ResNetV2 (BiT) [resnetv2.py]
</a>
</li>
<li class="md-nav__item"> <li class="md-nav__item">
<a href="#cross-stage-partial-networks-cspnetpy" class="md-nav__link"> <a href="#cross-stage-partial-networks-cspnetpy" class="md-nav__link">
Cross-Stage Partial Networks [cspnet.py] Cross-Stage Partial Networks [cspnet.py]
@ -251,6 +258,13 @@
Dual-Path Networks [dpn.py] Dual-Path Networks [dpn.py]
</a> </a>
</li>
<li class="md-nav__item">
<a href="#gpu-efficient-networks-byobnetpy" class="md-nav__link">
GPU-Efficient Networks [byobnet.py]
</a>
</li> </li>
<li class="md-nav__item"> <li class="md-nav__item">
@ -314,6 +328,13 @@
RegNet [regnet.py] RegNet [regnet.py]
</a> </a>
</li>
<li class="md-nav__item">
<a href="#repvgg-byobnetpy" class="md-nav__link">
RepVGG [byobnet.py]
</a>
</li> </li>
<li class="md-nav__item"> <li class="md-nav__item">
@ -370,6 +391,13 @@
TResNet [tresnet.py] TResNet [tresnet.py]
</a> </a>
</li>
<li class="md-nav__item">
<a href="#vgg-vggpy" class="md-nav__link">
VGG [vgg.py]
</a>
</li> </li>
<li class="md-nav__item"> <li class="md-nav__item">
@ -511,6 +539,13 @@
</label> </label>
<ul class="md-nav__list" data-md-scrollfix> <ul class="md-nav__list" data-md-scrollfix>
<li class="md-nav__item">
<a href="#big-transfer-resnetv2-bit-resnetv2py" class="md-nav__link">
Big Transfer ResNetV2 (BiT) [resnetv2.py]
</a>
</li>
<li class="md-nav__item"> <li class="md-nav__item">
<a href="#cross-stage-partial-networks-cspnetpy" class="md-nav__link"> <a href="#cross-stage-partial-networks-cspnetpy" class="md-nav__link">
Cross-Stage Partial Networks [cspnet.py] Cross-Stage Partial Networks [cspnet.py]
@ -537,6 +572,13 @@
Dual-Path Networks [dpn.py] Dual-Path Networks [dpn.py]
</a> </a>
</li>
<li class="md-nav__item">
<a href="#gpu-efficient-networks-byobnetpy" class="md-nav__link">
GPU-Efficient Networks [byobnet.py]
</a>
</li> </li>
<li class="md-nav__item"> <li class="md-nav__item">
@ -600,6 +642,13 @@
RegNet [regnet.py] RegNet [regnet.py]
</a> </a>
</li>
<li class="md-nav__item">
<a href="#repvgg-byobnetpy" class="md-nav__link">
RepVGG [byobnet.py]
</a>
</li> </li>
<li class="md-nav__item"> <li class="md-nav__item">
@ -656,6 +705,13 @@
TResNet [tresnet.py] TResNet [tresnet.py]
</a> </a>
</li>
<li class="md-nav__item">
<a href="#vgg-vggpy" class="md-nav__link">
VGG [vgg.py]
</a>
</li> </li>
<li class="md-nav__item"> <li class="md-nav__item">
@ -722,6 +778,11 @@
<li>trained from scratch using the included training script</li> <li>trained from scratch using the included training script</li>
</ol> </ol>
<p>The validation results for the pretrained weights can be found <a href="../results/">here</a></p> <p>The validation results for the pretrained weights can be found <a href="../results/">here</a></p>
<h2 id="big-transfer-resnetv2-bit-resnetv2py">Big Transfer ResNetV2 (BiT) [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/resnetv2.py">resnetv2.py</a>]</h2>
<ul>
<li>Paper: <code>Big Transfer (BiT): General Visual Representation Learning</code> - <a href="https://arxiv.org/abs/1912.11370">https://arxiv.org/abs/1912.11370</a></li>
<li>Reference code: <a href="https://github.com/google-research/big_transfer">https://github.com/google-research/big_transfer</a></li>
</ul>
<h2 id="cross-stage-partial-networks-cspnetpy">Cross-Stage Partial Networks [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/cspnet.py">cspnet.py</a>]</h2> <h2 id="cross-stage-partial-networks-cspnetpy">Cross-Stage Partial Networks [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/cspnet.py">cspnet.py</a>]</h2>
<ul> <ul>
<li>Paper: <code>CSPNet: A New Backbone that can Enhance Learning Capability of CNN</code> - <a href="https://arxiv.org/abs/1911.11929">https://arxiv.org/abs/1911.11929</a></li> <li>Paper: <code>CSPNet: A New Backbone that can Enhance Learning Capability of CNN</code> - <a href="https://arxiv.org/abs/1911.11929">https://arxiv.org/abs/1911.11929</a></li>
@ -743,6 +804,11 @@
<li>My PyTorch code: <a href="https://github.com/rwightman/pytorch-dpn-pretrained">https://github.com/rwightman/pytorch-dpn-pretrained</a></li> <li>My PyTorch code: <a href="https://github.com/rwightman/pytorch-dpn-pretrained">https://github.com/rwightman/pytorch-dpn-pretrained</a></li>
<li>Reference code: <a href="https://github.com/cypw/DPNs">https://github.com/cypw/DPNs</a></li> <li>Reference code: <a href="https://github.com/cypw/DPNs">https://github.com/cypw/DPNs</a></li>
</ul> </ul>
<h2 id="gpu-efficient-networks-byobnetpy">GPU-Efficient Networks [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/byobnet.py">byobnet.py</a>]</h2>
<ul>
<li>Paper: <code>Neural Architecture Design for GPU-Efficient Networks</code> - <a href="https://arxiv.org/abs/2006.14090">https://arxiv.org/abs/2006.14090</a></li>
<li>Reference code: <a href="https://github.com/idstcv/GPU-Efficient-Networks">https://github.com/idstcv/GPU-Efficient-Networks</a></li>
</ul>
<h2 id="hrnet-hrnetpy">HRNet [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/hrnet.py">hrnet.py</a>]</h2> <h2 id="hrnet-hrnetpy">HRNet [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/hrnet.py">hrnet.py</a>]</h2>
<ul> <ul>
<li>Paper: <code>Deep High-Resolution Representation Learning for Visual Recognition</code> - <a href="https://arxiv.org/abs/1908.07919">https://arxiv.org/abs/1908.07919</a></li> <li>Paper: <code>Deep High-Resolution Representation Learning for Visual Recognition</code> - <a href="https://arxiv.org/abs/1908.07919">https://arxiv.org/abs/1908.07919</a></li>
@ -804,6 +870,11 @@
<li>Paper: <code>Designing Network Design Spaces</code> - <a href="https://arxiv.org/abs/2003.13678">https://arxiv.org/abs/2003.13678</a></li> <li>Paper: <code>Designing Network Design Spaces</code> - <a href="https://arxiv.org/abs/2003.13678">https://arxiv.org/abs/2003.13678</a></li>
<li>Reference code: <a href="https://github.com/facebookresearch/pycls/blob/master/pycls/models/regnet.py">https://github.com/facebookresearch/pycls/blob/master/pycls/models/regnet.py</a></li> <li>Reference code: <a href="https://github.com/facebookresearch/pycls/blob/master/pycls/models/regnet.py">https://github.com/facebookresearch/pycls/blob/master/pycls/models/regnet.py</a></li>
</ul> </ul>
<h2 id="repvgg-byobnetpy">RepVGG [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/byobnet.py">byobnet.py</a>]</h2>
<ul>
<li>Paper: <code>Making VGG-style ConvNets Great Again</code> - <a href="https://arxiv.org/abs/2101.03697">https://arxiv.org/abs/2101.03697</a></li>
<li>Reference code: <a href="https://github.com/DingXiaoH/RepVGG">https://github.com/DingXiaoH/RepVGG</a></li>
</ul>
<h2 id="resnet-resnext-resnetpy">ResNet, ResNeXt [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/resnet.py">resnet.py</a>]</h2> <h2 id="resnet-resnext-resnetpy">ResNet, ResNeXt [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/resnet.py">resnet.py</a>]</h2>
<ul> <ul>
<li>ResNet (V1B)<ul> <li>ResNet (V1B)<ul>
@ -878,6 +949,11 @@
<li>Paper: <code>TResNet: High Performance GPU-Dedicated Architecture</code> - <a href="https://arxiv.org/abs/2003.13630">https://arxiv.org/abs/2003.13630</a></li> <li>Paper: <code>TResNet: High Performance GPU-Dedicated Architecture</code> - <a href="https://arxiv.org/abs/2003.13630">https://arxiv.org/abs/2003.13630</a></li>
<li>Code: <a href="https://github.com/mrT23/TResNet">https://github.com/mrT23/TResNet</a></li> <li>Code: <a href="https://github.com/mrT23/TResNet">https://github.com/mrT23/TResNet</a></li>
</ul> </ul>
<h2 id="vgg-vggpy">VGG [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/vgg.py">vgg.py</a>]</h2>
<ul>
<li>Paper: <code>Very Deep Convolutional Networks For Large-Scale Image Recognition</code> - <a href="https://arxiv.org/pdf/1409.1556.pdf">https://arxiv.org/pdf/1409.1556.pdf</a></li>
<li>Reference code: <a href="https://github.com/pytorch/vision/blob/master/torchvision/models/vgg.py">https://github.com/pytorch/vision/blob/master/torchvision/models/vgg.py</a></li>
</ul>
<h2 id="vision-transformer-vision_transformerpy">Vision Transformer [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/vision_transformer.py">vision_transformer.py</a>]</h2> <h2 id="vision-transformer-vision_transformerpy">Vision Transformer [<a href="https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/vision_transformer.py">vision_transformer.py</a>]</h2>
<ul> <ul>
<li>Paper: <code>An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale</code> - <a href="https://arxiv.org/abs/2010.11929">https://arxiv.org/abs/2010.11929</a></li> <li>Paper: <code>An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale</code> - <a href="https://arxiv.org/abs/2010.11929">https://arxiv.org/abs/2010.11929</a></li>

@ -384,14 +384,14 @@
<h2 id="training-script">Training Script</h2> <h2 id="training-script">Training Script</h2>
<p>The variety of training args is large and not all combinations of options (or even options) have been fully tested. For the training dataset folder, specify the folder to the base that contains a <code>train</code> and <code>validation</code> folder.</p> <p>The variety of training args is large and not all combinations of options (or even options) have been fully tested. For the training dataset folder, specify the folder to the base that contains a <code>train</code> and <code>validation</code> folder.</p>
<p>To train an SE-ResNet34 on ImageNet, locally distributed, 4 GPUs, one process per GPU w/ cosine schedule, random-erasing prob of 50% and per-pixel random value:</p> <p>To train an SE-ResNet34 on ImageNet, locally distributed, 4 GPUs, one process per GPU w/ cosine schedule, random-erasing prob of 50% and per-pixel random value:</p>
<p><code>./distributed_train.sh 4 /data/imagenet --model seresnet34 --sched cosine --epochs 150 --warmup-epochs 5 --lr 0.4 --reprob 0.5 --remode pixel --batch-size 256 -j 4</code></p> <p><code>./distributed_train.sh 4 /data/imagenet --model seresnet34 --sched cosine --epochs 150 --warmup-epochs 5 --lr 0.4 --reprob 0.5 --remode pixel --batch-size 256 --amp -j 4</code></p>
<p>NOTE: NVIDIA APEX should be installed to run in per-process distributed via DDP or to enable AMP mixed precision with the --amp flag</p> <p>NOTE: It is recommended to use PyTorch 1.7+ w/ PyTorch native AMP and DDP instead of APEX AMP. <code>--amp</code> defaults to native AMP as of timm ver 0.4.3. <code>--apex-amp</code> will force use of APEX components if they are installed.</p>
<h2 id="validation-inference-scripts">Validation / Inference Scripts</h2> <h2 id="validation-inference-scripts">Validation / Inference Scripts</h2>
<p>Validation and inference scripts are similar in usage. One outputs metrics on a validation set and the other outputs topk class ids in a csv. Specify the folder containing validation images, not the base as in training script. </p> <p>Validation and inference scripts are similar in usage. One outputs metrics on a validation set and the other outputs topk class ids in a csv. Specify the folder containing validation images, not the base as in training script. </p>
<p>To validate with the model's pretrained weights (if they exist):</p> <p>To validate with the model's pretrained weights (if they exist):</p>
<p><code>python validate.py /imagenet/validation/ --model seresnext26_32x4d --pretrained</code></p> <p><code>python validate.py /imagenet/validation/ --model seresnext26_32x4d --pretrained</code></p>
<p>To run inference from a checkpoint:</p> <p>To run inference from a checkpoint:</p>
<p><code>python inference.py /imagenet/validation/ --model mobilenetv3_large_100 --checkpoint ./output/model_best.pth.tar</code></p> <p><code>python inference.py /imagenet/validation/ --model mobilenetv3_large_100 --checkpoint ./output/train/model_best.pth.tar</code></p>

File diff suppressed because one or more lines are too long

@ -1,35 +1,35 @@
<?xml version="1.0" encoding="UTF-8"?> <?xml version="1.0" encoding="UTF-8"?>
<urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9"><url> <urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9"><url>
<loc>None</loc> <loc>None</loc>
<lastmod>2020-10-30</lastmod> <lastmod>2021-02-11</lastmod>
<changefreq>daily</changefreq> <changefreq>daily</changefreq>
</url><url> </url><url>
<loc>None</loc> <loc>None</loc>
<lastmod>2020-10-30</lastmod> <lastmod>2021-02-11</lastmod>
<changefreq>daily</changefreq> <changefreq>daily</changefreq>
</url><url> </url><url>
<loc>None</loc> <loc>None</loc>
<lastmod>2020-10-30</lastmod> <lastmod>2021-02-11</lastmod>
<changefreq>daily</changefreq> <changefreq>daily</changefreq>
</url><url> </url><url>
<loc>None</loc> <loc>None</loc>
<lastmod>2020-10-30</lastmod> <lastmod>2021-02-11</lastmod>
<changefreq>daily</changefreq> <changefreq>daily</changefreq>
</url><url> </url><url>
<loc>None</loc> <loc>None</loc>
<lastmod>2020-10-30</lastmod> <lastmod>2021-02-11</lastmod>
<changefreq>daily</changefreq> <changefreq>daily</changefreq>
</url><url> </url><url>
<loc>None</loc> <loc>None</loc>
<lastmod>2020-10-30</lastmod> <lastmod>2021-02-11</lastmod>
<changefreq>daily</changefreq> <changefreq>daily</changefreq>
</url><url> </url><url>
<loc>None</loc> <loc>None</loc>
<lastmod>2020-10-30</lastmod> <lastmod>2021-02-11</lastmod>
<changefreq>daily</changefreq> <changefreq>daily</changefreq>
</url><url> </url><url>
<loc>None</loc> <loc>None</loc>
<lastmod>2020-10-30</lastmod> <lastmod>2021-02-11</lastmod>
<changefreq>daily</changefreq> <changefreq>daily</changefreq>
</url> </url>
</urlset> </urlset>

Binary file not shown.
Loading…
Cancel
Save