Skip to content

Commit 758c443

Browse files
committed
Update README.md
1 parent d413eef commit 758c443

File tree

3 files changed

+24
-112
lines changed

3 files changed

+24
-112
lines changed

README.md

Lines changed: 6 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,12 @@ I'm fortunate to be able to dedicate significant time and money of my own suppor
2323

2424
## What's New
2525

26+
### June 8, 2021
27+
* Add first ResMLP weights, trained in PyTorch XLA on TPU-VM w/ my XLA branch. 24 block variant, 79.2 top-1.
28+
* Add ResNet51-Q model w/ pretrained weights at 82.36 top-1.
29+
* NFNet inspired block layout with quad layer stem and no maxpool
30+
* Same param count (35.7M) and throughput as ResNetRS-50 but +1.5 top-1 @ 224x224 and +2.5 top-1 at 288x288
31+
2632
### May 25, 2021
2733
* Add LeViT, Visformer, ConViT (PR by Aman Arora), Twins (PR by paper authors) transformer models
2834
* Add ResMLP and gMLP MLP vision models to the existing MLP Mixer impl
@@ -163,17 +169,6 @@ I'm fortunate to be able to dedicate significant time and money of my own suppor
163169
* 320x320 val, 1.0 crop - 84.36
164170
* Update [results files](results/)
165171

166-
### Dec 18, 2020
167-
* Add ResNet-101D, ResNet-152D, and ResNet-200D weights trained @ 256x256
168-
* 256x256 val, 0.94 crop (top-1) - 101D (82.33), 152D (83.08), 200D (83.25)
169-
* 288x288 val, 1.0 crop - 101D (82.64), 152D (83.48), 200D (83.76)
170-
* 320x320 val, 1.0 crop - 101D (83.00), 152D (83.66), 200D (84.01)
171-
172-
### Dec 7, 2020
173-
* Simplify EMA module (ModelEmaV2), compatible with fully torchscripted models
174-
* Misc fixes for SiLU ONNX export, default_cfg missing from Feature extraction models, Linear layer w/ AMP + torchscript
175-
* PyPi release @ 0.3.2 (needed by EfficientDet)
176-
177172

178173
## Introduction
179174

docs/archived_changes.md

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,17 @@
11
# Archived Changes
22

3+
### Dec 18, 2020
4+
* Add ResNet-101D, ResNet-152D, and ResNet-200D weights trained @ 256x256
5+
* 256x256 val, 0.94 crop (top-1) - 101D (82.33), 152D (83.08), 200D (83.25)
6+
* 288x288 val, 1.0 crop - 101D (82.64), 152D (83.48), 200D (83.76)
7+
* 320x320 val, 1.0 crop - 101D (83.00), 152D (83.66), 200D (84.01)
8+
9+
### Dec 7, 2020
10+
* Simplify EMA module (ModelEmaV2), compatible with fully torchscripted models
11+
* Misc fixes for SiLU ONNX export, default_cfg missing from Feature extraction models, Linear layer w/ AMP + torchscript
12+
* PyPi release @ 0.3.2 (needed by EfficientDet)
13+
14+
315
### Oct 30, 2020
416
* Test with PyTorch 1.7 and fix a small top-n metric view vs reshape issue.
517
* Convert newly added 224x224 Vision Transformer weights from official JAX repo. 81.8 top-1 for B/16, 83.1 L/16.

docs/changes.md

Lines changed: 6 additions & 101 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,11 @@
11
# Recent Changes
22

3+
### June 8, 2021
4+
* Add first ResMLP weights, trained in PyTorch XLA on TPU-VM w/ my XLA branch. 24 block variant, 79.2 top-1.
5+
* Add ResNet51-Q model w/ pretrained weights at 82.36 top-1.
6+
* NFNet inspired block layout with quad layer stem and no maxpool
7+
* Same param count (35.7M) and throughput as ResNetRS-50 but +1.5 top-1 @ 224x224 and +2.5 top-1 at 288x288
8+
39
### May 25, 2021
410
* Add LeViT, Visformer, Convit (PR by Aman Arora), Twins (PR by paper authors) transformer models
511
* Cleanup input_size/img_size override handling and testing for all vision transformer models
@@ -122,104 +128,3 @@
122128
* 256x256 val, 0.94 crop top-1 - 83.75
123129
* 320x320 val, 1.0 crop - 84.36
124130
* Update results files
125-
126-
### Dec 18, 2020
127-
* Add ResNet-101D, ResNet-152D, and ResNet-200D weights trained @ 256x256
128-
* 256x256 val, 0.94 crop (top-1) - 101D (82.33), 152D (83.08), 200D (83.25)
129-
* 288x288 val, 1.0 crop - 101D (82.64), 152D (83.48), 200D (83.76)
130-
* 320x320 val, 1.0 crop - 101D (83.00), 152D (83.66), 200D (84.01)
131-
132-
### Dec 7, 2020
133-
* Simplify EMA module (ModelEmaV2), compatible with fully torchscripted models
134-
* Misc fixes for SiLU ONNX export, default_cfg missing from Feature extraction models, Linear layer w/ AMP + torchscript
135-
* PyPi release @ 0.3.2 (needed by EfficientDet)
136-
137-
### Oct 30, 2020
138-
* Test with PyTorch 1.7 and fix a small top-n metric view vs reshape issue.
139-
* Convert newly added 224x224 Vision Transformer weights from official JAX repo. 81.8 top-1 for B/16, 83.1 L/16.
140-
* Support PyTorch 1.7 optimized, native SiLU (aka Swish) activation. Add mapping to 'silu' name, custom swish will eventually be deprecated.
141-
* Fix regression for loading pretrained classifier via direct model entrypoint functions. Didn't impact create_model() factory usage.
142-
* PyPi release @ 0.3.0 version!
143-
144-
### Oct 26, 2020
145-
* Update Vision Transformer models to be compatible with official code release at https://github.com/google-research/vision_transformer
146-
* Add Vision Transformer weights (ImageNet-21k pretrain) for 384x384 base and large models converted from official jax impl
147-
* ViT-B/16 - 84.2
148-
* ViT-B/32 - 81.7
149-
* ViT-L/16 - 85.2
150-
* ViT-L/32 - 81.5
151-
152-
### Oct 21, 2020
153-
* Weights added for Vision Transformer (ViT) models. 77.86 top-1 for 'small' and 79.35 for 'base'. Thanks to [Christof](https://www.kaggle.com/christofhenkel) for training the base model w/ lots of GPUs.
154-
155-
### Oct 13, 2020
156-
* Initial impl of Vision Transformer models. Both patch and hybrid (CNN backbone) variants. Currently trying to train...
157-
* Adafactor and AdaHessian (FP32 only, no AMP) optimizers
158-
* EdgeTPU-M (`efficientnet_em`) model trained in PyTorch, 79.3 top-1
159-
* Pip release, doc updates pending a few more changes...
160-
161-
### Sept 18, 2020
162-
* New ResNet 'D' weights. 72.7 (top-1) ResNet-18-D, 77.1 ResNet-34-D, 80.5 ResNet-50-D
163-
* Added a few untrained defs for other ResNet models (66D, 101D, 152D, 200/200D)
164-
165-
### Sept 3, 2020
166-
* New weights
167-
* Wide-ResNet50 - 81.5 top-1 (vs 78.5 torchvision)
168-
* SEResNeXt50-32x4d - 81.3 top-1 (vs 79.1 cadene)
169-
* Support for native Torch AMP and channels_last memory format added to train/validate scripts (`--channels-last`, `--native-amp` vs `--apex-amp`)
170-
* Models tested with channels_last on latest NGC 20.08 container. AdaptiveAvgPool in attn layers changed to mean((2,3)) to work around bug with NHWC kernel.
171-
172-
### Aug 12, 2020
173-
* New/updated weights from training experiments
174-
* EfficientNet-B3 - 82.1 top-1 (vs 81.6 for official with AA and 81.9 for AdvProp)
175-
* RegNetY-3.2GF - 82.0 top-1 (78.9 from official ver)
176-
* CSPResNet50 - 79.6 top-1 (76.6 from official ver)
177-
* Add CutMix integrated w/ Mixup. See [pull request](https://github.com/rwightman/pytorch-image-models/pull/218) for some usage examples
178-
* Some fixes for using pretrained weights with `in_chans` != 3 on several models.
179-
180-
### Aug 5, 2020
181-
Universal feature extraction, new models, new weights, new test sets.
182-
183-
* All models support the `features_only=True` argument for `create_model` call to return a network that extracts features from the deepest layer at each stride.
184-
* New models
185-
* CSPResNet, CSPResNeXt, CSPDarkNet, DarkNet
186-
* ReXNet
187-
* (Modified Aligned) Xception41/65/71 (a proper port of TF models)
188-
* New trained weights
189-
* SEResNet50 - 80.3 top-1
190-
* CSPDarkNet53 - 80.1 top-1
191-
* CSPResNeXt50 - 80.0 top-1
192-
* DPN68b - 79.2 top-1
193-
* EfficientNet-Lite0 (non-TF ver) - 75.5 (submitted by [@hal-314](https://github.com/hal-314))
194-
* Add 'real' labels for ImageNet and ImageNet-Renditions test set, see [`results/README.md`](results/README.md)
195-
* Test set ranking/top-n diff script by [@KushajveerSingh](https://github.com/KushajveerSingh)
196-
* Train script and loader/transform tweaks to punch through more aug arguments
197-
* README and documentation overhaul. See initial (WIP) documentation at https://rwightman.github.io/pytorch-image-models/
198-
* adamp and sgdp optimizers added by [@hellbell](https://github.com/hellbell)
199-
200-
### June 11, 2020
201-
Bunch of changes:
202-
203-
* DenseNet models updated with memory efficient addition from torchvision (fixed a bug), blur pooling and deep stem additions
204-
* VoVNet V1 and V2 models added, 39 V2 variant (ese_vovnet_39b) trained to 79.3 top-1
205-
* Activation factory added along with new activations:
206-
* select act at model creation time for more flexibility in using activations compatible with scripting or tracing (ONNX export)
207-
* hard_mish (experimental) added with memory-efficient grad, along with ME hard_swish
208-
* context mgr for setting exportable/scriptable/no_jit states
209-
* Norm + Activation combo layers added with initial trial support in DenseNet and VoVNet along with impl of EvoNorm and InplaceAbn wrapper that fit the interface
210-
* Torchscript works for all but two of the model types as long as using Pytorch 1.5+, tests added for this
211-
* Some import cleanup and classifier reset changes, all models will have classifier reset to nn.Identity on reset_classifer(0) call
212-
* Prep for 0.1.28 pip release
213-
214-
### May 12, 2020
215-
* Add ResNeSt models (code adapted from https://github.com/zhanghang1989/ResNeSt, paper https://arxiv.org/abs/2004.08955))
216-
217-
### May 3, 2020
218-
* Pruned EfficientNet B1, B2, and B3 (https://arxiv.org/abs/2002.08258) contributed by [Yonathan Aflalo](https://github.com/yoniaflalo)
219-
220-
### May 1, 2020
221-
* Merged a number of execellent contributions in the ResNet model family over the past month
222-
* BlurPool2D and resnetblur models initiated by [Chris Ha](https://github.com/VRandme), I trained resnetblur50 to 79.3.
223-
* TResNet models and SpaceToDepth, AntiAliasDownsampleLayer layers by [mrT23](https://github.com/mrT23)
224-
* ecaresnet (50d, 101d, light) models and two pruned variants using pruning as per (https://arxiv.org/abs/2002.08258) by [Yonathan Aflalo](https://github.com/yoniaflalo)
225-
* 200 pretrained models in total now with updated results csv in results folder

0 commit comments

Comments
 (0)