Skip to content

Commit 318360c

Browse files
committed
Update README.md before merge. Bump version to 0.4.10
1 parent 99d97e0 commit 318360c

File tree

4 files changed

+66
-25
lines changed

4 files changed

+66
-25
lines changed

README.md

Lines changed: 13 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,14 @@ I'm fortunate to be able to dedicate significant time and money of my own suppor
2323

2424
## What's New
2525

26+
### May 25, 2021
27+
* Add LeViT, Visformer, ConViT (PR by Aman Arora), Twins (PR by paper authors) transformer models
28+
* Add ResMLP and gMLP MLP vision models to the existing MLP Mixer impl
29+
* Fix a number of torchscript issues with various vision transformer models
30+
* Cleanup input_size/img_size override handling and improve testing / test coverage for all vision transformer and MLP models
31+
* More flexible pos embedding resize (non-square) for ViT and TnT. Thanks [Alexander Soare](https://github.com/alexander-soare)
32+
* Add `efficientnetv2_rw_m` model and weights (started training before official code). 84.8 top-1, 53M params.
33+
2634
### May 14, 2021
2735
* Add EfficientNet-V2 official model defs w/ ported weights from official [Tensorflow/Keras](https://github.com/google/automl/tree/master/efficientnetv2) impl.
2836
* 1k trained variants: `tf_efficientnetv2_s/m/l`
@@ -166,30 +174,6 @@ I'm fortunate to be able to dedicate significant time and money of my own suppor
166174
* Misc fixes for SiLU ONNX export, default_cfg missing from Feature extraction models, Linear layer w/ AMP + torchscript
167175
* PyPi release @ 0.3.2 (needed by EfficientDet)
168176

169-
### Oct 30, 2020
170-
* Test with PyTorch 1.7 and fix a small top-n metric view vs reshape issue.
171-
* Convert newly added 224x224 Vision Transformer weights from official JAX repo. 81.8 top-1 for B/16, 83.1 L/16.
172-
* Support PyTorch 1.7 optimized, native SiLU (aka Swish) activation. Add mapping to 'silu' name, custom swish will eventually be deprecated.
173-
* Fix regression for loading pretrained classifier via direct model entrypoint functions. Didn't impact create_model() factory usage.
174-
* PyPi release @ 0.3.0 version!
175-
176-
### Oct 26, 2020
177-
* Update Vision Transformer models to be compatible with official code release at https://github.com/google-research/vision_transformer
178-
* Add Vision Transformer weights (ImageNet-21k pretrain) for 384x384 base and large models converted from official jax impl
179-
* ViT-B/16 - 84.2
180-
* ViT-B/32 - 81.7
181-
* ViT-L/16 - 85.2
182-
* ViT-L/32 - 81.5
183-
184-
### Oct 21, 2020
185-
* Weights added for Vision Transformer (ViT) models. 77.86 top-1 for 'small' and 79.35 for 'base'. Thanks to [Christof](https://www.kaggle.com/christofhenkel) for training the base model w/ lots of GPUs.
186-
187-
### Oct 13, 2020
188-
* Initial impl of Vision Transformer models. Both patch and hybrid (CNN backbone) variants. Currently trying to train...
189-
* Adafactor and AdaHessian (FP32 only, no AMP) optimizers
190-
* EdgeTPU-M (`efficientnet_em`) model trained in PyTorch, 79.3 top-1
191-
* Pip release, doc updates pending a few more changes...
192-
193177

194178
## Introduction
195179

@@ -207,6 +191,7 @@ A full version of the list below with source links can be found in the [document
207191
* Bottleneck Transformers - https://arxiv.org/abs/2101.11605
208192
* CaiT (Class-Attention in Image Transformers) - https://arxiv.org/abs/2103.17239
209193
* CoaT (Co-Scale Conv-Attentional Image Transformers) - https://arxiv.org/abs/2104.06399
194+
* ConViT (Soft Convolutional Inductive Biases Vision Transformers)- https://arxiv.org/abs/2103.10697
210195
* CspNet (Cross-Stage Partial Networks) - https://arxiv.org/abs/1911.11929
211196
* DeiT (Vision Transformer) - https://arxiv.org/abs/2012.12877
212197
* DenseNet - https://arxiv.org/abs/1608.06993
@@ -224,13 +209,15 @@ A full version of the list below with source links can be found in the [document
224209
* MobileNet-V2 - https://arxiv.org/abs/1801.04381
225210
* Single-Path NAS - https://arxiv.org/abs/1904.02877
226211
* GhostNet - https://arxiv.org/abs/1911.11907
212+
* gMLP - https://arxiv.org/abs/2105.08050
227213
* GPU-Efficient Networks - https://arxiv.org/abs/2006.14090
228214
* Halo Nets - https://arxiv.org/abs/2103.12731
229215
* HardCoRe-NAS - https://arxiv.org/abs/2102.11646
230216
* HRNet - https://arxiv.org/abs/1908.07919
231217
* Inception-V3 - https://arxiv.org/abs/1512.00567
232218
* Inception-ResNet-V2 and Inception-V4 - https://arxiv.org/abs/1602.07261
233219
* Lambda Networks - https://arxiv.org/abs/2102.08602
220+
* LeViT (Vision Transformer in ConvNet's Clothing) - https://arxiv.org/abs/2104.01136
234221
* MLP-Mixer - https://arxiv.org/abs/2105.01601
235222
* MobileNet-V3 (MBConvNet w/ Efficient Head) - https://arxiv.org/abs/1905.02244
236223
* NASNet-A - https://arxiv.org/abs/1707.07012
@@ -240,6 +227,7 @@ A full version of the list below with source links can be found in the [document
240227
* Pooling-based Vision Transformer (PiT) - https://arxiv.org/abs/2103.16302
241228
* RegNet - https://arxiv.org/abs/2003.13678
242229
* RepVGG - https://arxiv.org/abs/2101.03697
230+
* ResMLP - https://arxiv.org/abs/2105.03404
243231
* ResNet/ResNeXt
244232
* ResNet (v1b/v1.5) - https://arxiv.org/abs/1512.03385
245233
* ResNeXt - https://arxiv.org/abs/1611.05431
@@ -257,6 +245,7 @@ A full version of the list below with source links can be found in the [document
257245
* Swin Transformer - https://arxiv.org/abs/2103.14030
258246
* Transformer-iN-Transformer (TNT) - https://arxiv.org/abs/2103.00112
259247
* TResNet - https://arxiv.org/abs/2003.13630
248+
* Twins (Spatial Attention in Vision Transformers) - https://arxiv.org/pdf/2104.13840.pdf
260249
* Vision Transformer - https://arxiv.org/abs/2010.11929
261250
* VovNet V2 and V1 - https://arxiv.org/abs/1911.06667
262251
* Xception - https://arxiv.org/abs/1610.02357

docs/archived_changes.md

Lines changed: 24 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,29 @@
11
# Archived Changes
22

3+
### Oct 30, 2020
4+
* Test with PyTorch 1.7 and fix a small top-n metric view vs reshape issue.
5+
* Convert newly added 224x224 Vision Transformer weights from official JAX repo. 81.8 top-1 for B/16, 83.1 L/16.
6+
* Support PyTorch 1.7 optimized, native SiLU (aka Swish) activation. Add mapping to 'silu' name, custom swish will eventually be deprecated.
7+
* Fix regression for loading pretrained classifier via direct model entrypoint functions. Didn't impact create_model() factory usage.
8+
* PyPi release @ 0.3.0 version!
9+
10+
### Oct 26, 2020
11+
* Update Vision Transformer models to be compatible with official code release at https://github.com/google-research/vision_transformer
12+
* Add Vision Transformer weights (ImageNet-21k pretrain) for 384x384 base and large models converted from official jax impl
13+
* ViT-B/16 - 84.2
14+
* ViT-B/32 - 81.7
15+
* ViT-L/16 - 85.2
16+
* ViT-L/32 - 81.5
17+
18+
### Oct 21, 2020
19+
* Weights added for Vision Transformer (ViT) models. 77.86 top-1 for 'small' and 79.35 for 'base'. Thanks to [Christof](https://www.kaggle.com/christofhenkel) for training the base model w/ lots of GPUs.
20+
21+
### Oct 13, 2020
22+
* Initial impl of Vision Transformer models. Both patch and hybrid (CNN backbone) variants. Currently trying to train...
23+
* Adafactor and AdaHessian (FP32 only, no AMP) optimizers
24+
* EdgeTPU-M (`efficientnet_em`) model trained in PyTorch, 79.3 top-1
25+
* Pip release, doc updates pending a few more changes...
26+
327
### Sept 18, 2020
428
* New ResNet 'D' weights. 72.7 (top-1) ResNet-18-D, 77.1 ResNet-34-D, 80.5 ResNet-50-D
529
* Added a few untrained defs for other ResNet models (66D, 101D, 152D, 200/200D)

docs/changes.md

Lines changed: 28 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,33 @@
11
# Recent Changes
22

3+
### May 25, 2021
4+
* Add LeViT, Visformer, Convit (PR by Aman Arora), Twins (PR by paper authors) transformer models
5+
* Cleanup input_size/img_size override handling and testing for all vision transformer models
6+
* Add `efficientnetv2_rw_m` model and weights (started training before official code). 84.8 top-1, 53M params.
7+
8+
### May 14, 2021
9+
* Add EfficientNet-V2 official model defs w/ ported weights from official [Tensorflow/Keras](https://github.com/google/automl/tree/master/efficientnetv2) impl.
10+
* 1k trained variants: `tf_efficientnetv2_s/m/l`
11+
* 21k trained variants: `tf_efficientnetv2_s/m/l_in21k`
12+
* 21k pretrained -> 1k fine-tuned: `tf_efficientnetv2_s/m/l_in21ft1k`
13+
* v2 models w/ v1 scaling: `tf_efficientnetv2_b0` through `b3`
14+
* Rename my prev V2 guess `efficientnet_v2s` -> `efficientnetv2_rw_s`
15+
* Some blank `efficientnetv2_*` models in-place for future native PyTorch training
16+
17+
### May 5, 2021
18+
* Add MLP-Mixer models and port pretrained weights from [Google JAX impl](https://github.com/google-research/vision_transformer/tree/linen)
19+
* Add CaiT models and pretrained weights from [FB](https://github.com/facebookresearch/deit)
20+
* Add ResNet-RS models and weights from [TF](https://github.com/tensorflow/tpu/tree/master/models/official/resnet/resnet_rs). Thanks [Aman Arora](https://github.com/amaarora)
21+
* Add CoaT models and weights. Thanks [Mohammed Rizin](https://github.com/morizin)
22+
* Add new ImageNet-21k weights & finetuned weights for TResNet, MobileNet-V3, ViT models. Thanks [mrT](https://github.com/mrT23)
23+
* Add GhostNet models and weights. Thanks [Kai Han](https://github.com/iamhankai)
24+
* Update ByoaNet attention modles
25+
* Improve SA module inits
26+
* Hack together experimental stand-alone Swin based attn module and `swinnet`
27+
* Consistent '26t' model defs for experiments.
28+
* Add improved Efficientnet-V2S (prelim model def) weights. 83.8 top-1.
29+
* WandB logging support
30+
331
### April 13, 2021
432
* Add Swin Transformer models and weights from https://github.com/microsoft/Swin-Transformer
533

timm/version.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
__version__ = '0.4.9'
1+
__version__ = '0.4.10'

0 commit comments

Comments
 (0)