You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+13-24Lines changed: 13 additions & 24 deletions
Original file line number
Diff line number
Diff line change
@@ -23,6 +23,14 @@ I'm fortunate to be able to dedicate significant time and money of my own suppor
23
23
24
24
## What's New
25
25
26
+
### May 25, 2021
27
+
* Add LeViT, Visformer, ConViT (PR by Aman Arora), Twins (PR by paper authors) transformer models
28
+
* Add ResMLP and gMLP MLP vision models to the existing MLP Mixer impl
29
+
* Fix a number of torchscript issues with various vision transformer models
30
+
* Cleanup input_size/img_size override handling and improve testing / test coverage for all vision transformer and MLP models
31
+
* More flexible pos embedding resize (non-square) for ViT and TnT. Thanks [Alexander Soare](https://github.com/alexander-soare)
32
+
* Add `efficientnetv2_rw_m` model and weights (started training before official code). 84.8 top-1, 53M params.
33
+
26
34
### May 14, 2021
27
35
* Add EfficientNet-V2 official model defs w/ ported weights from official [Tensorflow/Keras](https://github.com/google/automl/tree/master/efficientnetv2) impl.
28
36
* 1k trained variants: `tf_efficientnetv2_s/m/l`
@@ -166,30 +174,6 @@ I'm fortunate to be able to dedicate significant time and money of my own suppor
166
174
* Misc fixes for SiLU ONNX export, default_cfg missing from Feature extraction models, Linear layer w/ AMP + torchscript
167
175
* PyPi release @ 0.3.2 (needed by EfficientDet)
168
176
169
-
### Oct 30, 2020
170
-
* Test with PyTorch 1.7 and fix a small top-n metric view vs reshape issue.
171
-
* Convert newly added 224x224 Vision Transformer weights from official JAX repo. 81.8 top-1 for B/16, 83.1 L/16.
172
-
* Support PyTorch 1.7 optimized, native SiLU (aka Swish) activation. Add mapping to 'silu' name, custom swish will eventually be deprecated.
173
-
* Fix regression for loading pretrained classifier via direct model entrypoint functions. Didn't impact create_model() factory usage.
174
-
* PyPi release @ 0.3.0 version!
175
-
176
-
### Oct 26, 2020
177
-
* Update Vision Transformer models to be compatible with official code release at https://github.com/google-research/vision_transformer
178
-
* Add Vision Transformer weights (ImageNet-21k pretrain) for 384x384 base and large models converted from official jax impl
179
-
* ViT-B/16 - 84.2
180
-
* ViT-B/32 - 81.7
181
-
* ViT-L/16 - 85.2
182
-
* ViT-L/32 - 81.5
183
-
184
-
### Oct 21, 2020
185
-
* Weights added for Vision Transformer (ViT) models. 77.86 top-1 for 'small' and 79.35 for 'base'. Thanks to [Christof](https://www.kaggle.com/christofhenkel) for training the base model w/ lots of GPUs.
186
-
187
-
### Oct 13, 2020
188
-
* Initial impl of Vision Transformer models. Both patch and hybrid (CNN backbone) variants. Currently trying to train...
189
-
* Adafactor and AdaHessian (FP32 only, no AMP) optimizers
190
-
* EdgeTPU-M (`efficientnet_em`) model trained in PyTorch, 79.3 top-1
191
-
* Pip release, doc updates pending a few more changes...
192
-
193
177
194
178
## Introduction
195
179
@@ -207,6 +191,7 @@ A full version of the list below with source links can be found in the [document
Copy file name to clipboardExpand all lines: docs/archived_changes.md
+24Lines changed: 24 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,29 @@
1
1
# Archived Changes
2
2
3
+
### Oct 30, 2020
4
+
* Test with PyTorch 1.7 and fix a small top-n metric view vs reshape issue.
5
+
* Convert newly added 224x224 Vision Transformer weights from official JAX repo. 81.8 top-1 for B/16, 83.1 L/16.
6
+
* Support PyTorch 1.7 optimized, native SiLU (aka Swish) activation. Add mapping to 'silu' name, custom swish will eventually be deprecated.
7
+
* Fix regression for loading pretrained classifier via direct model entrypoint functions. Didn't impact create_model() factory usage.
8
+
* PyPi release @ 0.3.0 version!
9
+
10
+
### Oct 26, 2020
11
+
* Update Vision Transformer models to be compatible with official code release at https://github.com/google-research/vision_transformer
12
+
* Add Vision Transformer weights (ImageNet-21k pretrain) for 384x384 base and large models converted from official jax impl
13
+
* ViT-B/16 - 84.2
14
+
* ViT-B/32 - 81.7
15
+
* ViT-L/16 - 85.2
16
+
* ViT-L/32 - 81.5
17
+
18
+
### Oct 21, 2020
19
+
* Weights added for Vision Transformer (ViT) models. 77.86 top-1 for 'small' and 79.35 for 'base'. Thanks to [Christof](https://www.kaggle.com/christofhenkel) for training the base model w/ lots of GPUs.
20
+
21
+
### Oct 13, 2020
22
+
* Initial impl of Vision Transformer models. Both patch and hybrid (CNN backbone) variants. Currently trying to train...
23
+
* Adafactor and AdaHessian (FP32 only, no AMP) optimizers
24
+
* EdgeTPU-M (`efficientnet_em`) model trained in PyTorch, 79.3 top-1
25
+
* Pip release, doc updates pending a few more changes...
Copy file name to clipboardExpand all lines: docs/changes.md
+28Lines changed: 28 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,33 @@
1
1
# Recent Changes
2
2
3
+
### May 25, 2021
4
+
* Add LeViT, Visformer, Convit (PR by Aman Arora), Twins (PR by paper authors) transformer models
5
+
* Cleanup input_size/img_size override handling and testing for all vision transformer models
6
+
* Add `efficientnetv2_rw_m` model and weights (started training before official code). 84.8 top-1, 53M params.
7
+
8
+
### May 14, 2021
9
+
* Add EfficientNet-V2 official model defs w/ ported weights from official [Tensorflow/Keras](https://github.com/google/automl/tree/master/efficientnetv2) impl.
0 commit comments