Skip to content

Commit 75f3c45

Browse files
committed
EfficientNetV2 Added
1 parent ee53084 commit 75f3c45

File tree

2 files changed

+37
-0
lines changed

2 files changed

+37
-0
lines changed

Images/efficientnetv2.png

41 KB
Loading

README.md

Lines changed: 37 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -671,4 +671,41 @@ Recommended Reading for NAS: https://lilianweng.github.io/lil-log/2020/08/06/neu
671671

672672
</p>
673673
<img src="Images/resmlp.png"; alt="ResMLP">
674+
</details>
675+
676+
<details>
677+
<summary>🔥 EfficientNetV2</summary>
678+
<p>
679+
680+
Paper: EfficientNetV2: Smaller Models and Faster Training by Mingxing Tan, Quoc V. Le
681+
682+
Training efficiency has gained significant interests recently. For instance,
683+
NFNets aim to improve training efficiency by removing the expensive batch normalization;
684+
Several recent works focus on improving training speed by adding attention layers into
685+
convolutional networks (ConvNets); Vision Transformers improves training efficiency on
686+
large-scale datasets by using Transformer blocks. However, these methods often come with
687+
significant overheads.
688+
689+
To develop these models, it uses a combination of training-aware neural search(NAS) and
690+
scaling, to jointly optimize training speed and parameter efficiency.
691+
692+
Drawbracks of previous version of EfficientNets
693+
694+
1. training with very large image sizes is slow.
695+
2. depthwise convolutions are slow in early layers.
696+
3. equally scaling up every stage is sub-optimal.
697+
698+
Whats New With EfficientNetV2
699+
700+
Based on the above observations, V2 is designed on a search space enriched with additional
701+
ops such as Fused-MBConv, and apply training-aware NAS and scaling to jointly optimize model
702+
accuracy, training speed, and parameter size. EfficientNetV2, train up to 4x faster than
703+
prior models, while being up to 6.8x smaller in parameter size.
704+
705+
To further increase the training speed, it uses progressive increase image size, previously
706+
done by FixRes, Mix&Match. The only difference between the current approach from the previous
707+
approach is the use of adaptive regularization as the image size is increased.
708+
709+
</p>
710+
<img src="Images/efficientnetv2.png"; alt="EfficientNetV2">
674711
</details>

0 commit comments

Comments
 (0)