Torch implementation for gradually updated neural networks:
Gradually Updated Neural Networks for Large-Scale Image Recognition
Siyuan Qiao, Zhishuai Zhang, Wei Shen, Bo Wang, Alan Yuille
In Thirty-fifth International Conference on Machine Learning (ICML), 2018.
The code is built on fb.resnet.torch.
@inproceedings{Gunn,
title = {Gradually Updated Neural Networks for Large-Scale Image Recognition},
author = {Siyuan Qiao and Zhishuai Zhang and Wei Shen and Bo Wang and Alan L. Yuille},
booktitle = {International Conference on Machine Learning (ICML)},
year = {2018}
}
The state-of-the-art network architectures usually increase the depths by cascading convolutional layers or building blocks. Gradually Updated Neural Network (GUNN) presents an alternative method to increase the depth. It introduces computation orderings to the channels within convolutional layers or blocks, based on which it gradually computes the outputs in a channel-wise manner. The added orderings not only increase the depths and the learning capacities of the networks without any additional computation costs, but also eliminate the overlap singularities so that the networks are able to converge faster and perform better.
Install Torch and required packages following here. Training on CIFAR
th main.lua -netType gunn-15 -dataset cifar10 -batchSize 64 -nGPU 4 -nThreads 8 -shareGradInput true -nEpochs 300
For CIFAR-100, please change cifar10 to cifar100 after -dataset. Training on ImageNet
th main.lua -netType gunn-18 -dataset imagenet -batchSize 256 -nGPU 4 -nThreads 16 -shareGradInput true -nEpochs 120 -data [data folder]
Model | Parameters | CIFAR-10 | CIFAR-100 |
---|---|---|---|
GUNN-15 | 1.6M | 4.15 | 20.45 |
GUNN-24 | 29.6M | 3.21 | 16.69 |
Model | Parameters | ImageNet Top-1 | ImageNet Top-5 |
---|---|---|---|
GUNN-18 | 28.9M | 21.65 | 5.87 |
Wide GUNN-18 | 45.6M | 20.59 | 5.52 |