RexNet

Last updated on Feb 14, 2021

rexnet_100

Parameters 5 Million
FLOPs 510 Million
File Size 18.52 MB
Training Data ImageNet
Training Resources 4x NVIDIA V100 GPUs
Training Time

Training Techniques Nesterov Accelerated Gradient, Weight Decay, Linear Warmup With Cosine Annealing, Label Smoothing
Architecture ReLU6, Convolution, Batch Normalization, Residual Connection, Dropout
ID rexnet_100
LR 0.5
Epochs 400
Dropout 0.2
Crop Pct 0.875
Momentum 0.9
Batch Size 512
Image Size 224
Weight Decay 0.00001
Interpolation bicubic
Label Smoothing 0.1
SHOW MORE
SHOW LESS
rexnet_130

Parameters 8 Million
FLOPs 848 Million
File Size 29.09 MB
Training Data ImageNet
Training Resources 4x NVIDIA V100 GPUs
Training Time

Training Techniques Nesterov Accelerated Gradient, Weight Decay, Linear Warmup With Cosine Annealing, Label Smoothing
Architecture ReLU6, Convolution, Batch Normalization, Residual Connection, Dropout
ID rexnet_130
LR 0.5
Epochs 400
Dropout 0.2
Crop Pct 0.875
Momentum 0.9
Batch Size 512
Image Size 224
Weight Decay 0.00001
Interpolation bicubic
Label Smoothing 0.1
SHOW MORE
SHOW LESS
rexnet_150

Parameters 10 Million
FLOPs 1 Billion
File Size 37.41 MB
Training Data ImageNet
Training Resources 4x NVIDIA V100 GPUs
Training Time

Training Techniques Nesterov Accelerated Gradient, Weight Decay, Linear Warmup With Cosine Annealing, Label Smoothing
Architecture ReLU6, Convolution, Batch Normalization, Residual Connection, Dropout
ID rexnet_150
LR 0.5
Epochs 400
Dropout 0.2
Crop Pct 0.875
Momentum 0.9
Batch Size 512
Image Size 224
Weight Decay 0.00001
Interpolation bicubic
Label Smoothing 0.1
SHOW MORE
SHOW LESS
rexnet_200

Parameters 16 Million
FLOPs 2 Billion
File Size 62.81 MB
Training Data ImageNet
Training Resources 4x NVIDIA V100 GPUs
Training Time

Training Techniques Nesterov Accelerated Gradient, Weight Decay, Linear Warmup With Cosine Annealing, Label Smoothing
Architecture ReLU6, Convolution, Batch Normalization, Residual Connection, Dropout
ID rexnet_200
LR 0.5
Epochs 400
Dropout 0.2
Crop Pct 0.875
Momentum 0.9
Batch Size 512
Image Size 224
Weight Decay 0.00001
Interpolation bicubic
Label Smoothing 0.1
SHOW MORE
SHOW LESS
README.md

Summary

Rank Expansion Networks (ReXNets) follow a set of new design principles for designing bottlenecks in image classification models. Authors refine each layer by 1) expanding the input channel size of the convolution layer and 2) replacing the ReLU6s.

How do I load this model?

To load a pretrained model:

import timm
m = timm.create_model('rexnet_100', pretrained=True)
m.eval()

Replace the model name with the variant you want to use, e.g. rexnet_100. You can find the IDs in the model summaries at the top of this page.

How do I train this model?

You can follow the timm recipe scripts for training a new model afresh.

Citation

@misc{han2020rexnet,
      title={ReXNet: Diminishing Representational Bottleneck on Convolutional Neural Network}, 
      author={Dongyoon Han and Sangdoo Yun and Byeongho Heo and YoungJoon Yoo},
      year={2020},
      eprint={2007.00992},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Results

Image Classification on ImageNet

Image Classification on ImageNet
MODEL TOP 1 ACCURACY TOP 5 ACCURACY
rexnet_200 81.63% 95.67%
rexnet_150 80.31% 95.16%
rexnet_130 79.49% 94.67%
rexnet_100 77.86% 93.88%