TResNet

Last updated on Feb 14, 2021

tresnet_l

Parameters 53 Million
FLOPs 11 Billion
File Size 214.04 MB
Training Data ImageNet
Training Resources 8x NVIDIA 100 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_l
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 224
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
tresnet_l_448

Parameters 53 Million
FLOPs 43 Billion
File Size 214.04 MB
Training Data ImageNet
Training Resources 8x NVIDIA 100 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_l_448
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 448
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
tresnet_m

Parameters 41 Million
FLOPs 6 Billion
File Size 120.03 MB
Training Data ImageNet
Training Resources 8x NVIDIA 100 GPUs
Training Time < 24 hours

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_m
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 224
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
tresnet_m_448

Parameters 29 Million
FLOPs 23 Billion
File Size 120.03 MB
Training Data ImageNet
Training Resources 8x NVIDIA 100 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_m_448
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 448
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
tresnet_xl

Parameters 76 Million
FLOPs 15 Billion
File Size 299.82 MB
Training Data ImageNet
Training Resources 8x NVIDIA 100 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_xl
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 224
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
tresnet_xl_448

Parameters 76 Million
FLOPs 61 Billion
File Size 214.04 MB
Training Data ImageNet
Training Resources 8x NVIDIA 100 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay, AutoAugment, Cutout, Label Smoothing
Architecture 1x1 Convolution, Anti-Alias Downsampling, Convolution, Global Average Pooling, InPlace-ABN, Leaky ReLU, ReLU, Residual Connection, Squeeze-and-Excitation Block
ID tresnet_xl_448
LR 0.01
Epochs 300
Crop Pct 0.875
Momentum 0.9
Image Size 448
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS
README.md

Summary

A TResNet is a variant on a ResNet that aim to boost accuracy while maintaining GPU training and inference efficiency. They contain several design tricks including a SpaceToDepth stem, Anti-Alias downsampling, In-Place Activated BatchNorm, Blocks selection and squeeze-and-excitation layers.

How do I load this model?

To load a pretrained model:

import timm
m = timm.create_model('tresnet_m', pretrained=True)
m.eval()

Replace the model name with the variant you want to use, e.g. tresnet_m. You can find the IDs in the model summaries at the top of this page.

How do I train this model?

You can follow the timm recipe scripts for training a new model afresh.

Citation

@misc{ridnik2020tresnet,
      title={TResNet: High Performance GPU-Dedicated Architecture}, 
      author={Tal Ridnik and Hussam Lawen and Asaf Noy and Emanuel Ben Baruch and Gilad Sharir and Itamar Friedman},
      year={2020},
      eprint={2003.13630},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Results

Image Classification on ImageNet

Image Classification on ImageNet
MODEL TOP 1 ACCURACY TOP 5 ACCURACY
tresnet_xl_448 83.06% 96.19%
tresnet_l_448 82.26% 95.98%
tresnet_xl 82.05% 95.93%
tresnet_m_448 81.72% 95.57%
tresnet_l 81.49% 95.62%
tresnet_m 80.8% 94.86%