EfficientNet

Last updated on Feb 14, 2021

efficientnet_b0

Parameters 5 Million
FLOPs 511 Million
File Size 20.39 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID efficientnet_b0
Layers 18
Crop Pct 0.875
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
efficientnet_b1

Parameters 8 Million
FLOPs 910 Million
File Size 30.04 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID efficientnet_b1
Crop Pct 0.875
Image Size 240
Interpolation bicubic
SHOW MORE
SHOW LESS
efficientnet_b2

Parameters 9 Million
FLOPs 1 Billion
File Size 35.08 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID efficientnet_b2
Crop Pct 0.875
Image Size 260
Interpolation bicubic
SHOW MORE
SHOW LESS
efficientnet_b2a

Parameters 9 Million
FLOPs 1 Billion
File Size 47.08 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID efficientnet_b2a
Crop Pct 1.0
Image Size 288
Interpolation bicubic
SHOW MORE
SHOW LESS
efficientnet_b3

Parameters 12 Million
FLOPs 2 Billion
File Size 47.08 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID efficientnet_b3
Crop Pct 0.904
Image Size 300
Interpolation bicubic
SHOW MORE
SHOW LESS
efficientnet_b3a

Parameters 12 Million
FLOPs 3 Billion
File Size 47.08 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID efficientnet_b3a
Crop Pct 1.0
Image Size 320
Interpolation bicubic
SHOW MORE
SHOW LESS
efficientnet_em

Parameters 7 Million
FLOPs 4 Billion
File Size 26.63 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID efficientnet_em
Crop Pct 0.882
Image Size 240
Interpolation bicubic
SHOW MORE
SHOW LESS
efficientnet_es

Parameters 5 Million
FLOPs 2 Billion
File Size 20.98 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID efficientnet_es
Crop Pct 0.875
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
efficientnet_lite0

Parameters 5 Million
FLOPs 511 Million
File Size 17.95 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID efficientnet_lite0
Crop Pct 0.875
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
README.md

Summary

EfficientNet is a convolutional neural network architecture and scaling method that uniformly scales all dimensions of depth/width/resolution using a compound coefficient. Unlike conventional practice that arbitrary scales these factors, the EfficientNet scaling method uniformly scales network width, depth, and resolution with a set of fixed scaling coefficients. For example, if we want to use $2^N$ times more computational resources, then we can simply increase the network depth by $\alpha ^ N$, width by $\beta ^ N$, and image size by $\gamma ^ N$, where $\alpha, \beta, \gamma$ are constant coefficients determined by a small grid search on the original small model. EfficientNet uses a compound coefficient $\phi$ to uniformly scales network width, depth, and resolution in a principled way.

The compound scaling method is justified by the intuition that if the input image is bigger, then the network needs more layers to increase the receptive field and more channels to capture more fine-grained patterns on the bigger image.

The base EfficientNet-B0 network is based on the inverted bottleneck residual blocks of MobileNetV2, in addition to squeeze-and-excitation blocks.

How do I load this model?

To load a pretrained model:

import timm
m = timm.create_model('efficientnet_b0', pretrained=True)
m.eval()

Replace the model name with the variant you want to use, e.g. efficientnet_b0. You can find the IDs in the model summaries at the top of this page.

How do I train this model?

You can follow the timm recipe scripts for training a new model afresh.

Citation

@misc{tan2020efficientnet,
      title={EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks}, 
      author={Mingxing Tan and Quoc V. Le},
      year={2020},
      eprint={1905.11946},
      archivePrefix={arXiv},
      primaryClass={cs.LG}
}

Results

Image Classification on ImageNet

Image Classification on ImageNet
MODEL TOP 1 ACCURACY TOP 5 ACCURACY
efficientnet_b3a 82.25% 96.11%
efficientnet_b3 82.08% 96.03%
efficientnet_b2a 80.61% 95.32%
efficientnet_b2 80.38% 95.08%
efficientnet_em 79.26% 94.79%
efficientnet_b1 78.71% 94.15%
efficientnet_es 78.09% 93.93%
efficientnet_b0 77.71% 93.52%
efficientnet_lite0 75.5% 92.51%