TF MobileNet V3

Last updated on Feb 14, 2021

tf_mobilenetv3_large_075

Parameters 4 Million
FLOPs 194 Million
File Size 15.35 MB
Training Data ImageNet
Training Resources 4x4 TPU Pod
Training Time

Training Techniques RMSProp, Weight Decay
Architecture 1x1 Convolution, Batch Normalization, Convolution, Dense Connections, Depthwise Separable Convolution, Dropout, Global Average Pooling, Hard Swish, Inverted Residual Block, Residual Connection, ReLU, Softmax, Squeeze-and-Excitation Block
ID tf_mobilenetv3_large_075
LR 0.1
Dropout 0.8
Crop Pct 0.875
Momentum 0.9
Batch Size 4096
Image Size 224
Weight Decay 0.00001
Interpolation bilinear
SHOW MORE
SHOW LESS
tf_mobilenetv3_large_100

Parameters 5 Million
FLOPs 275 Million
File Size 21.05 MB
Training Data ImageNet
Training Resources 4x4 TPU Pod
Training Time

Training Techniques RMSProp, Weight Decay
Architecture 1x1 Convolution, Batch Normalization, Convolution, Dense Connections, Depthwise Separable Convolution, Dropout, Global Average Pooling, Hard Swish, Inverted Residual Block, Residual Connection, ReLU, Softmax, Squeeze-and-Excitation Block
ID tf_mobilenetv3_large_100
LR 0.1
Dropout 0.8
Crop Pct 0.875
Momentum 0.9
Batch Size 4096
Image Size 224
Weight Decay 0.00001
Interpolation bilinear
SHOW MORE
SHOW LESS
tf_mobilenetv3_large_minimal_100

Parameters 4 Million
FLOPs 267 Million
File Size 15.10 MB
Training Data ImageNet
Training Resources 4x4 TPU Pod
Training Time

Training Techniques RMSProp, Weight Decay
Architecture 1x1 Convolution, Batch Normalization, Convolution, Dense Connections, Depthwise Separable Convolution, Dropout, Global Average Pooling, Hard Swish, Inverted Residual Block, Residual Connection, ReLU, Softmax, Squeeze-and-Excitation Block
ID tf_mobilenetv3_large_minimal_100
LR 0.1
Dropout 0.8
Crop Pct 0.875
Momentum 0.9
Batch Size 4096
Image Size 224
Weight Decay 0.00001
Interpolation bilinear
SHOW MORE
SHOW LESS
tf_mobilenetv3_small_075

Parameters 2 Million
FLOPs 48 Million
File Size 7.86 MB
Training Data ImageNet
Training Resources 16x GPUs
Training Time

Training Techniques RMSProp, Weight Decay
Architecture 1x1 Convolution, Batch Normalization, Convolution, Dense Connections, Depthwise Separable Convolution, Dropout, Global Average Pooling, Hard Swish, Inverted Residual Block, Residual Connection, ReLU, Softmax, Squeeze-and-Excitation Block
ID tf_mobilenetv3_small_075
LR 0.045
Crop Pct 0.875
Momentum 0.9
Batch Size 4096
Image Size 224
Weight Decay 0.00004
Interpolation bilinear
RMSProp Decay 0.9
SHOW MORE
SHOW LESS
tf_mobilenetv3_small_100

Parameters 3 Million
FLOPs 65 Million
File Size 9.78 MB
Training Data ImageNet
Training Resources 16x GPUs
Training Time

Training Techniques RMSProp, Weight Decay
Architecture 1x1 Convolution, Batch Normalization, Convolution, Dense Connections, Depthwise Separable Convolution, Dropout, Global Average Pooling, Hard Swish, Inverted Residual Block, Residual Connection, ReLU, Softmax, Squeeze-and-Excitation Block
ID tf_mobilenetv3_small_100
LR 0.045
Crop Pct 0.875
Momentum 0.9
Batch Size 4096
Image Size 224
Weight Decay 0.00004
Interpolation bilinear
RMSProp Decay 0.9
SHOW MORE
SHOW LESS
tf_mobilenetv3_small_minimal_100

Parameters 2 Million
FLOPs 61 Million
File Size 7.88 MB
Training Data ImageNet
Training Resources 16x GPUs
Training Time

Training Techniques RMSProp, Weight Decay
Architecture 1x1 Convolution, Batch Normalization, Convolution, Dense Connections, Depthwise Separable Convolution, Dropout, Global Average Pooling, Hard Swish, Inverted Residual Block, Residual Connection, ReLU, Softmax, Squeeze-and-Excitation Block
ID tf_mobilenetv3_small_minimal_100
LR 0.045
Crop Pct 0.875
Momentum 0.9
Batch Size 4096
Image Size 224
Weight Decay 0.00004
Interpolation bilinear
RMSProp Decay 0.9
SHOW MORE
SHOW LESS
README.md

Summary

MobileNetV3 is a convolutional neural network that is designed for mobile phone CPUs. The network design includes the use of a hard swish activation and squeeze-and-excitation modules in the MBConv blocks.

How do I load this model?

To load a pretrained model:

import timm
m = timm.create_model('tf_mobilenetv3_large_075', pretrained=True)
m.eval()

Replace the model name with the variant you want to use, e.g. tf_mobilenetv3_large_075. You can find the IDs in the model summaries at the top of this page.

How do I train this model?

You can follow the timm recipe scripts for training a new model afresh.

Citation

@article{DBLP:journals/corr/abs-1905-02244,
  author    = {Andrew Howard and
               Mark Sandler and
               Grace Chu and
               Liang{-}Chieh Chen and
               Bo Chen and
               Mingxing Tan and
               Weijun Wang and
               Yukun Zhu and
               Ruoming Pang and
               Vijay Vasudevan and
               Quoc V. Le and
               Hartwig Adam},
  title     = {Searching for MobileNetV3},
  journal   = {CoRR},
  volume    = {abs/1905.02244},
  year      = {2019},
  url       = {http://arxiv.org/abs/1905.02244},
  archivePrefix = {arXiv},
  eprint    = {1905.02244},
  timestamp = {Tue, 12 Jan 2021 15:30:06 +0100},
  biburl    = {https://dblp.org/rec/journals/corr/abs-1905-02244.bib},
  bibsource = {dblp computer science bibliography, https://dblp.org}
}

Results

Image Classification on ImageNet

Image Classification on ImageNet
MODEL TOP 1 ACCURACY TOP 5 ACCURACY
tf_mobilenetv3_large_100 75.51% 92.61%
tf_mobilenetv3_large_075 73.45% 91.34%
tf_mobilenetv3_large_minimal_100 72.24% 90.64%
tf_mobilenetv3_small_100 67.92% 87.68%
tf_mobilenetv3_small_075 65.72% 86.13%
tf_mobilenetv3_small_minimal_100 62.91% 84.24%