DPN

Last updated on Feb 14, 2021

dpn107

Parameters 87 Million
FLOPs 24 Billion
File Size 332.46 MB
Training Data ImageNet
Training Resources 40x K80 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture Batch Normalization, Convolution, DPN Block, Dense Connections, Global Average Pooling, Max Pooling, Softmax
ID dpn107
LR 0.316
Layers 107
Crop Pct 0.875
Batch Size 1280
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
dpn131

Parameters 79 Million
FLOPs 21 Billion
File Size 303.28 MB
Training Data ImageNet
Training Resources 40x K80 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture Batch Normalization, Convolution, DPN Block, Dense Connections, Global Average Pooling, Max Pooling, Softmax
ID dpn131
LR 0.316
Layers 131
Crop Pct 0.875
Batch Size 960
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
dpn68

Parameters 13 Million
FLOPs 3 Billion
File Size 48.41 MB
Training Data ImageNet
Training Resources 40x K80 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture Batch Normalization, Convolution, DPN Block, Dense Connections, Global Average Pooling, Max Pooling, Softmax
ID dpn68
LR 0.316
Layers 68
Crop Pct 0.875
Batch Size 1280
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
dpn68b

Parameters 13 Million
FLOPs 3 Billion
File Size 48.43 MB
Training Data ImageNet
Training Resources 40x K80 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture Batch Normalization, Convolution, DPN Block, Dense Connections, Global Average Pooling, Max Pooling, Softmax
ID dpn68b
LR 0.316
Layers 68
Crop Pct 0.875
Batch Size 1280
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
dpn92

Parameters 38 Million
FLOPs 8 Billion
File Size 144.24 MB
Training Data ImageNet
Training Resources 40x K80 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture Batch Normalization, Convolution, DPN Block, Dense Connections, Global Average Pooling, Max Pooling, Softmax
ID dpn92
LR 0.316
Layers 92
Crop Pct 0.875
Batch Size 1280
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
dpn98

Parameters 62 Million
FLOPs 15 Billion
File Size 235.58 MB
Training Data ImageNet
Training Resources 40x K80 GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture Batch Normalization, Convolution, DPN Block, Dense Connections, Global Average Pooling, Max Pooling, Softmax
ID dpn98
LR 0.4
Layers 98
Crop Pct 0.875
Batch Size 1280
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
README.md

Summary

A Dual Path Network (DPN) is a convolutional neural network which presents a new topology of connection paths internally. The intuition is that ResNets enables feature re-usage while DenseNet enables new feature exploration, and both are important for learning good representations. To enjoy the benefits from both path topologies, Dual Path Networks share common features while maintaining the flexibility to explore new features through dual path architectures.

The principal building block is an DPN Block.

How do I load this model?

To load a pretrained model:

import timm
m = timm.create_model('dpn68', pretrained=True)
m.eval()

Replace the model name with the variant you want to use, e.g. dpn68. You can find the IDs in the model summaries at the top of this page.

How do I train this model?

You can follow the timm recipe scripts for training a new model afresh.

Citation

@misc{chen2017dual,
      title={Dual Path Networks}, 
      author={Yunpeng Chen and Jianan Li and Huaxin Xiao and Xiaojie Jin and Shuicheng Yan and Jiashi Feng},
      year={2017},
      eprint={1707.01629},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Results

Image Classification on ImageNet

Image Classification on ImageNet
MODEL TOP 1 ACCURACY TOP 5 ACCURACY
dpn107 80.16% 94.91%
dpn92 79.99% 94.84%
dpn131 79.83% 94.71%
dpn98 79.65% 94.61%
dpn68b 79.21% 94.42%
dpn68 76.31% 92.97%