GreedyNAS: Towards Fast One-Shot NAS with Greedy Supernet

Training a supernet matters for one-shot neural architecture search (NAS) methods since it serves as a basic performance estimator for different architectures (paths). Current methods mainly hold the assumption that a supernet should give a reasonable ranking over all paths. They thus treat all paths equally, and spare much effort to train paths. However, it is harsh for a single supernet to evaluate accurately on such a huge-scale search space (e.g., $7^{21}$). In this paper, instead of covering all paths, we ease the burden of supernet by encouraging it to focus more on evaluation of those potentially-good ones, which are identified using a surrogate portion of validation data. Concretely, during training, we propose a multi-path sampling strategy with rejection, and greedily filter the weak paths. The training efficiency is thus boosted since the training space has been greedily shrunk from all paths to those potentially-good ones. Moreover, we further adopt an exploration and exploitation policy by introducing an empirical candidate path pool. Our proposed method GreedyNAS is easy-to-follow, and experimental results on ImageNet dataset indicate that it can achieve better Top-1 accuracy under same search space and FLOPs or latency level, but with only $\sim$60\% of supernet training cost. By searching on a larger space, our GreedyNAS can also obtain new state-of-the-art architectures.

PDF Abstract CVPR 2020 PDF CVPR 2020 Abstract

Datasets


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Image Classification ImageNet GreedyNAS-B Top 1 Accuracy 76.8% # 828
Number of params 5.2M # 410
GFLOPs 0.324 # 33
Image Classification ImageNet GreedyNAS-C Top 1 Accuracy 76.2% # 851
Number of params 4.7M # 392
GFLOPs 0.284 # 26
Image Classification ImageNet GreedyNAS-A Top 1 Accuracy 77.1% # 819
Number of params 6.5M # 444
Hardware Burden None # 1
Operations per network pass None # 1
GFLOPs 0.366 # 39
Neural Architecture Search ImageNet GreedyNAS-C Top-1 Error Rate 23.8 # 89
Accuracy 76.2 # 72
Params 4.7M # 47
Neural Architecture Search ImageNet GreedyNAS-B Top-1 Error Rate 23.2 # 79
Accuracy 76.8 # 63
Params 5.2M # 39
Neural Architecture Search ImageNet GreedyNAS-A Top-1 Error Rate 22.9 # 72
Accuracy 77.1 # 58
Params 6.5M # 16

Methods