1 code implementation • 25 Mar 2022 • Bingqian Lu, Zheyu Yan, Yiyu Shi, Shaolei Ren
We first perform neural architecture search to obtain a small set of optimal architectures for one accelerator candidate.
1 code implementation • 1 Nov 2021 • Bingqian Lu, Jianyi Yang, Weiwen Jiang, Yiyu Shi, Shaolei Ren
A key requirement of efficient hardware-aware NAS is the fast evaluation of inference latencies in order to rank different architectures.
Hardware Aware Neural Architecture Search Neural Architecture Search
no code implementations • 1 Sep 2020 • Bingqian Lu, Jianyi Yang, Shaolei Ren
In the first approach, we reuse the performance predictors built on a proxy device, and leverage the performance monotonicity to scale up the DNN optimization without re-building performance predictors for each different device.
no code implementations • 29 Feb 2020 • Luting Yang, Bingqian Lu, Shaolei Ren
Running deep neural network (DNN) inference on mobile devices, i. e., mobile inference, has become a growing trend, making inference less dependent on network connections and keeping private data locally.