Search Results for author: Bingqian Lu

Found 4 papers, 2 papers with code

A Semi-Decoupled Approach to Fast and Optimal Hardware-Software Co-Design of Neural Accelerators

1 code implementation25 Mar 2022 Bingqian Lu, Zheyu Yan, Yiyu Shi, Shaolei Ren

We first perform neural architecture search to obtain a small set of optimal architectures for one accelerator candidate.

Neural Architecture Search

One Proxy Device Is Enough for Hardware-Aware Neural Architecture Search

1 code implementation1 Nov 2021 Bingqian Lu, Jianyi Yang, Weiwen Jiang, Yiyu Shi, Shaolei Ren

A key requirement of efficient hardware-aware NAS is the fast evaluation of inference latencies in order to rank different architectures.

Hardware Aware Neural Architecture Search Neural Architecture Search

Scaling Up Deep Neural Network Optimization for Edge Inference

no code implementations1 Sep 2020 Bingqian Lu, Jianyi Yang, Shaolei Ren

In the first approach, we reuse the performance predictors built on a proxy device, and leverage the performance monotonicity to scale up the DNN optimization without re-building performance predictors for each different device.

Quantization

A Note on Latency Variability of Deep Neural Networks for Mobile Inference

no code implementations29 Feb 2020 Luting Yang, Bingqian Lu, Shaolei Ren

Running deep neural network (DNN) inference on mobile devices, i. e., mobile inference, has become a growing trend, making inference less dependent on network connections and keeping private data locally.

Cannot find the paper you are looking for? You can Submit a new open access paper.