Search Results for author: Ruizhou Ding

Found 14 papers, 6 papers with code

QUIDAM: A Framework for Quantization-Aware DNN Accelerator and Model Co-Exploration

no code implementations30 Jun 2022 Ahmet Inci, Siri Garudanagiri Virupaksha, Aman Jain, Ting-Wu Chin, Venkata Vivek Thallam, Ruizhou Ding, Diana Marculescu

As the machine learning and systems communities strive to achieve higher energy-efficiency through custom deep neural network (DNN) accelerators, varied precision or quantization levels, and model compression techniques, there is a need for design space exploration frameworks that incorporate quantization-aware processing elements into the accelerator design space while having accurate and fast power, performance, and area models.

Model Compression Quantization

QADAM: Quantization-Aware DNN Accelerator Modeling for Pareto-Optimality

no code implementations20 May 2022 Ahmet Inci, Siri Garudanagiri Virupaksha, Aman Jain, Venkata Vivek Thallam, Ruizhou Ding, Diana Marculescu

We also show that the proposed lightweight processing elements (LightPEs) consistently achieve Pareto-optimal results in terms of accuracy and hardware-efficiency.

Quantization

QAPPA: Quantization-Aware Power, Performance, and Area Modeling of DNN Accelerators

no code implementations17 May 2022 Ahmet Inci, Siri Garudanagiri Virupaksha, Aman Jain, Venkata Vivek Thallam, Ruizhou Ding, Diana Marculescu

As the machine learning and systems community strives to achieve higher energy-efficiency through custom DNN accelerators and model compression techniques, there is a need for a design space exploration framework that incorporates quantization-aware processing elements into the accelerator design space while having accurate and fast power, performance, and area models.

Model Compression Quantization

Single-Path Mobile AutoML: Efficient ConvNet Design and NAS Hyperparameter Optimization

1 code implementation1 Jul 2019 Dimitrios Stamoulis, Ruizhou Ding, Di Wang, Dimitrios Lymberopoulos, Bodhi Priyantha, Jie Liu, Diana Marculescu

In this work, we alleviate the NAS search cost down to less than 3 hours, while achieving state-of-the-art image classification results under mobile latency constraints.

Hyperparameter Optimization Image Classification +1

ViP: Virtual Pooling for Accelerating CNN-based Image Classification and Object Detection

1 code implementation19 Jun 2019 Zhuo Chen, Jiyuan Zhang, Ruizhou Ding, Diana Marculescu

In this paper, we propose Virtual Pooling (ViP), a model-level approach to improve speed and energy consumption of CNN-based image classification and object detection tasks, with a provable error bound.

General Classification Image Classification +3

Single-Path NAS: Device-Aware Efficient ConvNet Design

no code implementations10 May 2019 Dimitrios Stamoulis, Ruizhou Ding, Di Wang, Dimitrios Lymberopoulos, Bodhi Priyantha, Jie Liu, Diana Marculescu

Can we automatically design a Convolutional Network (ConvNet) with the highest image classification accuracy under the latency constraint of a mobile device?

General Classification Image Classification +1

FLightNNs: Lightweight Quantized Deep Neural Networks for Fast and Accurate Inference

no code implementations5 Apr 2019 Ruizhou Ding, Zeye Liu, Ting-Wu Chin, Diana Marculescu, R. D., Blanton

Over 46 FPGA-design experiments involving eight configurations and four data sets reveal that lightweight neural networks with a flexible $k$ value (dubbed FLightNNs) fully utilize the hardware resources on Field Programmable Gate Arrays (FPGAs), our experimental results show that FLightNNs can achieve 2$\times$ speedup when compared to lightweight NNs with $k=2$, with only 0. 1\% accuracy degradation.

Quantization

Single-Path NAS: Designing Hardware-Efficient ConvNets in less than 4 Hours

9 code implementations5 Apr 2019 Dimitrios Stamoulis, Ruizhou Ding, Di Wang, Dimitrios Lymberopoulos, Bodhi Priyantha, Jie Liu, Diana Marculescu

Can we automatically design a Convolutional Network (ConvNet) with the highest image classification accuracy under the runtime constraint of a mobile device?

General Classification Image Classification +1

Regularizing Activation Distribution for Training Binarized Deep Networks

1 code implementation CVPR 2019 Ruizhou Ding, Ting-Wu Chin, Zeye Liu, Diana Marculescu

Binarized Neural Networks (BNNs) can significantly reduce the inference latency and energy consumption in resource-constrained devices due to their pure-logical computation and fewer memory accesses.

AdaScale: Towards Real-time Video Object Detection Using Adaptive Scaling

no code implementations8 Feb 2019 Ting-Wu Chin, Ruizhou Ding, Diana Marculescu

In vision-enabled autonomous systems such as robots and autonomous cars, video object detection plays a crucial role, and both its speed and accuracy are important factors to provide reliable operation.

object-detection Video Object Detection

Understanding the Impact of Label Granularity on CNN-based Image Classification

1 code implementation21 Jan 2019 Zhuo Chen, Ruizhou Ding, Ting-Wu Chin, Diana Marculescu

In this paper, we conduct extensive experiments using various datasets to demonstrate and analyze how and why training based on fine-grain labeling, such as "Persian cat" can improve CNN accuracy on classifying coarse-grain classes, in this case "cat."

General Classification Image Classification

Differentiable Training for Hardware Efficient LightNNs

no code implementations NIPS Workshop CDNNRIA 2018 Ruizhou Ding, Zeye Liu, Ting-Wu Chin, Diana Marculescu, R.D. (Shawn) Blanton

To reduce runtime and resource utilization of Deep Neural Networks (DNNs) on customized hardware, LightNN has been proposed by constraining the weights of DNNs to be a sum of a limited number (denoted as $k\in\{1, 2\}$) of powers of 2.

Quantization

LightNN: Filling the Gap between Conventional Deep Neural Networks and Binarized Networks

no code implementations2 Dec 2017 Ruizhou Ding, Zeye Liu, Rongye Shi, Diana Marculescu, R. D. Blanton

For a fixed DNN configuration, LightNNs have better accuracy at a slight energy increase than BNNs, yet are more energy efficient with only slightly less accuracy than conventional DNNs.

Cannot find the paper you are looking for? You can Submit a new open access paper.