no code implementations • 30 Jun 2022 • Ahmet Inci, Siri Garudanagiri Virupaksha, Aman Jain, Ting-Wu Chin, Venkata Vivek Thallam, Ruizhou Ding, Diana Marculescu
As the machine learning and systems communities strive to achieve higher energy-efficiency through custom deep neural network (DNN) accelerators, varied precision or quantization levels, and model compression techniques, there is a need for design space exploration frameworks that incorporate quantization-aware processing elements into the accelerator design space while having accurate and fast power, performance, and area models.
no code implementations • 22 Jun 2022 • Yang Zhou, Feng Liang, Ting-Wu Chin, Diana Marculescu
Machine learning (ML) has entered the mobile era where an enormous number of ML models are deployed on edge devices.
no code implementations • 24 Apr 2021 • Ting-Wu Chin, Diana Marculescu, Ari S. Morcos
In this work, we propose width transfer, a technique that harnesses the assumptions that the optimized widths (or channel counts) are regular across sizes and depths.
no code implementations • 22 Aug 2020 • Ting-Wu Chin, Pierce I-Jen Chuang, Vikas Chandra, Diana Marculescu
Weight quantization for deep ConvNets has shown promising results for applications such as image classification and semantic segmentation and is especially important for applications where memory storage is limited.
2 code implementations • 23 Jul 2020 • Ting-Wu Chin, Ari S. Morcos, Diana Marculescu
In this work, we propose a general framework to enable joint optimization for both width configurations and weights of slimmable networks.
1 code implementation • 7 Feb 2020 • Ting-Wu Chin, Cha Zhang, Diana Marculescu
Fine-tuning through knowledge transfer from a pre-trained model on a large-scale dataset is a widely spread approach to effectively build models on small-scale datasets.
no code implementations • 25 Sep 2019 • Ting-Wu Chin, Pierce I-Jen Chuang, Vikas Chandra, Diana Marculescu
Weight Quantization for deep convolutional neural networks (CNNs) has shown promising results in compressing and accelerating CNN-powered applications such as semantic segmentation, gesture recognition, and scene understanding.
1 code implementation • CVPR 2020 • Ting-Wu Chin, Ruizhou Ding, Cha Zhang, Diana Marculescu
First, both the accuracy and the speed of ConvNets can affect the performance of the application.
no code implementations • 5 Apr 2019 • Ruizhou Ding, Zeye Liu, Ting-Wu Chin, Diana Marculescu, R. D., Blanton
Over 46 FPGA-design experiments involving eight configurations and four data sets reveal that lightweight neural networks with a flexible $k$ value (dubbed FLightNNs) fully utilize the hardware resources on Field Programmable Gate Arrays (FPGAs), our experimental results show that FLightNNs can achieve 2$\times$ speedup when compared to lightweight NNs with $k=2$, with only 0. 1\% accuracy degradation.
1 code implementation • CVPR 2019 • Ruizhou Ding, Ting-Wu Chin, Zeye Liu, Diana Marculescu
Binarized Neural Networks (BNNs) can significantly reduce the inference latency and energy consumption in resource-constrained devices due to their pure-logical computation and fewer memory accesses.
no code implementations • 8 Feb 2019 • Ting-Wu Chin, Ruizhou Ding, Diana Marculescu
In vision-enabled autonomous systems such as robots and autonomous cars, video object detection plays a crucial role, and both its speed and accuracy are important factors to provide reliable operation.
1 code implementation • 21 Jan 2019 • Zhuo Chen, Ruizhou Ding, Ting-Wu Chin, Diana Marculescu
In this paper, we conduct extensive experiments using various datasets to demonstrate and analyze how and why training based on fine-grain labeling, such as "Persian cat" can improve CNN accuracy on classifying coarse-grain classes, in this case "cat."
no code implementations • NIPS Workshop CDNNRIA 2018 • Ruizhou Ding, Zeye Liu, Ting-Wu Chin, Diana Marculescu, R.D. (Shawn) Blanton
To reduce runtime and resource utilization of Deep Neural Networks (DNNs) on customized hardware, LightNN has been proposed by constraining the weights of DNNs to be a sum of a limited number (denoted as $k\in\{1, 2\}$) of powers of 2.
no code implementations • 4 Oct 2018 • Ting-Wu Chin, Chia-Lin Yu, Matthew Halpern, Hasan Genc, Shiao-Li Tsao, Vijay Janapa Reddi
There is growing interest in object detection in advanced driver assistance systems and autonomous robots and vehicles.
1 code implementation • 1 Oct 2018 • Ting-Wu Chin, Cha Zhang, Diana Marculescu
Resource-efficient convolution neural networks enable not only the intelligence on edge devices but also opportunities in system-level optimization such as scheduling.
no code implementations • 5 Aug 2018 • Dimitrios Stamoulis, Ting-Wu Chin, Anand Krishnan Prakash, Haocheng Fang, Sribhuvan Sajja, Mitchell Bognar, Diana Marculescu
We cast the design of adaptive CNNs as a hyper-parameter optimization problem with respect to energy, accuracy, and communication constraints imposed by the mobile device.