Search Results for author: Venkata Vivek Thallam

Found 3 papers, 0 papers with code

QUIDAM: A Framework for Quantization-Aware DNN Accelerator and Model Co-Exploration

no code implementations30 Jun 2022 Ahmet Inci, Siri Garudanagiri Virupaksha, Aman Jain, Ting-Wu Chin, Venkata Vivek Thallam, Ruizhou Ding, Diana Marculescu

As the machine learning and systems communities strive to achieve higher energy-efficiency through custom deep neural network (DNN) accelerators, varied precision or quantization levels, and model compression techniques, there is a need for design space exploration frameworks that incorporate quantization-aware processing elements into the accelerator design space while having accurate and fast power, performance, and area models.

Model Compression Quantization

QADAM: Quantization-Aware DNN Accelerator Modeling for Pareto-Optimality

no code implementations20 May 2022 Ahmet Inci, Siri Garudanagiri Virupaksha, Aman Jain, Venkata Vivek Thallam, Ruizhou Ding, Diana Marculescu

We also show that the proposed lightweight processing elements (LightPEs) consistently achieve Pareto-optimal results in terms of accuracy and hardware-efficiency.

Quantization

QAPPA: Quantization-Aware Power, Performance, and Area Modeling of DNN Accelerators

no code implementations17 May 2022 Ahmet Inci, Siri Garudanagiri Virupaksha, Aman Jain, Venkata Vivek Thallam, Ruizhou Ding, Diana Marculescu

As the machine learning and systems community strives to achieve higher energy-efficiency through custom DNN accelerators and model compression techniques, there is a need for a design space exploration framework that incorporates quantization-aware processing elements into the accelerator design space while having accurate and fast power, performance, and area models.

Model Compression Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.