Search Results for author: Hongjie Wang

Found 3 papers, 0 papers with code

Zero-TPrune: Zero-Shot Token Pruning through Leveraging of the Attention Graph in Pre-Trained Transformers

no code implementations27 May 2023 Hongjie Wang, Bhishma Dedhia, Niraj K. Jha

Deployment of Transformer models on edge devices is becoming increasingly challenging due to the exponentially growing inference cost that scales quadratically with the number of tokens in the input sequence.

A New MRAM-based Process In-Memory Accelerator for Efficient Neural Network Training with Floating Point Precision

no code implementations2 Mar 2020 Hongjie Wang, Yang Zhao, Chaojian Li, Yue Wang, Yingyan Lin

The excellent performance of modern deep neural networks (DNNs) comes at an often prohibitive training cost, limiting the rapid development of DNN innovations and raising various environmental concerns.

Efficient Neural Network

Cannot find the paper you are looking for? You can Submit a new open access paper.