1 code implementation • 21 Feb 2024 • Chenyang Song, Xu Han, Zhengyan Zhang, Shengding Hu, Xiyu Shi, Kuai Li, Chen Chen, Zhiyuan Liu, Guangli Li, Tao Yang, Maosong Sun
Some recent efforts have explored introducing ReLU or its variants as the substitutive activation function to help LLMs achieve activation sparsity and inference acceleration, but few can simultaneously obtain high sparsity and comparable model performance.
no code implementations • 1 Apr 2021 • Jiansong Li, Xiao Dong, Guangli Li, Peng Zhao, Xueying Wang, Xiaobing Chen, Xianzhi Yu, Yongxin Yang, Zihan Jiang, Wei Cao, Lei Liu, Xiaobing Feng
The training of deep neural networks (DNNs) is usually memory-hungry due to the limited device memory capacity of DNN accelerators.
no code implementations • 30 Oct 2020 • Guangli Li, Xiu Ma, Xueying Wang, Lei Liu, Jingling Xue, Xiaobing Feng
The increasing computational cost of deep neural network models limits the applicability of intelligent applications on resource-constrained edge devices.
no code implementations • 19 Mar 2020 • Guangli Li, Lei Liu, Xueying Wang, Xiu Ma, Xiaobing Feng
Accelerating deep convolutional neural networks has become an active topic and sparked an interest in academia and industry.
no code implementations • 17 Jan 2019 • Xueying Wang, Lei Liu, Guangli Li, Xiao Dong, Peng Zhao, Xiaobing Feng
Background subtraction is a significant component of computer vision systems.
no code implementations • 16 Dec 2018 • Guangli Li, Lei Liu, Xueying Wang, Xiao Dong, Peng Zhao, Xiaobing Feng
By analyzing the characteristics of layers in DNNs, an auto-tuning neural network quantization framework for collaborative inference is proposed.