no code implementations • 1 Apr 2021 • Jiansong Li, Xiao Dong, Guangli Li, Peng Zhao, Xueying Wang, Xiaobing Chen, Xianzhi Yu, Yongxin Yang, Zihan Jiang, Wei Cao, Lei Liu, Xiaobing Feng
The training of deep neural networks (DNNs) is usually memory-hungry due to the limited device memory capacity of DNN accelerators.
no code implementations • 8 Nov 2019 • Zihan Jiang, Jiansong Li, Jiangfeng Zhan
To reveal this pitfall, we evaluates several frequently-used optimizations on a typical AI accelerator and quantifies their impact on accuracy and throughout under representative DL inference workloads.