Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning

17 Sep 2020 Bingbing Li Zhenglun Kong Tianyun Zhang Ji Li Zhengang Li Hang Liu Caiwen Ding

Pre-trained large-scale language models have increasingly demonstrated high accuracy on many natural language processing (NLP) tasks. However, the limited weight storage and computational speed on hardware platforms have impeded the popularity of pre-trained models, especially in the era of edge computing... (read more)

PDF Abstract
No code implementations yet. Submit your code now


Results from the Paper

  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods used in the Paper