no code implementations • 31 Oct 2023 • Gabrielle Cohn, Rishika Agarwal, Deepanshu Gupta, Siddharth Patwardhan
We introduce EELBERT, an approach for compression of transformer-based models (e. g., BERT), with minimal impact on the accuracy of downstream tasks.