no code implementations • NAACL 2022 • Marzieh Tahaei, Ella Charlaix, Vahid Nia, Ali Ghodsi, Mehdi Rezagholizadeh
We push the limits of state-of-the-art Transformer-based pre-trained language model compression using Kronecker decomposition.
no code implementations • 13 Sep 2021 • Marzieh S. Tahaei, Ella Charlaix, Vahid Partovi Nia, Ali Ghodsi, Mehdi Rezagholizadeh
We present our KroneckerBERT, a compressed version of the BERT_BASE model obtained using this framework.
1 code implementation • EMNLP 2021 • François Lagunas, Ella Charlaix, Victor Sanh, Alexander M. Rush
Pre-training has improved model accuracy for both classification and generation tasks at the cost of introducing much larger and slower models.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Gabriele Prato, Ella Charlaix, Mehdi Rezagholizadeh
State-of-the-art neural machine translation methods employ massive amounts of parameters.