no code implementations • 27 Jan 2024 • Seungcheol Park, Jaehyeon Choi, Sojin Lee, U Kang
How can we compress language models without sacrificing accuracy?
Knowledge Distillation Quantization