no code implementations • 24 Dec 2021 • Souvik Kundu, Shikai Wang, Qirui Sun, Peter A. Beerel, Massoud Pedram
Compared to the baseline FP-32 models, BMPQ can yield models that have 15. 4x fewer parameter bits with a negligible drop in accuracy.
no code implementations • NeurIPS 2021 • Souvik Kundu, Qirui Sun, Yao Fu, Massoud Pedram, Peter Beerel
Knowledge distillation (KD) has recently been identified as a method that can unintentionally leak private information regarding the details of a teacher model to an unauthorized student.