Search Results for author: Shouda Liu

Found 3 papers, 2 papers with code

decoupleQ: Towards 2-bit Post-Training Uniform Quantization via decoupling Parameters into Integer and Floating Points

1 code implementation19 Apr 2024 Yi Guo, Fanliu Kong, Xiaoyang Li, Hui Li, Wei Chen, Xiaogang Tian, Jinping Cai, Yang Zhang, Shouda Liu

However, existing quantization schemes suffer from significant accuracy degradation at very low bits, or require some additional computational overhead when deployed, making it difficult to be applied to large-scale applications in industry.

Quantization

Accurate LoRA-Finetuning Quantization of LLMs via Information Retention

1 code implementation8 Feb 2024 Haotong Qin, Xudong Ma, Xingyu Zheng, Xiaoyang Li, Yang Zhang, Shouda Liu, Jie Luo, Xianglong Liu, Michele Magno

This paper proposes a novel IR-QLoRA for pushing quantized LLMs with LoRA to be highly accurate through information retention.

Quantization

RdimKD: Generic Distillation Paradigm by Dimensionality Reduction

no code implementations14 Dec 2023 Yi Guo, Yiqian He, Xiaoyang Li, Haotong Qin, Van Tung Pham, Yang Zhang, Shouda Liu

Knowledge Distillation (KD) emerges as one of the most promising compression technologies to run advanced deep neural networks on resource-limited devices.

Dimensionality Reduction Knowledge Distillation

Cannot find the paper you are looking for? You can Submit a new open access paper.