Search Results for author: Mickael Seznec

Found 1 papers, 1 papers with code

SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models

4 code implementations18 Nov 2022 Guangxuan Xiao, Ji Lin, Mickael Seznec, Hao Wu, Julien Demouth, Song Han

We propose SmoothQuant, a training-free, accuracy-preserving, and general-purpose post-training quantization (PTQ) solution to enable 8-bit weight, 8-bit activation (W8A8) quantization for LLMs.

Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.