Search Results for author: Jungyu Jin

Found 1 papers, 1 papers with code

OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Models

2 code implementations4 Jun 2023 Changhun Lee, Jungyu Jin, Taesu Kim, HyungJun Kim, Eunhyeok Park

Large language models (LLMs) with hundreds of billions of parameters require powerful server-grade GPUs for inference, limiting their practical deployment.

Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.