Search Results for author: Xin Men

Found 2 papers, 1 papers with code

ShortGPT: Layers in Large Language Models are More Redundant Than You Expect

no code implementations6 Mar 2024 Xin Men, Mingyu Xu, Qingyu Zhang, Bingning Wang, Hongyu Lin, Yaojie Lu, Xianpei Han, WeiPeng Chen

As Large Language Models (LLMs) continue to advance in performance, their size has escalated significantly, with current LLMs containing billions or even trillions of parameters.

Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.