Search Results for author: Jae W. Lee

Found 4 papers, 2 papers with code

Any-Precision LLM: Low-Cost Deployment of Multiple, Different-Sized LLMs

no code implementations16 Feb 2024 Yeonhong Park, Jake Hyun, SangLyul Cho, Bonggeun Sim, Jae W. Lee

Recently, considerable efforts have been directed towards compressing Large Language Models (LLMs), which showcase groundbreaking capabilities across diverse applications but entail significant deployment costs due to their large sizes.

Quantization

Ginex: SSD-enabled Billion-scale Graph Neural Network Training on a Single Machine via Provably Optimal In-memory Caching

1 code implementation19 Aug 2022 Yeonhong Park, Sunhong Min, Jae W. Lee

Thus, we propose Ginex, the first SSD-based GNN training system that can process billion-scale graph datasets on a single machine.

Compiler Optimization

L3: Accelerator-Friendly Lossless Image Format for High-Resolution, High-Throughput DNN Training

1 code implementation18 Aug 2022 Jonghyun Bae, Woohyeon Baek, Tae Jun Ham, Jae W. Lee

The decoding process of L3 is effectively parallelized on the accelerator, thus minimizing CPU intervention for data preparation during DNN training.

Vocal Bursts Intensity Prediction

A$^3$: Accelerating Attention Mechanisms in Neural Networks with Approximation

no code implementations22 Feb 2020 Tae Jun Ham, Sung Jun Jung, Seonghak Kim, Young H. Oh, Yeonhong Park, Yoonho Song, Jung-Hun Park, Sanghee Lee, Kyoung Park, Jae W. Lee, Deog-Kyoon Jeong

The attention mechanism is widely adopted by many state-of-the-art neural networks for computer vision, natural language processing, and machine translation, and accounts for a large portion of total execution time.

Machine Translation Translation

Cannot find the paper you are looking for? You can Submit a new open access paper.