Search Results for author: Ranggi Hwang

Found 5 papers, 1 papers with code

Pre-gated MoE: An Algorithm-System Co-Design for Fast and Scalable Mixture-of-Expert Inference

1 code implementation23 Aug 2023 Ranggi Hwang, Jianyu Wei, Shijie Cao, Changho Hwang, Xiaohu Tang, Ting Cao, Mao Yang

To tackle the high compute requirements of LLMs, the Mixture-of-Experts (MoE) architecture was introduced which is able to scale its model size without proportionally scaling up its computational requirements.

DiVa: An Accelerator for Differentially Private Machine Learning

no code implementations26 Aug 2022 Beomsik Park, Ranggi Hwang, Dongho Yoon, Yoonhyuk Choi, Minsoo Rhu

The widespread deployment of machine learning (ML) is raising serious concerns on protecting the privacy of users who contributed to the collection of training data.

GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for Memory-Efficient Graph Convolutional Neural Networks

no code implementations1 Mar 2022 Ranggi Hwang, Minhoo Kang, Jiwon Lee, Dongyun Kam, Youngjoo Lee, Minsoo Rhu

Graph convolutional neural networks (GCNs) have emerged as a key technology in various application domains where the input data is relational.

Centaur: A Chiplet-based, Hybrid Sparse-Dense Accelerator for Personalized Recommendations

no code implementations12 May 2020 Ranggi Hwang, Taehun Kim, Youngeun Kwon, Minsoo Rhu

Personalized recommendations are the backbone machine learning (ML) algorithm that powers several important application domains (e. g., ads, e-commerce, etc) serviced from cloud datacenters.

Cannot find the paper you are looking for? You can Submit a new open access paper.