no code implementations • 28 Feb 2024 • Bedionita Soro, Bruno Andreis, Hayeon Lee, Song Chong, Frank Hutter, Sung Ju Hwang
By learning the distribution of a neural network on a variety pretrained models, our approach enables adaptive sampling weights for unseen datasets achieving faster convergence and reaching competitive performance.
1 code implementation • 5 Jul 2022 • Mingyu Kim, Jihwan Oh, Yongsik Lee, Joonkee Kim, SeongHwan Kim, Song Chong, Se-Young Yun
This challenge, on the other hand, is interested in the exploration capability of MARL algorithms to efficiently learn implicit multi-stage tasks and environmental factors as well as micro-control.
Ranked #1 on SMAC+ on Off_Superhard_parallel
1 code implementation • NeurIPS 2021 • Hayeon Lee, Sewoong Lee, Song Chong, Sung Ju Hwang
To overcome such limitations, we propose Hardware-adaptive Efficient Latency Predictor (HELP), which formulates the device-specific latency estimation problem as a meta-learning problem, such that we can estimate the latency of a model's performance for a given task on an unseen device with a few samples.
1 code implementation • 16 Jun 2021 • Hayeon Lee, Sewoong Lee, Song Chong, Sung Ju Hwang
To overcome such limitations, we propose Hardware-adaptive Efficient Latency Predictor (HELP), which formulates the device-specific latency estimation problem as a meta-learning problem, such that we can estimate the latency of a model's performance for a given task on an unseen device with a few samples.
no code implementations • 11 Apr 2017 • Kimin Lee, Jaehyung Kim, Song Chong, Jinwoo Shin
In this paper, we aim at developing efficient training methods for SFNN, in particular using known architectures and pre-trained parameters of DNN.