1 code implementation • 25 Oct 2022 • Soyoung Yoon, Sungjoon Park, Gyuwan Kim, Junhee Cho, Kihyo Park, Gyutae Kim, Minjoon Seo, Alice Oh
We show that the model trained with our datasets significantly outperforms the currently used statistical Korean GEC system (Hanspell) on a wider range of error types, demonstrating the diversity and usefulness of the datasets.
no code implementations • 25 Oct 2022 • Gyuwan Kim, Jinhyuk Lee, Barlas Oguz, Wenhan Xiong, Yizhe Zhang, Yashar Mehdad, William Yang Wang
Building dense retrievers requires a series of standard procedures, including training and validating neural models and creating indexes for efficient search.
no code implementations • 18 Nov 2021 • Shira Guskin, Moshe Wasserblat, Ke Ding, Gyuwan Kim
Additionally, a separate model must be trained for each inference scenario with its distinct computational budget.
1 code implementation • Findings (ACL) 2021 • Soyoung Yoon, Gyuwan Kim, Kyumin Park
Data augmentation with mixup has shown to be effective on various computer vision tasks.
1 code implementation • NAACL 2022 • Jungsoo Park, Gyuwan Kim, Jaewoo Kang
Consistency training regularizes a model by enforcing predictions of original and perturbed inputs to be similar.
1 code implementation • 25 Oct 2020 • Seongbin Kim, Gyuwan Kim, Seongjin Shin, Sangmin Lee
End-to-end approaches open a new way for more accurate and efficient spoken language understanding (SLU) systems by alleviating the drawbacks of traditional pipeline systems.
Ranked #3 on Spoken Language Understanding on Fluent Speech Commands (using extra training data)
Automatic Speech Recognition Automatic Speech Recognition (ASR) +4
no code implementations • 23 Oct 2020 • Minjeong Kim, Gyuwan Kim, Sang-Woo Lee, Jung-Woo Ha
Language model pre-training has shown promising results in various downstream tasks.
1 code implementation • ACL 2021 • Gyuwan Kim, Kyunghyun Cho
We then conduct a multi-objective evolutionary search to find a length configuration that maximizes the accuracy and minimizes the efficiency metric under any given computational budget.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Gyuwan Kim, Tae-Hwan Jung
Product key memory (PKM) proposed by Lample et al. (2019) enables to improve prediction accuracy by increasing model capacity efficiently with insignificant computational overhead.
4 code implementations • ICLR 2021 • Byeongho Heo, Sanghyuk Chun, Seong Joon Oh, Dongyoon Han, Sangdoo Yun, Gyuwan Kim, Youngjung Uh, Jung-Woo Ha
Because of the scale invariance, this modification only alters the effective step sizes without changing the effective update directions, thus enjoying the original convergence properties of GD optimizers.
3 code implementations • ACL 2020 • Sungdong Kim, Sohee Yang, Gyuwan Kim, Sang-Woo Lee
This mechanism consists of two steps: (1) predicting state operation on each of the memory slots, and (2) overwriting the memory with new values, of which only a few are generated according to the predicted state operations.
Ranked #10 on Multi-domain Dialogue State Tracking on MULTIWOZ 2.0
Dialogue State Tracking Multi-domain Dialogue State Tracking
1 code implementation • IJCNLP 2019 • Gyuwan Kim
We present how to utilize subword language models for the fast and accurate generation of query completion candidates.
no code implementations • 12 Nov 2016 • Jangho Lee, Gyuwan Kim, Jaeyoon Yoo, Changwoo Jung, Minseok Kim, Sungroh Yoon
Under the assumption that using such an automatically generated dataset could relieve the burden of manual question-answer generation, we tried to use this dataset to train an instance of Watson and checked the training efficiency and accuracy.
no code implementations • 6 Nov 2016 • Gyuwan Kim, Hayoon Yi, Jangho Lee, Yunheung Paek, Sungroh Yoon
In computer security, designing a robust intrusion detection system is one of the most fundamental and important problems.
1 code implementation • 16 Nov 2015 • Sunyoung Kwon, Gyuwan Kim, Byunghan Lee, Jongsik Chun, Sungroh Yoon, Young-Han Kim
Motivated by the need for fast and accurate classification of unlabeled nucleotide sequences on a large scale, we developed NASCUP, a new classification method that captures statistical structures of nucleotide sequences by compact context-tree models and universal probability from information theory.
Genomics Information Theory Information Theory