1 code implementation • Findings (ACL) 2022 • Yong-Ho Jung, Jun-Hyung Park, Joon-Young Choi, Mingyu Lee, Junho Kim, Kang-Min Kim, SangKeun Lee
Commonsense inference poses a unique challenge to reason and generate the physical, social, and causal conditions of a given event.
no code implementations • EMNLP (ACL) 2021 • San-Hee Park, Kang-Min Kim, Seonhee Cho, Jun-Hyung Park, Hyuntae Park, Hyuna Kim, Seongwon Chung, SangKeun Lee
Warning: This manuscript contains a certain level of offensive expression.
1 code implementation • 15 Dec 2022 • Mingyu Lee, Jun-Hyung Park, Junho Kim, Kang-Min Kim, SangKeun Lee
Masked language modeling (MLM) has been widely used for pre-training effective bidirectional representations, but incurs substantial training costs.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Kang-Min Kim, Bumsu Hyeon, Yeachan Kim, Jun-Hyung Park, SangKeun Lee
In addition, we propose a weakly supervised pretraining, where labels for text classification are obtained automatically from an existing approach.
no code implementations • ACL 2020 • Yeachan Kim, Kang-Min Kim, SangKeun Lee
However, unlike prior works that assign the same length of codes to all words, we adaptively assign different lengths of codes to each word by learning downstream tasks.
no code implementations • LREC 2020 • Yeachan Kim, Kang-Min Kim, SangKeun Lee
In the first stage, we learn subword embeddings from the pre-trained word embeddings by using an additive composition function of subwords.
no code implementations • COLING 2018 • Yeachan Kim, Kang-Min Kim, Ji-Min Lee, SangKeun Lee
Unlike previous models that learn word representations from a large corpus, we take a set of pre-trained word embeddings and generalize it to word entries, including OOV words.
no code implementations • 3 Apr 2018 • Kang-Min Kim, Aliyeva Dinara, Byung-Ju Choi, SangKeun Lee
However, these approaches are limited to small- or moderate-scale text classification.