no code implementations • 23 Mar 2024 • Phong Nguyen-Thuan Do, Son Quoc Tran, Phu Gia Hoang, Kiet Van Nguyen, Ngan Luu-Thuy Nguyen
The success of Natural Language Understanding (NLU) benchmarks in various languages, such as GLUE for English, CLUE for Chinese, KLUE for Korean, and IndoNLU for Indonesian, has facilitated the evaluation of new NLU models across a wide range of tasks.
1 code implementation • 10 Sep 2023 • Son Quoc Tran, Gia-Huy Do, Phong Nguyen-Thuan Do, Matt Kretchmar, Xinya Du
In this paper, we demonstrate the usefulness of this AGent pipeline by creating two sets of unanswerable questions from answerable questions in SQuAD and HotpotQA.
1 code implementation • 8 Aug 2023 • Son Quoc Tran, Matt Kretchmar
Machine Reading Comprehension (MRC) models tend to take advantage of spurious correlations (also known as dataset bias or annotation artifacts in the research community).
no code implementations • 16 Mar 2023 • Son Quoc Tran, Phong Nguyen-Thuan Do, Kiet Van Nguyen, Ngan Luu-Thuy Nguyen
From the analysis results, we suggest new directions for developing Vietnamese language models.
no code implementations • 31 Jan 2023 • Son Quoc Tran, Phong Nguyen-Thuan Do, Uyen Le, Matt Kretchmar
Pretrained language models have achieved super-human performances on many Machine Reading Comprehension (MRC) benchmarks.
no code implementations • 22 Mar 2022 • Kiet Van Nguyen, Son Quoc Tran, Luan Thanh Nguyen, Tin Van Huynh, Son T. Luu, Ngan Luu-Thuy Nguyen
To address the weakness, we provide the research community with a benchmark dataset named UIT-ViQuAD 2. 0 for evaluating the MRC task and question answering systems for the Vietnamese language.