no code implementations • 27 Apr 2024 • Kazuki Fujii, Taishi Nakamura, Mengsay Loem, Hiroki Iida, Masanari Ohi, Kakeru Hattori, Hirai Shota, Sakae Mizuki, Rio Yokota, Naoaki Okazaki
The results showed that the efficiency gained through vocabulary expansion had no negative impact on performance, except for the summarization task, and that the combined use of parallel corpora enhanced translation ability.
no code implementations • 27 Apr 2024 • Naoaki Okazaki, Kakeru Hattori, Hirai Shota, Hiroki Iida, Masanari Ohi, Kazuki Fujii, Taishi Nakamura, Mengsay Loem, Rio Yokota, Sakae Mizuki
Open Japanese large language models (LLMs) have been trained on the Japanese portions of corpora such as CC-100, mC4, and OSCAR.
no code implementations • 25 Feb 2024 • Masanari Ohi, Masahiro Kaneko, Ryuto Koike, Mengsay Loem, Naoaki Okazaki
In this paper, we investigate the presence and impact of likelihood bias in LLM-based evaluators.
no code implementations • 14 Nov 2023 • Mengsay Loem, Masahiro Kaneko, Naoaki Okazaki
Large Language Models (LLMs) can justify or critique their predictions through discussions with other models or humans, thereby enriching their intrinsic understanding of instances.
no code implementations • 29 May 2023 • Mengsay Loem, Masahiro Kaneko, Sho Takase, Naoaki Okazaki
Large-scale pre-trained language models such as GPT-3 have shown remarkable performance across various natural language processing tasks.
no code implementations • 27 Jul 2022 • Mengsay Loem, Sho Takase, Masahiro Kaneko, Naoaki Okazaki
Impressive performance of Transformer has been attributed to self-attention, where dependencies between entire input in a sequence are considered at every position.
no code implementations • NAACL (ACL) 2022 • Mengsay Loem, Sho Takase, Masahiro Kaneko, Naoaki Okazaki
Through experiments, we show that ExtraPhrase improves the performance of abstractive summarization tasks by more than 0. 50 points in ROUGE scores compared to the setting without data augmentation.