Search Results for author: Sugyeong Eo

Found 17 papers, 1 papers with code

Dealing with the Paradox of Quality Estimation

no code implementations MTSummit 2021 Sugyeong Eo, Chanjun Park, Hyeonseok Moon, Jaehyung Seo, Heuiseok Lim

In quality estimation (QE), the quality of translation can be predicted by referencing the source sentence and the machine translation (MT) output without access to the reference sentence.

Machine Translation Sentence +1

BTS: Back TranScription for Speech-to-Text Post-Processor using Text-to-Speech-to-Text

no code implementations ACL (WAT) 2021 Chanjun Park, Jaehyung Seo, Seolhwa Lee, Chanhee Lee, Hyeonseok Moon, Sugyeong Eo, Heuiseok Lim

Automatic speech recognition (ASR) is arguably the most critical component of such systems, as errors in speech recognition propagate to the downstream components and drastically degrade the user experience.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

A Dog Is Passing Over The Jet? A Text-Generation Dataset for Korean Commonsense Reasoning and Evaluation

no code implementations Findings (NAACL) 2022 Jaehyung Seo, Seounghoon Lee, Chanjun Park, Yoonna Jang, Hyeonseok Moon, Sugyeong Eo, Seonmin Koo, Heuiseok Lim

However, Korean pretrained language models still struggle to generate a short sentence with a given condition based on compositionality and commonsense reasoning (i. e., generative commonsense reasoning).

Language Modelling Natural Language Understanding +2

Focus on FoCus: Is FoCus focused on Context, Knowledge and Persona?

no code implementations CCGPK (COLING) 2022 SeungYoon Lee, Jungseob Lee, Chanjun Park, Sugyeong Eo, Hyeonseok Moon, Jaehyung Seo, Jeongbae Park, Heuiseok Lim

As a result of the experiment, we present that the FoCus model could not correctly blend the knowledge according to the input dialogue and that the dataset design is unsuitable for the multi-turn conversation.

Dialogue Generation Question Answering

Synthetic Alone: Exploring the Dark Side of Synthetic Data for Grammatical Error Correction

no code implementations26 Jun 2023 Chanjun Park, Seonmin Koo, Seolhwa Lee, Jaehyung Seo, Sugyeong Eo, Hyeonseok Moon, Heuiseok Lim

Data-centric AI approach aims to enhance the model performance without modifying the model and has been shown to impact model performance positively.

Grammatical Error Correction

Towards Diverse and Effective Question-Answer Pair Generation from Children Storybooks

1 code implementation11 Jun 2023 Sugyeong Eo, Hyeonseok Moon, Jinsung Kim, Yuna Hur, Jeongwook Kim, Songeun Lee, Changwoo Chun, Sungsoo Park, Heuiseok Lim

In this paper, we propose a QAG framework that enhances QA type diversity by producing different interrogative sentences and implicit/explicit answers.

Self-Improving-Leaderboard(SIL): A Call for Real-World Centric Natural Language Processing Leaderboards

no code implementations20 Mar 2023 Chanjun Park, Hyeonseok Moon, Seolhwa Lee, Jaehyung Seo, Sugyeong Eo, Heuiseok Lim

Leaderboard systems allow researchers to objectively evaluate Natural Language Processing (NLP) models and are typically used to identify models that exhibit superior performance on a given task in a predetermined setting.

A Self-Supervised Automatic Post-Editing Data Generation Tool

no code implementations24 Nov 2021 Hyeonseok Moon, Chanjun Park, Sugyeong Eo, Jaehyung Seo, Seungjun Lee, Heuiseok Lim

Data building for automatic post-editing (APE) requires extensive and expert-level human effort, as it contains an elaborate process that involves identifying errors in sentences and providing suitable revisions.

Automatic Post-Editing

A New Tool for Efficiently Generating Quality Estimation Datasets

no code implementations1 Nov 2021 Sugyeong Eo, Chanjun Park, Jaehyung Seo, Hyeonseok Moon, Heuiseok Lim

Building of data for quality estimation (QE) training is expensive and requires significant human labor.

Data Augmentation

Automatic Knowledge Augmentation for Generative Commonsense Reasoning

no code implementations30 Oct 2021 Jaehyung Seo, Chanjun Park, Sugyeong Eo, Hyeonseok Moon, Heuiseok Lim

Generative commonsense reasoning is the capability of a language model to generate a sentence with a given concept-set that is based on commonsense knowledge.

Language Modelling Sentence

How should human translation coexist with NMT? Efficient tool for building high quality parallel corpus

no code implementations30 Oct 2021 Chanjun Park, Seolhwa Lee, Hyeonseok Moon, Sugyeong Eo, Jaehyung Seo, Heuiseok Lim

This paper proposes a tool for efficiently constructing high-quality parallel corpora with minimizing human labor and making this tool publicly available.

Machine Translation NMT +1

Should we find another model?: Improving Neural Machine Translation Performance with ONE-Piece Tokenization Method without Model Modification

no code implementations NAACL 2021 Chanjun Park, Sugyeong Eo, Hyeonseok Moon, Heuiseok Lim

We derive an optimal subword tokenization result for Korean-English machine translation by conducting a case study that combines the subword tokenization method, morphological segmentation, and vocabulary method.

Machine Translation Translation

Cannot find the paper you are looking for? You can Submit a new open access paper.