Search Results for author: Minki Kang

Found 14 papers, 8 papers with code

Knowledge-Augmented Language Model Verification

1 code implementation19 Oct 2023 Jinheon Baek, Soyeong Jeong, Minki Kang, Jong C. Park, Sung Ju Hwang

Recent Language Models (LMs) have shown impressive capabilities in generating texts with the knowledge internalized in parameters.

Language Modelling Question Answering +1

Face-StyleSpeech: Improved Face-to-Voice latent mapping for Natural Zero-shot Speech Synthesis from a Face Image

no code implementations25 Sep 2023 Minki Kang, Wooseok Han, Eunho Yang

The prosody encoder is specifically designed to model prosodic features that are not captured only with a face image, allowing the face encoder to focus solely on capturing the speaker identity from the face image.

Speech Synthesis

Knowledge Graph-Augmented Language Models for Knowledge-Grounded Dialogue Generation

no code implementations30 May 2023 Minki Kang, Jin Myung Kwak, Jinheon Baek, Sung Ju Hwang

To overcome this limitation, we propose SUbgraph Retrieval-augmented GEneration (SURGE), a framework for generating context-relevant and knowledge-grounded dialogues with the KG.

Contrastive Learning Dialogue Generation +3

Knowledge-Augmented Reasoning Distillation for Small Language Models in Knowledge-Intensive Tasks

1 code implementation NeurIPS 2023 Minki Kang, Seanie Lee, Jinheon Baek, Kenji Kawaguchi, Sung Ju Hwang

Large Language Models (LLMs) have shown promising performance in knowledge-intensive reasoning tasks that require a compound understanding of knowledge.

Memorization StrategyQA

ZET-Speech: Zero-shot adaptive Emotion-controllable Text-to-Speech Synthesis with Diffusion and Style-based Models

no code implementations23 May 2023 Minki Kang, Wooseok Han, Sung Ju Hwang, Eunho Yang

Emotional Text-To-Speech (TTS) is an important task in the development of systems (e. g., human-like dialogue agents) that require natural and emotional speech.

Speech Synthesis Text-To-Speech Synthesis

Grad-StyleSpeech: Any-speaker Adaptive Text-to-Speech Synthesis with Diffusion Models

no code implementations17 Nov 2022 Minki Kang, Dongchan Min, Sung Ju Hwang

There has been a significant progress in Text-To-Speech (TTS) synthesis technology in recent years, thanks to the advancement in neural generative modeling.

Speech Synthesis Text-To-Speech Synthesis

Self-Distillation for Further Pre-training of Transformers

no code implementations30 Sep 2022 Seanie Lee, Minki Kang, Juho Lee, Sung Ju Hwang, Kenji Kawaguchi

Pre-training a large transformer model on a massive amount of unlabeled data and fine-tuning it on labeled datasets for diverse downstream tasks has proven to be a successful strategy, for a variety of vision and natural language processing tasks.

text-classification Text Classification

KALA: Knowledge-Augmented Language Model Adaptation

1 code implementation NAACL 2022 Minki Kang, Jinheon Baek, Sung Ju Hwang

Pre-trained language models (PLMs) have achieved remarkable success on various natural language understanding tasks.

Domain Adaptation General Knowledge +6

Learning to Perturb Word Embeddings for Out-of-distribution QA

1 code implementation ACL 2021 Seanie Lee, Minki Kang, Juho Lee, Sung Ju Hwang

QA models based on pretrained language mod-els have achieved remarkable performance on various benchmark datasets. However, QA models do not generalize well to unseen data that falls outside the training distribution, due to distributional shifts. Data augmentation (DA) techniques which drop/replace words have shown to be effective in regularizing the model from overfitting to the training data. Yet, they may adversely affect the QA tasks since they incur semantic changes that may lead to wrong answers for the QA task.

Data Augmentation Domain Generalization +1

Accurate Learning of Graph Representations with Graph Multiset Pooling

1 code implementation ICLR 2021 Jinheon Baek, Minki Kang, Sung Ju Hwang

Graph neural networks have been widely used on modeling graph data, achieving impressive results on node classification and link prediction tasks.

Graph Classification Graph Clustering +5

Neural Mask Generator: Learning to Generate Adaptive Word Maskings for Language Model Adaptation

1 code implementation EMNLP 2020 Minki Kang, Moonsu Han, Sung Ju Hwang

We propose a method to automatically generate a domain- and task-adaptive maskings of the given text for self-supervised pre-training, such that we can effectively adapt the language model to a particular target task (e. g. question answering).

Language Modelling Question Answering +4

Episodic Memory Reader: Learning What to Remember for Question Answering from Streaming Data

1 code implementation ACL 2019 Moonsu Han, Minki Kang, Hyunwoo Jung, Sung Ju Hwang

We consider a novel question answering (QA) task where the machine needs to read from large streaming data (long documents or videos) without knowing when the questions will be given, which is difficult to solve with existing QA methods due to their lack of scalability.

Question Answering Reading Comprehension +2

Learning What to Remember: Long-term Episodic Memory Networks for Learning from Streaming Data

no code implementations ICLR 2019 Hyunwoo Jung, Moonsu Han, Minki Kang, Sungju Hwang

We tackle this problem by proposing a memory network fit for long-term lifelong learning scenario, which we refer to as Long-term Episodic Memory Networks (LEMN), that features a RNN-based retention agent that learns to replace less important memory entries based on the retention probability generated on each entry that is learned to identify data instances of generic importance relative to other memory entries, as well as its historical importance.

Question Answering Scheduling

Cannot find the paper you are looking for? You can Submit a new open access paper.