Search Results for author: Taehwan Kwon

Found 7 papers, 2 papers with code

Hexa: Self-Improving for Knowledge-Grounded Dialogue System

no code implementations10 Oct 2023 DaeJin Jo, Daniel Wontae Nam, Gunsoo Han, Kyoung-Woon On, Taehwan Kwon, Seungeun Rho, Sungwoong Kim

A common practice in knowledge-grounded dialogue generation is to explicitly utilize intermediate steps (e. g., web-search, memory retrieval) with modular approaches.

Dialogue Generation Retrieval

LECO: Learnable Episodic Count for Task-Specific Intrinsic Reward

1 code implementation11 Oct 2022 DaeJin Jo, Sungwoong Kim, Daniel Wontae Nam, Taehwan Kwon, Seungeun Rho, Jongmin Kim, Donghoon Lee

In order to resolve these issues, in this paper, we propose a learnable hash-based episodic count, which we name LECO, that efficiently performs as a task-specific intrinsic reward in hard exploration problems.

Efficient Exploration reinforcement-learning

Selective Token Generation for Few-shot Natural Language Generation

1 code implementation COLING 2022 DaeJin Jo, Taehwan Kwon, Eun-Sol Kim, Sungwoong Kim

Natural language modeling with limited training data is a challenging problem, and many algorithms make use of large-scale pretrained language models (PLMs) for this due to its great generalization ability.

Data-to-Text Generation Language Modelling +3

Selective Token Generation for Few-shot Language Modeling

no code implementations29 Sep 2021 DaeJin Jo, Taehwan Kwon, Sungwoong Kim, Eun-Sol Kim

Therefore, in this work, we develop a novel additive learning algorithm based on reinforcement learning (RL) for few-shot natural language generation (NLG) tasks.

Data-to-Text Generation Language Modelling +3

Variational Intrinsic Control Revisited

no code implementations ICLR 2021 Taehwan Kwon

In this paper, we revisit variational intrinsic control (VIC), an unsupervised reinforcement learning method for finding the largest set of intrinsic options available to an agent.

reinforcement-learning Reinforcement Learning (RL) +1

Learning NP-Hard Multi-Agent Assignment Planning using GNN: Inference on a Random Graph and Provable Auction-Fitted Q-learning

no code implementations29 May 2019 Hyunwook Kang, Taehwan Kwon, Jinkyoo Park, James R. Morrison

In representing the MRRC problem as a sequential decision-making problem, we observe that each state can be represented as an extension of probabilistic graphical models (PGMs), which we refer to as random PGMs.

Combinatorial Optimization Decision Making +3

Cannot find the paper you are looking for? You can Submit a new open access paper.