Search Results for author: Han-Dong Lim

Found 8 papers, 0 papers with code

Finite-Time Error Analysis of Online Model-Based Q-Learning with a Relaxed Sampling Model

no code implementations19 Feb 2024 Han-Dong Lim, HyeAnn Lee, Donghwan Lee

Reinforcement learning has witnessed significant advancements, particularly with the emergence of model-based approaches.

Q-Learning reinforcement-learning

A primal-dual perspective for distributed TD-learning

no code implementations1 Oct 2023 Han-Dong Lim, Donghwan Lee

The goal of this paper is to investigate distributed temporal difference (TD) learning for a networked multi-agent Markov decision process.

Distributed Optimization

Continuous-Time Distributed Dynamic Programming for Networked Multi-Agent Markov Decision Processes

no code implementations31 Jul 2023 Donghwan Lee, Han-Dong Lim, Do Wan Kim

The main goal of this paper is to investigate continuous-time distributed dynamic programming (DP) algorithms for networked multi-agent Markov decision problems (MAMDPs).

Distributed Optimization

Temporal Difference Learning with Experience Replay

no code implementations16 Jun 2023 Han-Dong Lim, Donghwan Lee

Temporal-difference (TD) learning is widely regarded as one of the most popular algorithms in reinforcement learning (RL).

Reinforcement Learning (RL)

Backstepping Temporal Difference Learning

no code implementations20 Feb 2023 Han-Dong Lim, Donghwan Lee

Off-policy learning ability is an important feature of reinforcement learning (RL) for practical applications.

Reinforcement Learning (RL)

Finite-Time Analysis of Asynchronous Q-learning under Diminishing Step-Size from Control-Theoretic View

no code implementations25 Jul 2022 Han-Dong Lim, Donghwan Lee

Q-learning has long been one of the most popular reinforcement learning algorithms, and theoretical analysis of Q-learning has been an active research topic for decades.

Q-Learning

Regularized Q-learning

no code implementations11 Feb 2022 Han-Dong Lim, Do Wan Kim, Donghwan Lee

This paper develops a new Q-learning algorithm that converges when linear function approximation is used.

Q-Learning reinforcement-learning +1

New Versions of Gradient Temporal Difference Learning

no code implementations9 Sep 2021 Donghwan Lee, Han-Dong Lim, Jihoon Park, Okyong Choi

Sutton, Szepesv\'{a}ri and Maei introduced the first gradient temporal-difference (GTD) learning algorithms compatible with both linear function approximation and off-policy training.

Cannot find the paper you are looking for? You can Submit a new open access paper.