Search Results for author: Daniel Palenicek

Found 6 papers, 4 papers with code

Iterated $Q$-Network: Beyond the One-Step Bellman Operator

no code implementations4 Mar 2024 Théo Vincent, Daniel Palenicek, Boris Belousov, Jan Peters, Carlo D'Eramo

Value-based Reinforcement Learning (RL) methods rely on the application of the Bellman operator, which needs to be approximated from samples.

Atari Games Continuous Control +1

Diminishing Return of Value Expansion Methods in Model-Based Reinforcement Learning

1 code implementation7 Mar 2023 Daniel Palenicek, Michael Lutter, Joao Carvalho, Jan Peters

Therefore, we conclude that the limitation of model-based value expansion methods is not the model accuracy of the learned models.

Continuous Control Model-based Reinforcement Learning +2

Revisiting Model-based Value Expansion

no code implementations28 Mar 2022 Daniel Palenicek, Michael Lutter, Jan Peters

Model-based value expansion methods promise to improve the quality of value function targets and, thereby, the effectiveness of value function learning.

Model-based Reinforcement Learning

SAMBA: Safe Model-Based & Active Reinforcement Learning

1 code implementation12 Jun 2020 Alexander I. Cowen-Rivers, Daniel Palenicek, Vincent Moens, Mohammed Abdullah, Aivar Sootla, Jun Wang, Haitham Ammar

In this paper, we propose SAMBA, a novel framework for safe reinforcement learning that combines aspects from probabilistic modelling, information theory, and statistics.

Reinforcement Learning (RL) Safe Reinforcement Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.