no code implementations • 4 Mar 2024 • Théo Vincent, Daniel Palenicek, Boris Belousov, Jan Peters, Carlo D'Eramo
Value-based Reinforcement Learning (RL) methods rely on the application of the Bellman operator, which needs to be approximated from samples.
1 code implementation • 7 Mar 2023 • Daniel Palenicek, Michael Lutter, Joao Carvalho, Jan Peters
Therefore, we conclude that the limitation of model-based value expansion methods is not the model accuracy of the learned models.
no code implementations • 28 Mar 2022 • Daniel Palenicek, Michael Lutter, Jan Peters
Model-based value expansion methods promise to improve the quality of value function targets and, thereby, the effectiveness of value function learning.
3 code implementations • 19 Oct 2020 • Ming Zhou, Jun Luo, Julian Villella, Yaodong Yang, David Rusu, Jiayu Miao, Weinan Zhang, Montgomery Alban, Iman Fadakar, Zheng Chen, Aurora Chongxi Huang, Ying Wen, Kimia Hassanzadeh, Daniel Graves, Dong Chen, Zhengbang Zhu, Nhat Nguyen, Mohamed Elsayed, Kun Shao, Sanjeevan Ahilan, Baokuan Zhang, Jiannan Wu, Zhengang Fu, Kasra Rezaee, Peyman Yadmellat, Mohsen Rohani, Nicolas Perez Nieves, Yihan Ni, Seyedershad Banijamali, Alexander Cowen Rivers, Zheng Tian, Daniel Palenicek, Haitham Bou Ammar, Hongbo Zhang, Wulong Liu, Jianye Hao, Jun Wang
We open-source the SMARTS platform and the associated benchmark tasks and evaluation metrics to encourage and empower research on multi-agent learning for autonomous driving.
1 code implementation • 12 Jun 2020 • Alexander I. Cowen-Rivers, Daniel Palenicek, Vincent Moens, Mohammed Abdullah, Aivar Sootla, Jun Wang, Haitham Ammar
In this paper, we propose SAMBA, a novel framework for safe reinforcement learning that combines aspects from probabilistic modelling, information theory, and statistics.
3 code implementations • 14 Feb 2019 • Aditya Bhatt, Daniel Palenicek, Boris Belousov, Max Argus, Artemij Amiranashvili, Thomas Brox, Jan Peters
Sample efficiency is a crucial problem in deep reinforcement learning.