1 code implementation • 27 Jan 2022 • Rituraj Kaushik, Karol Arndt, Ville Kyrki
In this work, we introduce a novel learning algorithm called SafeAPT that leverages a diverse repertoire of policies evolved in the simulation and transfers the most promising safe policy to the real robot through episodic interaction.
1 code implementation • 10 Mar 2020 • Rituraj Kaushik, Timothée Anne, Jean-Baptiste Mouret
Meta-learning algorithms can accelerate the model-based reinforcement learning (MBRL) algorithms by finding an initial set of parameters for the dynamical model such that the model can be trained to match the actual dynamics of the system with only a few data-points.
1 code implementation • 16 Jul 2019 • Rituraj Kaushik, Pierre Desreumaux, Jean-Baptiste Mouret
Repertoire-based learning is a data-efficient adaptation approach based on a two-step process in which (1) a large and diverse set of policies is learned in simulation, and (2) a planning or learning algorithm chooses the most appropriate policies according to the current situation (e. g., a damaged robot, a new object, etc.).
1 code implementation • 25 Jun 2018 • Rituraj Kaushik, Konstantinos Chatzilygeroudis, Jean-Baptiste Mouret
The most data-efficient algorithms for reinforcement learning in robotics are model-based policy search algorithms, which alternate between learning a dynamical model of the robot and optimizing a policy to maximize the expected return given the model and its uncertainties.
1 code implementation • 21 Mar 2017 • Konstantinos Chatzilygeroudis, Roberto Rama, Rituraj Kaushik, Dorian Goepp, Vassilis Vassiliades, Jean-Baptiste Mouret
The most data-efficient algorithms for reinforcement learning (RL) in robotics are based on uncertain dynamical models: after each episode, they first learn a dynamical model of the robot, then they use an optimization algorithm to find a policy that maximizes the expected return given the model and its uncertainties.