no code implementations • 18 Aug 2022 • Manfred Eppe, Christian Gumbsch, Matthias Kerzel, Phuong D. H. Nguyen, Martin V. Butz, Stefan Wermter
According to cognitive psychology and related disciplines, the development of complex problem-solving behaviour in biological agents depends on hierarchical cognitive mechanisms.
Hierarchical Reinforcement Learning reinforcement-learning +1
no code implementations • 18 Dec 2020 • Manfred Eppe, Christian Gumbsch, Matthias Kerzel, Phuong D. H. Nguyen, Martin V. Butz, Stefan Wermter
We then relate these insights with contemporary hierarchical reinforcement learning methods, and identify the key machine intelligence approaches that realise these mechanisms.
Hierarchical Reinforcement Learning reinforcement-learning +1
no code implementations • 25 Nov 2020 • Phuong D. H. Nguyen, Yasmin Kim Georgie, Ezgi Kayhan, Manfred Eppe, Verena Vanessa Hafner, Stefan Wermter
Safe human-robot interactions require robots to be able to learn how to behave appropriately in \sout{humans' world} \rev{spaces populated by people} and thus to cope with the challenges posed by our dynamic and unstructured environment, rather than being provided a rigid set of rules for operations.
no code implementations • 13 Nov 2020 • Phuong D. H. Nguyen, Manfred Eppe, Stefan Wermter
Cognitive science suggests that the self-representation is critical for learning and problem-solving.
no code implementations • 11 Nov 2020 • Thilo Fryen, Manfred Eppe, Phuong D. H. Nguyen, Timo Gerkmann, Stefan Wermter
Reinforcement learning is a promising method to accomplish robotic control tasks.
1 code implementation • 7 May 2020 • Frank Röder, Manfred Eppe, Phuong D. H. Nguyen, Stefan Wermter
Hierarchical abstraction and curiosity-driven exploration are two common paradigms in current reinforcement learning approaches to break down difficult problems into a sequence of simpler ones and to overcome reward sparsity.
no code implementations • 23 May 2019 • Manfred Eppe, Phuong D. H. Nguyen, Stefan Wermter
In this article, we build on these novel methods to facilitate the integration of action planning with reinforcement learning by exploiting the reward-sparsity as a bridge between the high-level and low-level state- and control spaces.