no code implementations • 4 Sep 2023 • Joost Broekens, Bernhard Hilpert, Suzan Verberne, Kim Baraka, Patrick Gebhard, Aske Plaat
Large language models, in particular generative pre-trained transformers (GPTs), show impressive results on a wide variety of language-related tasks.
no code implementations • 27 Jan 2023 • Bernd Dudzik, Joost Broekens
In this article, we highlight the important influence of the temporal distance between a stimulus event and the moment when its experience is reported on the provided information's validity.
no code implementations • 27 Aug 2020 • Bernd Dudzik, Joost Broekens, Mark Neerincx, Hayley Hung
A key challenge in the accurate prediction of viewers' emotional responses to video stimuli in real-world applications is accounting for person- and situation-specific variation.
no code implementations • 30 Jun 2020 • Thomas M. Moerland, Joost Broekens, Aske Plaat, Catholijn M. Jonker
Two key approaches to this problem are reinforcement learning (RL) and planning.
no code implementations • 26 Jun 2020 • Thomas M. Moerland, Joost Broekens, Aske Plaat, Catholijn M. Jonker
Therefore, this paper presents a unifying algorithmic framework for reinforcement learning and planning (FRAP), which identifies underlying dimensions on which MDP planning and learning algorithms have to decide.
1 code implementation • 19 May 2020 • Thomas M. Moerland, Joost Broekens, Aske Plaat, Catholijn M. Jonker
Monte Carlo Tree Search (MCTS) efficiently balances exploration and exploitation in tree search based on count-derived uncertainty.
1 code implementation • 15 May 2020 • Thomas M. Moerland, Anna Deichler, Simone Baldi, Joost Broekens, Catholijn M. Jonker
Planning and reinforcement learning are two key approaches to sequential decision making.
no code implementations • 24 Jul 2018 • Joost Broekens
Emotions are intimately tied to motivation and the adaptation of behavior, and many animal species show evidence of emotions in their behavior.
1 code implementation • 11 Jun 2018 • Thomas M. Moerland, Joost Broekens, Catholijn M. Jonker
This paper studies the potential of the return distribution for exploration in deterministic reinforcement learning (RL) environments.
2 code implementations • 24 May 2018 • Thomas M. Moerland, Joost Broekens, Aske Plaat, Catholijn M. Jonker
A core novelty of Alpha Zero is the interleaving of tree search and deep learning, which has proven very successful in board games like Chess, Shogi and Go.
2 code implementations • 23 May 2018 • Thomas M. Moerland, Joost Broekens, Aske Plaat, Catholijn M. Jonker
Asymmetric termination of search trees introduces a type of uncertainty for which the standard upper confidence bound (UCB) formula does not account.
no code implementations • 29 Nov 2017 • Thomas M. Moerland, Joost Broekens, Catholijn M. Jonker
This paper studies directed exploration for reinforcement learning agents by tracking uncertainty about the value of each available action.
no code implementations • 15 May 2017 • Thomas M. Moerland, Joost Broekens, Catholijn M. Jonker
This article provides the first survey of computational models of emotion in reinforcement learning (RL) agents.
1 code implementation • 1 May 2017 • Thomas M. Moerland, Joost Broekens, Catholijn M. Jonker
In this paper we study how to learn stochastic, multimodal transition dynamics in reinforcement learning (RL) tasks.
Model-based Reinforcement Learning reinforcement-learning +2
no code implementations • 8 Apr 2014 • Joost Broekens, Tim Baarslag
We found that overly optimistic risk perception (outcome optimism) results in risk taking and in persistent gambling behaviour in addition to high intensity of sensations.