1 code implementation • 30 Jan 2023 • Alexandra Cimpean, Timothy Verstraeten, Lander Willem, Niel Hens, Ann Nowé, Pieter Libin
$m$-top exploration allows the algorithm to learn $m$ policies for which it expects the highest utility, enabling experts to inspect this small set of alternative strategies, along with their quantified uncertainty.
no code implementations • 11 Apr 2022 • Mathieu Reymond, Conor F. Hayes, Lander Willem, Roxana Rădulescu, Steven Abrams, Diederik M. Roijers, Enda Howley, Patrick Mannion, Niel Hens, Ann Nowé, Pieter Libin
As decision making in the context of epidemic mitigation is hard, reinforcement learning provides a methodology to automatically learn prevention strategies in combination with complex epidemic models.
1 code implementation • 30 Mar 2020 • Pieter Libin, Arno Moonens, Timothy Verstraeten, Fabian Perez-Sanjines, Niel Hens, Philippe Lemey, Ann Nowé
For this reason, we investigate a deep reinforcement learning approach to automatically learn prevention strategies in the context of pandemic influenza.
no code implementations • 30 Sep 2019 • Felipe Gomez Marulanda, Pieter Libin, Timothy Verstraeten, Ann Nowé
In general, our approach outperforms PointNet on every family of 3D geometries on which the models were tested.
no code implementations • 16 Nov 2017 • Pieter Libin, Timothy Verstraeten, Diederik M. Roijers, Jelena Grujic, Kristof Theys, Philippe Lemey, Ann Nowé
We evaluate these algorithms in a realistic experimental setting and demonstrate that it is possible to identify the optimal strategy using only a limited number of model evaluations, i. e., 2-to-3 times faster compared to the uniform sampling method, the predominant technique used for epidemiological decision making in the literature.