Search Results for author: Julien Pérolat

Found 9 papers, 2 papers with code

Learning in Mean Field Games: A Survey

no code implementations25 May 2022 Mathieu Laurière, Sarah Perrin, Julien Pérolat, Sertan Girgin, Paul Muller, Romuald Élie, Matthieu Geist, Olivier Pietquin

Non-cooperative and cooperative games with a very large number of players have many applications but remain generally intractable when the number of players increases.

Reinforcement Learning (RL)

Scalable Deep Reinforcement Learning Algorithms for Mean Field Games

no code implementations22 Mar 2022 Mathieu Laurière, Sarah Perrin, Sertan Girgin, Paul Muller, Ayush Jain, Theophile Cabannes, Georgios Piliouras, Julien Pérolat, Romuald Élie, Olivier Pietquin, Matthieu Geist

One limiting factor to further scale up using RL is that existing algorithms to solve MFGs require the mixing of approximated quantities such as strategies or $q$-values.

reinforcement-learning Reinforcement Learning (RL)

Generalization in Mean Field Games by Learning Master Policies

no code implementations20 Sep 2021 Sarah Perrin, Mathieu Laurière, Julien Pérolat, Romuald Élie, Matthieu Geist, Olivier Pietquin

Mean Field Games (MFGs) can potentially scale multi-agent systems to extremely large populations of agents.

Mean Field Games Flock! The Reinforcement Learning Way

no code implementations17 May 2021 Sarah Perrin, Mathieu Laurière, Julien Pérolat, Matthieu Geist, Romuald Élie, Olivier Pietquin

We present a method enabling a large number of agents to learn how to flock, which is a natural behavior observed in large populations of animals.

reinforcement-learning Reinforcement Learning (RL)

On the Convergence of Model Free Learning in Mean Field Games

no code implementations4 Jul 2019 Romuald Elie, Julien Pérolat, Mathieu Laurière, Matthieu Geist, Olivier Pietquin

In order to design scalable algorithms for systems with a large population of interacting agents (e. g. swarms), this paper focuses on Mean Field MAS, where the number of agents is asymptotically infinite.

Computing Approximate Equilibria in Sequential Adversarial Games by Exploitability Descent

no code implementations13 Mar 2019 Edward Lockhart, Marc Lanctot, Julien Pérolat, Jean-Baptiste Lespiau, Dustin Morrill, Finbarr Timbers, Karl Tuyls

In this paper, we present exploitability descent, a new algorithm to compute approximate equilibria in two-player zero-sum extensive-form games with imperfect information, by direct policy optimization against worst-case opponents.

counterfactual

Cannot find the paper you are looking for? You can Submit a new open access paper.