Search Results for author: Jakob N. Foerster

Found 19 papers, 11 papers with code

Game-Theoretical Perspectives on Active Equilibria: A Preferred Solution Concept over Nash Equilibria

no code implementations28 Oct 2022 Dong-Ki Kim, Matthew Riemer, Miao Liu, Jakob N. Foerster, Gerald Tesauro, Jonathan P. How

By directly comparing active equilibria to Nash equilibria in these examples, we find that active equilibria find more effective solutions than Nash equilibria, concluding that an active equilibrium is the desired solution for multiagent learning settings.

Learning to Optimize Quasi-Newton Methods

no code implementations11 Oct 2022 Isaac Liao, Rumen R. Dangovski, Jakob N. Foerster, Marin Soljačić

This paper introduces a novel machine learning optimizer called LODO, which tries to online meta-learn the best preconditioner during optimization.

K-level Reasoning for Zero-Shot Coordination in Hanabi

no code implementations NeurIPS 2021 Brandon Cui, Hengyuan Hu, Luis Pineda, Jakob N. Foerster

The standard problem setting in cooperative multi-agent settings is self-play (SP), where the goal is to train a team of agents that works well together.

Self-Explaining Deviations for Coordination

no code implementations13 Jul 2022 Hengyuan Hu, Samuel Sokota, David Wu, Anton Bakhtin, Andrei Lupu, Brandon Cui, Jakob N. Foerster

Fully cooperative, partially observable multi-agent problems are ubiquitous in the real world.

Influencing Long-Term Behavior in Multiagent Reinforcement Learning

1 code implementation7 Mar 2022 Dong-Ki Kim, Matthew Riemer, Miao Liu, Jakob N. Foerster, Michael Everett, Chuangchuang Sun, Gerald Tesauro, Jonathan P. How

An effective approach that has recently emerged for addressing this non-stationarity is for each agent to anticipate the learning of other agents and influence the evolution of future policies towards desirable behavior for its own benefit.

reinforcement-learning Reinforcement Learning (RL)

Simplified Action Decoder for Deep Multi-Agent Reinforcement Learning

4 code implementations ICLR 2020 Hengyuan Hu, Jakob N. Foerster

Learning to be informative when observed by others is an interesting challenge for Reinforcement Learning (RL): Fundamentally, RL requires agents to explore in order to discover good policies.

Multi-agent Reinforcement Learning reinforcement-learning +1

Robust Visual Domain Randomization for Reinforcement Learning

2 code implementations23 Oct 2019 Reda Bahi Slaoui, William R. Clements, Jakob N. Foerster, Sébastien Toth

Producing agents that can generalize to a wide range of visually different environments is a significant challenge in reinforcement learning.

reinforcement-learning Reinforcement Learning (RL)

Robust Domain Randomization for Reinforcement Learning

no code implementations25 Sep 2019 Reda Bahi Slaoui, William R. Clements, Jakob N. Foerster, Sébastien Toth

In this work, we formalize the domain randomization problem, and show that minimizing the policy's Lipschitz constant with respect to the randomization parameters leads to low variance in the learned policies.

reinforcement-learning Reinforcement Learning (RL)

Exploratory Combinatorial Optimization with Reinforcement Learning

2 code implementations9 Sep 2019 Thomas D. Barrett, William R. Clements, Jakob N. Foerster, A. I. Lvovsky

Our approach of exploratory combinatorial optimization (ECO-DQN) is, in principle, applicable to any combinatorial problem that can be defined on a graph.

Combinatorial Optimization reinforcement-learning +1

Bayesian Action Decoder for Deep Multi-Agent Reinforcement Learning

1 code implementation4 Nov 2018 Jakob N. Foerster, Francis Song, Edward Hughes, Neil Burch, Iain Dunning, Shimon Whiteson, Matthew Botvinick, Michael Bowling

We present the Bayesian action decoder (BAD), a new multi-agent learning method that uses an approximate Bayesian update to obtain a public belief that conditions on the actions taken by all agents in the environment.

Multi-agent Reinforcement Learning Policy Gradient Methods +2

Learning with Opponent-Learning Awareness

6 code implementations13 Sep 2017 Jakob N. Foerster, Richard Y. Chen, Maruan Al-Shedivat, Shimon Whiteson, Pieter Abbeel, Igor Mordatch

We also show that the LOLA update rule can be efficiently calculated using an extension of the policy gradient estimator, making the method suitable for model-free RL.

Multi-agent Reinforcement Learning

Learning to Communicate to Solve Riddles with Deep Distributed Recurrent Q-Networks

no code implementations8 Feb 2016 Jakob N. Foerster, Yannis M. Assael, Nando de Freitas, Shimon Whiteson

We propose deep distributed recurrent Q-networks (DDRQN), which enable teams of agents to learn to solve communication-based coordination tasks.

reinforcement-learning Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.