Search Results for author: Jacob Beck

Found 14 papers, 5 papers with code

SplAgger: Split Aggregation for Meta-Reinforcement Learning

no code implementations5 Mar 2024 Jacob Beck, Matthew Jackson, Risto Vuorio, Zheng Xiong, Shimon Whiteson

However, it remains unclear whether task inference sequence models are beneficial even when task inference objectives are not.

Continuous Control Meta Reinforcement Learning +2

Distilling Morphology-Conditioned Hypernetworks for Efficient Universal Morphology Control

no code implementations9 Feb 2024 Zheng Xiong, Risto Vuorio, Jacob Beck, Matthieu Zimmer, Kun Shao, Shimon Whiteson

Learning a universal policy across different robot morphologies can significantly improve learning efficiency and enable zero-shot generalization to unseen morphologies.

Zero-shot Generalization

Annotation Sensitivity: Training Data Collection Methods Affect Model Performance

1 code implementation23 Nov 2023 Christoph Kern, Stephanie Eckman, Jacob Beck, Rob Chew, Bolei Ma, Frauke Kreuter

We introduce the term annotation sensitivity to refer to the impact of annotation data collection methods on the annotations themselves and on downstream model performance and predictions.

Recurrent Hypernetworks are Surprisingly Strong in Meta-RL

1 code implementation NeurIPS 2023 Jacob Beck, Risto Vuorio, Zheng Xiong, Shimon Whiteson

While many specialized meta-RL methods have been proposed, recent work suggests that end-to-end learning in conjunction with an off-the-shelf sequential model, such as a recurrent network, is a surprisingly strong baseline.

Few-Shot Learning Reinforcement Learning (RL)

Universal Morphology Control via Contextual Modulation

1 code implementation22 Feb 2023 Zheng Xiong, Jacob Beck, Shimon Whiteson

Learning a universal policy across different robot morphologies can significantly improve learning efficiency and generalization in continuous control.

Continuous Control

A Survey of Meta-Reinforcement Learning

no code implementations19 Jan 2023 Jacob Beck, Risto Vuorio, Evan Zheran Liu, Zheng Xiong, Luisa Zintgraf, Chelsea Finn, Shimon Whiteson

Meta-RL is most commonly studied in a problem setting where, given a distribution of tasks, the goal is to learn a policy that is capable of adapting to any new task from the task distribution with as little data as possible.

Meta Reinforcement Learning reinforcement-learning +1

Hypernetworks in Meta-Reinforcement Learning

1 code implementation20 Oct 2022 Jacob Beck, Matthew Thomas Jackson, Risto Vuorio, Shimon Whiteson

In this paper, we 1) show that hypernetwork initialization is also a critical factor in meta-RL, and that naive initializations yield poor performance; 2) propose a novel hypernetwork initialization scheme that matches or exceeds the performance of a state-of-the-art approach proposed for supervised settings, as well as being simpler and more general; and 3) use this method to show that hypernetworks can improve performance in meta-RL by evaluating on multiple simulated robotics benchmarks.

Meta Reinforcement Learning reinforcement-learning +1

An Investigation of the Bias-Variance Tradeoff in Meta-Gradients

1 code implementation22 Sep 2022 Risto Vuorio, Jacob Beck, Shimon Whiteson, Jakob Foerster, Gregory Farquhar

Meta-gradients provide a general approach for optimizing the meta-parameters of reinforcement learning (RL) algorithms.

Meta-Learning Reinforcement Learning (RL)

Trust Region Bounds for Decentralized PPO Under Non-stationarity

no code implementations31 Jan 2022 Mingfei Sun, Sam Devlin, Jacob Beck, Katja Hofmann, Shimon Whiteson

We present trust region bounds for optimizing decentralized policies in cooperative Multi-Agent Reinforcement Learning (MARL), which holds even when the transition dynamics are non-stationary.

Multi-agent Reinforcement Learning

On the Practical Consistency of Meta-Reinforcement Learning Algorithms

no code implementations1 Dec 2021 Zheng Xiong, Luisa Zintgraf, Jacob Beck, Risto Vuorio, Shimon Whiteson

We further find that theoretically inconsistent algorithms can be made consistent by continuing to update all agent components on the OOD tasks, and adapt as well or better than originally consistent ones.

Meta-Learning Meta Reinforcement Learning +3

AMRL: Aggregated Memory For Reinforcement Learning

no code implementations ICLR 2020 Jacob Beck, Kamil Ciosek, Sam Devlin, Sebastian Tschiatschek, Cheng Zhang, Katja Hofmann

In many partially observable scenarios, Reinforcement Learning (RL) agents must rely on long-term memory in order to learn an optimal policy.

reinforcement-learning Reinforcement Learning (RL)

ReNeg and Backseat Driver: Learning from Demonstration with Continuous Human Feedback

no code implementations ICLR 2019 Jacob Beck, Zoe Papakipos, Michael Littman

Our framework learns continuous control from sub-optimal demonstration and evaluative feedback collected before training.

Continuous Control

Neural Mesh: Introducing a Notion of Space and Conservation of Energy to Neural Networks

no code implementations29 Jul 2018 Jacob Beck, Zoe Papakipos

Like in the brain, we only allow neurons to fire in a time step if they contain enough energy, or excitement.

Cannot find the paper you are looking for? You can Submit a new open access paper.