Search Results for author: Lucas Lehnert

Found 9 papers, 1 papers with code

Beyond A*: Better Planning with Transformers via Search Dynamics Bootstrapping

1 code implementation21 Feb 2024 Lucas Lehnert, Sainbayar Sukhbaatar, DiJia Su, Qinqing Zheng, Paul McVay, Michael Rabbat, Yuandong Tian

We fine tune this model to obtain a Searchformer, a Transformer model that optimally solves previously unseen Sokoban puzzles 93. 7% of the time, while using up to 26. 8% fewer search steps than the $A^*$ implementation that was used for training initially.

Decision Making

IQL-TD-MPC: Implicit Q-Learning for Hierarchical Model Predictive Control

no code implementations1 Jun 2023 Rohan Chitnis, Yingchen Xu, Bobak Hashemi, Lucas Lehnert, Urun Dogan, Zheqing Zhu, Olivier Delalleau

Model-based reinforcement learning (RL) has shown great promise due to its sample efficiency, but still struggles with long-horizon sparse-reward tasks, especially in offline settings where the agent learns from a fixed dataset.

D4RL Model-based Reinforcement Learning +4

Reward-Predictive Clustering

no code implementations7 Nov 2022 Lucas Lehnert, Michael J. Frank, Michael L. Littman

Recent advances in reinforcement-learning research have demonstrated impressive results in building algorithms that can out-perform humans in complex tasks.

Clustering reinforcement-learning +1

Mitigating Planner Overfitting in Model-Based Reinforcement Learning

no code implementations3 Dec 2018 Dilip Arumugam, David Abel, Kavosh Asadi, Nakul Gopalan, Christopher Grimm, Jun Ki Lee, Lucas Lehnert, Michael L. Littman

An agent with an inaccurate model of its environment faces a difficult choice: it can ignore the errors in its model and act in the real world in whatever way it determines is optimal with respect to its model.

Model-based Reinforcement Learning Position +2

Transfer with Model Features in Reinforcement Learning

no code implementations4 Jul 2018 Lucas Lehnert, Michael L. Littman

Further, we present a Successor Feature model which shows that learning Successor Features is equivalent to learning a Model-Reduction.

reinforcement-learning Reinforcement Learning (RL)

State Abstractions for Lifelong Reinforcement Learning

no code implementations ICML 2018 David Abel, Dilip Arumugam, Lucas Lehnert, Michael Littman

We introduce two new classes of abstractions: (1) transitive state abstractions, whose optimal form can be computed efficiently, and (2) PAC state abstractions, which are guaranteed to hold with respect to a distribution of tasks.

reinforcement-learning Reinforcement Learning (RL)

Advantages and Limitations of using Successor Features for Transfer in Reinforcement Learning

no code implementations31 Jul 2017 Lucas Lehnert, Stefanie Tellex, Michael L. Littman

One question central to Reinforcement Learning is how to learn a feature representation that supports algorithm scaling and re-use of learned information from different tasks.

reinforcement-learning Reinforcement Learning (RL)

Policy Gradient Methods for Off-policy Control

no code implementations13 Dec 2015 Lucas Lehnert, Doina Precup

Off-policy learning refers to the problem of learning the value function of a way of behaving, or policy, while following a different policy.

Policy Gradient Methods

Cannot find the paper you are looking for? You can Submit a new open access paper.