Search Results for author: Manan Tomar

Found 12 papers, 3 papers with code

Robotic Offline RL from Internet Videos via Value-Function Pre-Training

no code implementations22 Sep 2023 Chethan Bhateja, Derek Guo, Dibya Ghosh, Anikait Singh, Manan Tomar, Quan Vuong, Yevgen Chebotar, Sergey Levine, Aviral Kumar

Our system, called V-PTR, combines the benefits of pre-training on video data with robotic offline RL approaches that train on diverse robot data, resulting in value functions and policies for manipulation tasks that perform better, act robustly, and generalize broadly.

Offline RL Reinforcement Learning (RL)

Ignorance is Bliss: Robust Control via Information Gating

no code implementations NeurIPS 2023 Manan Tomar, Riashat Islam, Matthew E. Taylor, Sergey Levine, Philip Bachman

We propose \textit{information gating} as a way to learn parsimonious representations that identify the minimal information required for a task.

Inductive Bias Q-Learning

Agent-Controller Representations: Principled Offline RL with Rich Exogenous Information

1 code implementation31 Oct 2022 Riashat Islam, Manan Tomar, Alex Lamb, Yonathan Efroni, Hongyu Zang, Aniket Didolkar, Dipendra Misra, Xin Li, Harm van Seijen, Remi Tachet des Combes, John Langford

We find that contemporary representation learning techniques can fail on datasets where the noise is a complex and time dependent process, which is prevalent in practical applications.

Offline RL Reinforcement Learning (RL) +1

Learning Representations for Pixel-based Control: What Matters and Why?

no code implementations15 Nov 2021 Manan Tomar, Utkarsh A. Mishra, Amy Zhang, Matthew E. Taylor

A wide range of methods have been proposed to enable efficient learning, leading to sample complexities similar to those in the full state setting.

Contrastive Learning Representation Learning

Learning Minimal Representations with Model Invariance

no code implementations29 Sep 2021 Manan Tomar, Amy Zhang, Matthew E. Taylor

The common representation acts as a implicit invariance objective to avoid the different spurious correlations captured by individual predictors.

Self-Supervised Learning

Mirror Descent Policy Optimization

1 code implementation ICLR 2022 Manan Tomar, Lior Shani, Yonathan Efroni, Mohammad Ghavamzadeh

Overall, MDPO is derived from the MD principles, offers a unified approach to viewing a number of popular RL algorithms, and performs better than or on-par with TRPO, PPO, and SAC in a number of continuous control tasks.

Continuous Control Reinforcement Learning (RL)

Multi-step Greedy Reinforcement Learning Algorithms

no code implementations ICML 2020 Manan Tomar, Yonathan Efroni, Mohammad Ghavamzadeh

We derive model-free RL algorithms based on $\kappa$-PI and $\kappa$-VI in which the surrogate problem can be solved by any discrete or continuous action RL method, such as DQN and TRPO.

Continuous Control Game of Go +3

Multi-step Greedy Policies in Model-Free Deep Reinforcement Learning

no code implementations25 Sep 2019 Yonathan Efroni, Manan Tomar, Mohammad Ghavamzadeh

In this work, we explore the benefits of multi-step greedy policies in model-free RL when employed in the framework of multi-step Dynamic Programming (DP): multi-step Policy and Value Iteration.

Continuous Control Game of Go +3

MaMiC: Macro and Micro Curriculum for Robotic Reinforcement Learning

no code implementations17 May 2019 Manan Tomar, Akhil Sathuluri, Balaraman Ravindran

Shaping in humans and animals has been shown to be a powerful tool for learning complex tasks as compared to learning in a randomized fashion.

reinforcement-learning Reinforcement Learning (RL) +1

Successor Options: An Option Discovery Framework for Reinforcement Learning

1 code implementation14 May 2019 Rahul Ramesh, Manan Tomar, Balaraman Ravindran

This work adopts a complementary approach, where we attempt to discover options that navigate to landmark states.

Navigate reinforcement-learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.