Search Results for author: Mariano Phielipp

Found 21 papers, 7 papers with code

MOTO: Offline Pre-training to Online Fine-tuning for Model-based Robot Learning

no code implementations6 Jan 2024 Rafael Rafailov, Kyle Hatch, Victor Kolev, John D. Martin, Mariano Phielipp, Chelsea Finn

We study the problem of offline pre-training and online fine-tuning for reinforcement learning from high-dimensional observations in the context of realistic robot tasks.

Offline RL Robot Manipulation

Group SELFIES: A Robust Fragment-Based Molecular String Representation

1 code implementation23 Nov 2022 Austin Cheng, Andy Cai, Santiago Miret, Gustavo Malkomes, Mariano Phielipp, Alán Aspuru-Guzik

We introduce Group SELFIES, a molecular string representation that leverages group tokens to represent functional groups or entire substructures while maintaining chemical robustness guarantees.

AnyMorph: Learning Transferable Polices By Inferring Agent Morphology

no code implementations17 Jun 2022 Brandon Trabucco, Mariano Phielipp, Glen Berseth

Ours is the first reinforcement learning algorithm that can train a policy to generalize to new agent morphologies without requiring a description of the agent's morphology in advance.

reinforcement-learning Reinforcement Learning (RL) +1

Offline Policy Comparison with Confidence: Benchmarks and Baselines

2 code implementations22 May 2022 Anurag Koul, Mariano Phielipp, Alan Fern

Decision makers often wish to use offline historical data to compare sequential-action policies at various world states.

Pretraining Graph Neural Networks for few-shot Analog Circuit Modeling and Design

1 code implementation29 Mar 2022 Kourosh Hakhamaneshi, Marcel Nassar, Mariano Phielipp, Pieter Abbeel, Vladimir Stojanović

We show that pretraining GNNs on prediction of output node voltages can encourage learning representations that can be adapted to new unseen topologies or prediction of new circuit level properties with up to 10x more sample efficiency compared to a randomly initialized model.

DNS: Determinantal Point Process Based Neural Network Sampler for Ensemble Reinforcement Learning

1 code implementation31 Jan 2022 Hassam Sheikh, Kizza Frisbee, Mariano Phielipp

Application of ensemble of neural networks is becoming an imminent tool for advancing the state-of-the-art in deep reinforcement learning algorithms.

Continuous Control Ensemble Learning +2

Maximizing Ensemble Diversity in Deep Reinforcement Learning

no code implementations ICLR 2022 Hassam Sheikh, Mariano Phielipp, Ladislau Boloni

In this paper, we describe Maximize Ensemble Diversity in Reinforcement Learning (MED-RL), a set of regularization methods inspired from the economics and consensus optimization to improve diversity in the ensemble-based deep reinforcement learning methods by encouraging inequality between the networks during training.

Atari Games Decision Making +2

Neuroevolution-Enhanced Multi-Objective Optimization for Mixed-Precision Quantization

no code implementations14 Jun 2021 Santiago Miret, Vui Seng Chua, Mattias Marder, Mariano Phielipp, Nilesh Jain, Somdeb Majumdar

In this work, we present a flexible and scalable framework for automated mixed-precision quantization that concurrently optimizes task performance, memory compression, and compute savings through multi-objective evolutionary computing.

Quantization

ReaPER: Improving Sample Efficiency in Model-Based Latent Imagination

no code implementations1 Jan 2021 Martin A Bertran, Guillermo Sapiro, Mariano Phielipp

Deep Reinforcement Learning (DRL) can distill behavioural policies from sensory input that solve complex tasks, however, the policies tend to be task-specific and sample inefficient, requiring a large number of interactions with the environment that may be costly or impractical for many real world applications.

Instance-based Generalization in Reinforcement Learning

no code implementations NeurIPS 2020 Martin Bertran, Natalia Martinez, Mariano Phielipp, Guillermo Sapiro

Agents trained via deep reinforcement learning (RL) routinely fail to generalize to unseen environments, even when these share the same underlying dynamics as the training levels.

Generalization Bounds reinforcement-learning +1

Instance based Generalization in Reinforcement Learning

1 code implementation2 Nov 2020 Martin Bertran, Natalia Martinez, Mariano Phielipp, Guillermo Sapiro

Agents trained via deep reinforcement learning (RL) routinely fail to generalize to unseen environments, even when these share the same underlying dynamics as the training levels.

Generalization Bounds reinforcement-learning +1

Motion2Vec: Semi-Supervised Representation Learning from Surgical Videos

no code implementations31 May 2020 Ajay Kumar Tanwani, Pierre Sermanet, Andy Yan, Raghav Anand, Mariano Phielipp, Ken Goldberg

We demonstrate the use of this representation to imitate surgical suturing motions from publicly available videos of the JIGSAWS dataset.

Action Segmentation Metric Learning +1

Clone Swarms: Learning to Predict and Control Multi-Robot Systems by Imitation

no code implementations5 Dec 2019 Siyu Zhou, Mariano Phielipp, Jorge A. Sefair, Sara I. Walker, Heni Ben Amor

In this paper, we propose SwarmNet -- a neural network architecture that can learn to predict and imitate the behavior of an observed swarm of agents in a centralized manner.

Imitation Learning of Robot Policies by Combining Language, Vision and Demonstration

no code implementations26 Nov 2019 Simon Stepputtis, Joseph Campbell, Mariano Phielipp, Chitta Baral, Heni Ben Amor

In this work we propose a novel end-to-end imitation learning approach which combines natural language, vision, and motion information to produce an abstract representation of a task, which in turn is used to synthesize specific motion controllers at run-time.

Imitation Learning

Imitation Learning of Robot Policies using Language, Vision and Motion

no code implementations25 Sep 2019 Simon Stepputtis, Joseph Campbell, Mariano Phielipp, Chitta Baral, Heni Ben Amor

In this work we propose a novel end-to-end imitation learning approach which combines natural language, vision, and motion information to produce an abstract representation of a task, which in turn can be used to synthesize specific motion controllers at run-time.

Imitation Learning

On Training Flexible Robots using Deep Reinforcement Learning

no code implementations29 Jun 2019 Zach Dwiel, Madhavun Candadai, Mariano Phielipp

The use of robotics in controlled environments has flourished over the last several decades and training robots to perform tasks using control strategies developed from dynamical models of their hardware have proven very effective.

Industrial Robots reinforcement-learning +1

Goal-conditioned Imitation Learning

1 code implementation NeurIPS 2019 Yiming Ding, Carlos Florensa, Mariano Phielipp, Pieter Abbeel

Designing rewards for Reinforcement Learning (RL) is challenging because it needs to convey the desired task, be efficient to optimize, and be easy to compute.

Imitation Learning Reinforcement Learning (RL)

Hierarchical Policy Learning is Sensitive to Goal Space Design

no code implementations4 May 2019 Zach Dwiel, Madhavun Candadai, Mariano Phielipp, Arjun K. Bansal

Hierarchy in reinforcement learning agents allows for control at multiple time scales yielding improved sample efficiency, the ability to deal with long time horizons and transferability of sub-policies to tasks outside the training distribution.

Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.