Search Results for author: Brandon Trabucco

Found 10 papers, 6 papers with code

Effective Data Augmentation With Diffusion Models

1 code implementation7 Feb 2023 Brandon Trabucco, Kyle Doherty, Max Gurinas, Ruslan Salakhutdinov

Data augmentation is one of the most prevalent tools in deep learning, underpinning many recent advances, including those from classification, generative models, and representation learning.

Data Augmentation Few-Shot Image Classification +1

AnyMorph: Learning Transferable Polices By Inferring Agent Morphology

no code implementations17 Jun 2022 Brandon Trabucco, Mariano Phielipp, Glen Berseth

Ours is the first reinforcement learning algorithm that can train a policy to generalize to new agent morphologies without requiring a description of the agent's morphology in advance.

reinforcement-learning Reinforcement Learning (RL) +1

Design-Bench: Benchmarks for Data-Driven Offline Model-Based Optimization

3 code implementations17 Feb 2022 Brandon Trabucco, Xinyang Geng, Aviral Kumar, Sergey Levine

To address this, we present Design-Bench, a benchmark for offline MBO with a unified evaluation protocol and reference implementations of recent methods.

Conservative Objective Models for Effective Offline Model-Based Optimization

2 code implementations14 Jul 2021 Brandon Trabucco, Aviral Kumar, Xinyang Geng, Sergey Levine

Computational design problems arise in a number of settings, from synthetic biology to computer architectures.

Synthetic Datasets for Neural Program Synthesis

no code implementations ICLR 2019 Richard Shin, Neel Kant, Kavi Gupta, Christopher Bender, Brandon Trabucco, Rishabh Singh, Dawn Song

The goal of program synthesis is to automatically generate programs in a particular language from corresponding specifications, e. g. input-output behavior.

Program Synthesis

Inter-Level Cooperation in Hierarchical Reinforcement Learning

1 code implementation5 Dec 2019 Abdul Rahman Kreidieh, Glen Berseth, Brandon Trabucco, Samyak Parajuli, Sergey Levine, Alexandre M. Bayen

This allows us to draw on connections between communication and cooperation in multi-agent RL, and demonstrate the benefits of increased cooperation between sub-policies on the training performance of the overall policy.

Hierarchical Reinforcement Learning reinforcement-learning +1

Inferring the Optimal Policy using Markov Chain Monte Carlo

no code implementations16 Nov 2019 Brandon Trabucco, Albert Qu, Simon Li, Ganeshkumar Ashokavardhanan

Existing methods for estimating the optimal stochastic control policy rely on high variance estimates of the policy descent.

Cannot find the paper you are looking for? You can Submit a new open access paper.