Search Results for author: Hoang M. Le

Found 12 papers, 3 papers with code

OPSurv: Orthogonal Polynomials Quadrature Algorithm for Survival Analysis

no code implementations2 Feb 2024 Lilian W. Bialokozowicz, Hoang M. Le, Tristan Sylvain, Peter A. I. Forsyth, Vineel Nagisetty, Greg Mori

This paper introduces the Orthogonal Polynomials Quadrature Algorithm for Survival Analysis (OPSurv), a new method providing time-continuous functional outputs for both single and competing risks scenarios in survival analysis.

Survival Analysis

GamutMLP: A Lightweight MLP for Color Loss Recovery

no code implementations CVPR 2023 Hoang M. Le, Brian Price, Scott Cohen, Michael S. Brown

Inspired by neural implicit representations for 2D images, we propose a method that optimizes a lightweight multi-layer-perceptron (MLP) model during the gamut reduction step to predict the clipped values.

Policy Optimization with Linear Temporal Logic Constraints

no code implementations20 Jun 2022 Cameron Voloshin, Hoang M. Le, Swarat Chaudhuri, Yisong Yue

We study the problem of policy optimization (PO) with linear temporal logic (LTL) constraints.

Empirical Study of Off-Policy Policy Evaluation for Reinforcement Learning

3 code implementations15 Nov 2019 Cameron Voloshin, Hoang M. Le, Nan Jiang, Yisong Yue

We offer an experimental benchmark and empirical study for off-policy policy evaluation (OPE) in reinforcement learning, which is a key problem in many safety critical applications.

Benchmarking Experimental Design +2

Imitation-Projected Programmatic Reinforcement Learning

no code implementations NeurIPS 2019 Abhinav Verma, Hoang M. Le, Yisong Yue, Swarat Chaudhuri

First, we view our learning task as optimization in policy space, modulo the constraint that the desired policy has a programmatic representation, and solve this optimization problem using a form of mirror descent that takes a gradient step into the unconstrained policy space and then projects back onto the constrained space.

Continuous Control Imitation Learning +3

Batch Policy Learning under Constraints

2 code implementations20 Mar 2019 Hoang M. Le, Cameron Voloshin, Yisong Yue

When learning policies for real-world domains, two important questions arise: (i) how to efficiently use pre-collected off-policy, non-optimal behavior data; and (ii) how to mediate among different competing objectives and constraints.

A Control Lyapunov Perspective on Episodic Learning via Projection to State Stability

no code implementations18 Mar 2019 Andrew J. Taylor, Victor D. Dorobantu, Meera Krishnamoorthy, Hoang M. Le, Yisong Yue, Aaron D. Ames

The goal of this paper is to understand the impact of learning on control synthesis from a Lyapunov function perspective.

Episodic Learning with Control Lyapunov Functions for Uncertain Robotic Systems

no code implementations4 Mar 2019 Andrew J. Taylor, Victor D. Dorobantu, Hoang M. Le, Yisong Yue, Aaron D. Ames

Many modern nonlinear control methods aim to endow systems with guaranteed properties, such as stability or safety, and have been successfully applied to the domain of robotics.

Coordinated Multi-Agent Imitation Learning

no code implementations ICML 2017 Hoang M. Le, Yisong Yue, Peter Carr, Patrick Lucey

We study the problem of imitation learning from demonstrations of multiple coordinating agents.

Imitation Learning

Smooth Imitation Learning for Online Sequence Prediction

2 code implementations3 Jun 2016 Hoang M. Le, Andrew Kang, Yisong Yue, Peter Carr

We study the problem of smooth imitation learning for online sequence prediction, where the goal is to train a policy that can smoothly imitate demonstrated behavior in a dynamic and continuous environment in response to online, sequential context input.

Imitation Learning regression

Learning Online Smooth Predictors for Realtime Camera Planning Using Recurrent Decision Trees

no code implementations CVPR 2016 Jianhui Chen, Hoang M. Le, Peter Carr, Yisong Yue, James J. Little

We study the problem of online prediction for realtime camera planning, where the goal is to predict smooth trajectories that correctly track and frame objects of interest (e. g., players in a basketball game).

Cannot find the paper you are looking for? You can Submit a new open access paper.