Search Results for author: Daniel Graves

Found 15 papers, 3 papers with code

LISPR: An Options Framework for Policy Reuse with Reinforcement Learning

no code implementations29 Dec 2020 Daniel Graves, Jun Jin, Jun Luo

Our approach facilitates the learning of new policies by (1) maximizing the target MDP reward with the help of the black-box option, and (2) returning the agent to states in the learned initiation set of the black-box option where it is already optimal.

Continual Learning reinforcement-learning +1

Offline Learning of Counterfactual Predictions for Real-World Robotic Reinforcement Learning

no code implementations11 Nov 2020 Jun Jin, Daniel Graves, Cameron Haigh, Jun Luo, Martin Jagersand

We consider real-world reinforcement learning (RL) of robotic manipulation tasks that involve both visuomotor skills and contact-rich skills.

counterfactual reinforcement-learning +1

Hyperoctahedral Homology for Involutive Algebras

no code implementations6 Nov 2020 Daniel Graves

Hyperoctahedral homology is the homology theory associated to the hyperoctahedral crossed simplicial group.

Algebraic Topology 55N35, 13D03, 55U15, 55P47

Affordance as general value function: A computational model

no code implementations27 Oct 2020 Daniel Graves, Johannes Günther, Jun Luo

General value functions (GVFs) in the reinforcement learning (RL) literature are long-term predictive summaries of the outcomes of agents following specific policies in the environment.

Autonomous Driving Reinforcement Learning (RL)

What About Inputing Policy in Value Function: Policy Representation and Policy-extended Value Function Approximator

1 code implementation NeurIPS 2021 Hongyao Tang, Zhaopeng Meng, Jianye Hao, Chen Chen, Daniel Graves, Dong Li, Changmin Yu, Hangyu Mao, Wulong Liu, Yaodong Yang, Wenyuan Tao, Li Wang

We study Policy-extended Value Function Approximator (PeVFA) in Reinforcement Learning (RL), which extends conventional value function approximator (VFA) to take as input not only the state (and action) but also an explicit policy representation.

Continuous Control Contrastive Learning +3

Learning predictive representations in autonomous driving to improve deep reinforcement learning

no code implementations26 Jun 2020 Daniel Graves, Nhat M. Nguyen, Kimia Hassanzadeh, Jun Jin

Reinforcement learning using a novel predictive representation is applied to autonomous driving to accomplish the task of driving between lane markings where substantial benefits in performance and generalization are observed on unseen test roads in both simulation and on a real Jackal robot.

Autonomous Driving reinforcement-learning +1

Perception as prediction using general value functions in autonomous driving applications

no code implementations24 Jan 2020 Daniel Graves, Kasra Rezaee, Sean Scheideman

We demonstrate perception as prediction by learning to predict an agent's front safety and rear safety with GVFs, which encapsulate anticipation of the behavior of the vehicle in front and in the rear, respectively.

Autonomous Driving

Mapless Navigation among Dynamics with Social-safety-awareness: a reinforcement learning approach from 2D laser scans

no code implementations8 Nov 2019 Jun Jin, Nhat M. Nguyen, Nazmus Sakib, Daniel Graves, Hengshuai Yao, Martin Jagersand

We observe that our method demonstrates time-efficient path planning behavior with high success rate in mapless navigation tasks.

Robotics

Fixed-Horizon Temporal Difference Methods for Stable Reinforcement Learning

no code implementations9 Sep 2019 Kristopher De Asis, Alan Chan, Silviu Pitis, Richard S. Sutton, Daniel Graves

We explore fixed-horizon temporal difference (TD) methods, reinforcement learning algorithms for a new kind of value function that predicts the sum of rewards over a $\textit{fixed}$ number of future time steps.

Q-Learning reinforcement-learning +1

Importance Resampling for Off-policy Prediction

2 code implementations NeurIPS 2019 Matthew Schlegel, Wesley Chung, Daniel Graves, Jian Qian, Martha White

Importance sampling (IS) is a common reweighting strategy for off-policy prediction in reinforcement learning.

Importance Resampling for Off-policy Policy Evaluation

no code implementations27 Sep 2018 Matthew Schlegel, Wesley Chung, Daniel Graves, Martha White

We propose Importance Resampling (IR) for off-policy learning, that resamples experience from the replay buffer and applies a standard on-policy update.

Cannot find the paper you are looking for? You can Submit a new open access paper.