Search Results for author: Eric Graves

Found 4 papers, 1 papers with code

Value-aware Importance Weighting for Off-policy Reinforcement Learning

no code implementations27 Jun 2023 Kristopher De Asis, Eric Graves, Richard S. Sutton

Importance sampling is a central idea underlying off-policy prediction in reinforcement learning.

reinforcement-learning

Importance Sampling Placement in Off-Policy Temporal-Difference Methods

no code implementations18 Mar 2022 Eric Graves, Sina Ghiassian

A central challenge to applying many off-policy reinforcement learning algorithms to real world problems is the variance introduced by importance sampling.

Off-Policy Actor-Critic with Emphatic Weightings

1 code implementation16 Nov 2021 Eric Graves, Ehsan Imani, Raksha Kumaraswamy, Martha White

A variety of theoretically-sound policy gradient algorithms exist for the on-policy setting due to the policy gradient theorem, which provides a simplified form for the gradient.

An Off-policy Policy Gradient Theorem Using Emphatic Weightings

no code implementations NeurIPS 2018 Ehsan Imani, Eric Graves, Martha White

There have been a host of theoretically sound algorithms proposed for the on-policy setting, due to the existence of the policy gradient theorem which provides a simplified form for the gradient.

Policy Gradient Methods

Cannot find the paper you are looking for? You can Submit a new open access paper.