Search Results for author: Andrew Jacobsen

Found 6 papers, 0 papers with code

Unconstrained Online Learning with Unbounded Losses

no code implementations8 Jun 2023 Andrew Jacobsen, Ashok Cutkosky

Algorithms for online learning typically require one or more boundedness assumptions: that the domain is bounded, that the losses are Lipschitz, or both.

Parameter-free Mirror Descent

no code implementations26 Feb 2022 Andrew Jacobsen, Ashok Cutkosky

We develop a modified online mirror descent framework that is suitable for building adaptive and parameter-free algorithms in unbounded domains.

Parameter-free Gradient Temporal Difference Learning

no code implementations10 May 2021 Andrew Jacobsen, Alan Chan

In parallel, progress in online learning has provided parameter-free methods that achieve minimax optimal guarantees up to logarithmic terms, but their application in reinforcement learning has yet to be explored.

reinforcement-learning Reinforcement Learning (RL)

Meta-descent for Online, Continual Prediction

no code implementations17 Jul 2019 Andrew Jacobsen, Matthew Schlegel, Cameron Linke, Thomas Degris, Adam White, Martha White

This paper investigates different vector step-size adaptation approaches for non-stationary online, continual prediction problems.

Second-order methods Time Series +1

General Value Function Networks

no code implementations18 Jul 2018 Matthew Schlegel, Andrew Jacobsen, Zaheer Abbas, Andrew Patterson, Adam White, Martha White

A general purpose strategy for state construction is to learn the state update using a Recurrent Neural Network (RNN), which updates the internal state using the current internal state and the most recent observation.

Continuous Control Decision Making

Cannot find the paper you are looking for? You can Submit a new open access paper.