no code implementations • 13 Oct 2023 • Shengbo Wang, Jose Blanchet, Peter Glynn
In this context, the existing literature provides a sample complexity upper bound of $\widetilde O(|S||A|t_{\text{mix}}^2 \epsilon^{-2})$ and a lower bound of $\Omega(|S||A|t_{\text{mix}} \epsilon^{-2})$.
no code implementations • 15 Feb 2023 • Shengbo Wang, Jose Blanchet, Peter Glynn
We consider the optimal sample complexity theory of tabular reinforcement learning (RL) for maximizing the infinite horizon discounted reward in a Markov decision process (MDP).
no code implementations • 14 Mar 2022 • Jin Xie, Teng Zhang, Jose Blanchet, Peter Glynn, Matthew Randolph, David Scheinker
In order for an algorithm to see sustained use, it must be compatible with changes to hospital capacity, patient volumes, and scheduling practices.
no code implementations • 13 Feb 2022 • Yuan Shi, Saied Mahdian, Jose Blanchet, Peter Glynn, Andrew Y. Shin, David Scheinker
Using data from cardiovascular surgery patients with long and highly variable post-surgical lengths of stay (LOS), we develop a modeling framework to reduce recovery unit congestion.
no code implementations • 24 Aug 2019 • Shubhada Agrawal, Sandeep Juneja, Peter Glynn
We then propose a $\delta$-correct algorithm that matches the lower bound as $\delta$ reduces to zero under the mild restriction that a known bound on the expectation of $(1+\epsilon)^{th}$ moment of the underlying random variables exists, for $\epsilon > 0$.
no code implementations • 7 Jun 2019 • Saied Mahdian, Jose Blanchet, Peter Glynn
We propose a family of relaxations of the optimal transport problem which regularize the problem by introducing an additional minimization step over a small region around one of the underlying transporting measures.
no code implementations • 30 Jan 2019 • Casey Chu, Jose Blanchet, Peter Glynn
This paper provides a unifying view of a wide range of problems of interest in machine learning by framing them as the minimization of functionals defined on the space of probability measures.
no code implementations • ICML 2018 • Zhengyuan Zhou, Panayotis Mertikopoulos, Nicholas Bambos, Peter Glynn, Yinyu Ye, Li-Jia Li, Li Fei-Fei
One of the most widely used optimization methods for large-scale machine learning problems is distributed asynchronous stochastic gradient descent (DASGD).
no code implementations • 5 May 2018 • Mansur Arief, Peter Glynn, Ding Zhao
Various automobile and mobility companies, for instance Ford, Uber and Waymo, are currently testing their pre-produced autonomous vehicle (AV) fleets on the public roads.
no code implementations • 4 Apr 2018 • Abhishek Gupta, Rahul Jain, Peter Glynn
In many branches of engineering, Banach contraction mapping theorem is employed to establish the convergence of certain deterministic algorithms.
no code implementations • 18 Jun 2017 • Zhengyuan Zhou, Panayotis Mertikopoulos, Nicholas Bambos, Stephen Boyd, Peter Glynn
In this paper, we examine the convergence of mirror descent in a class of stochastic optimization problems that are not necessarily convex (or even quasi-convex), and which we call variationally coherent.
no code implementations • 11 Oct 2016 • John Duchi, Peter Glynn, Hongseok Namkoong
We study statistical inference and distributionally robust solution methods for stochastic optimization problems, focusing on confidence intervals for optimal values and solutions that achieve exact coverage asymptotically.
no code implementations • 16 Jul 2015 • Peter Glynn, Sandeep Juneja
Consider the problem of finding a population or a probability distribution amongst many with the largest mean when these means are unknown but population samples can be simulated or otherwise generated.