no code implementations • 31 Dec 2020 • Yuchen Xie, Raghu Bollapragada, Richard Byrd, Jorge Nocedal
The motivation for this paper stems from the desire to develop an adaptive sampling method for solving constrained optimization problems in which the objective function is stochastic and the constraints are deterministic.
1 code implementation • 9 Oct 2020 • Hao-Jun Michael Shi, Yuchen Xie, Richard Byrd, Jorge Nocedal
This paper describes an extension of the BFGS and L-BFGS methods for the minimization of a nonlinear function subject to errors.
Optimization and Control
no code implementations • 30 Oct 2017 • Raghu Bollapragada, Richard Byrd, Jorge Nocedal
In this paper, we propose a stochastic optimization method that adaptively controls the sample size used in the computation of gradient approximations.
no code implementations • 27 Sep 2016 • Raghu Bollapragada, Richard Byrd, Jorge Nocedal
The paper studies the solution of stochastic optimization problems in which approximations to the gradient and Hessian are obtained through subsampling.