no code implementations • 9 Sep 2023 • Neha S. Wadia, Yatin Dandi, Michael I. Jordan
The rapid progress in machine learning in recent years has been based on a highly productive connection to gradient-based optimization.
no code implementations • 28 Sep 2020 • Neha S. Wadia, Daniel Duckworth, Samuel Stern Schoenholz, Ethan Dyer, Jascha Sohl-Dickstein
We show that both data whitening and second order optimization can harm or entirely prevent generalization.
no code implementations • 17 Aug 2020 • Neha S. Wadia, Daniel Duckworth, Samuel S. Schoenholz, Ethan Dyer, Jascha Sohl-Dickstein
We show that both data whitening and second order optimization can harm or entirely prevent generalization.
no code implementations • 23 Mar 2020 • Charles G. Frye, James Simon, Neha S. Wadia, Andrew Ligeralde, Michael R. DeWeese, Kristofer E. Bouchard
Despite the fact that the loss functions of deep neural networks are highly non-convex, gradient-based optimization algorithms converge to approximately the same performance from many random initial points.
no code implementations • 29 Jan 2019 • Charles G. Frye, Neha S. Wadia, Michael R. DeWeese, Kristofer E. Bouchard
Numerically locating the critical points of non-convex surfaces is a long-standing problem central to many fields.