1 code implementation • 21 Oct 2023 • Nicholas S. Moore, Eric C. Cyr, Peter Ohm, Christopher M. Siefert, Raymond S. Tuminaro
With the recent interest in scientific machine learning, it is natural to ask how sparse matrix computations can leverage neural networks (NN).
1 code implementation • 7 Mar 2022 • Gordon Euhyun Moon, Eric C. Cyr
Parallelizing Gated Recurrent Unit (GRU) networks is a challenging task, as the training procedure of GRU is inherently sequential.
no code implementations • 27 Jan 2021 • Kookjin Lee, Nathaniel A. Trask, Ravi G. Patel, Mamikon A. Gulian, Eric C. Cyr
Approximation theorists have established best-in-class optimal approximation rates of deep neural networks by utilizing their ability to simultaneously emulate partitions of unity and monomials.
1 code implementation • 25 Sep 2020 • Ravi G. Patel, Nathaniel A. Trask, Mitchell A. Wood, Eric C. Cyr
The application of deep learning toward discovery of data-driven models requires careful application of inductive biases to obtain a description of physics which is both accurate and robust.
no code implementations • 17 Jun 2020 • Ravi G. Patel, Nathaniel A. Trask, Mamikon A. Gulian, Eric C. Cyr
By alternating between a second-order method to find globally optimal parameters for the linear layer and gradient descent to train the hidden layers, we ensure an optimal fit of the adaptive basis to data throughout training.
1 code implementation • 19 Dec 2019 • Eric C. Cyr, Stefanie Günther, Jacob B. Schroder
This paper investigates multilevel initialization strategies for training very deep neural networks with a layer-parallel multigrid solver.
no code implementations • 10 Dec 2019 • Eric C. Cyr, Mamikon A. Gulian, Ravi G. Patel, Mauro Perego, Nathaniel A. Trask
Motivated by the gap between theoretical optimal approximation rates of deep neural networks (DNNs) and the accuracy realized in practice, we seek to improve the training of DNNs.