Search Results for author: Andrew M. Stuart

Found 24 papers, 5 papers with code

Operator Learning: Algorithms and Analysis

no code implementations24 Feb 2024 Nikola B. Kovachki, Samuel Lanthaler, Andrew M. Stuart

This review article summarizes recent progress and the current state of our theoretical understanding of neural operators, focusing on an approximation theoretic point of view.

Model Discovery Operator learning

Modeling groundwater levels in California's Central Valley by hierarchical Gaussian process and neural network regression

no code implementations23 Oct 2023 Anshuman Pradhan, Kyra H. Adams, Venkat Chandrasekaran, Zhen Liu, John T. Reager, Andrew M. Stuart, Michael J. Turmon

Modeling groundwater levels continuously across California's Central Valley (CV) hydrological system is challenging due to low-quality well data which is sparsely and noisily sampled across time and space.

Gaussian Processes regression +1

The Parametric Complexity of Operator Learning

no code implementations28 Jun 2023 Samuel Lanthaler, Andrew M. Stuart

The first contribution of this paper is to prove that for general classes of operators which are characterized only by their $C^r$- or Lipschitz-regularity, operator learning suffers from a ``curse of parametric complexity'', which is an infinite-dimensional analogue of the well-known curse of dimensionality encountered in high-dimensional approximation problems.

Operator learning

Learning Homogenization for Elliptic Operators

2 code implementations21 Jun 2023 Kaushik Bhattacharya, Nikola Kovachki, Aakila Rajan, Andrew M. Stuart, Margaret Trautner

However, a major challenge in data-driven learning approaches for this problem has remained unexplored: the impact of discontinuities and corner interfaces in the underlying material.

The Nonlocal Neural Operator: Universal Approximation

no code implementations26 Apr 2023 Samuel Lanthaler, Zongyi Li, Andrew M. Stuart

A popular variant of neural operators is the Fourier neural operator (FNO).

Operator learning

Second Order Ensemble Langevin Method for Sampling and Inverse Problems

no code implementations9 Aug 2022 Ziming Liu, Andrew M. Stuart, YiXuan Wang

We propose a sampling method based on an ensemble approximation of second order Langevin dynamics.

Position

A Framework for Machine Learning of Model Error in Dynamical Systems

no code implementations14 Jul 2021 Matthew E. Levine, Andrew M. Stuart

For ergodic continuous-time systems, we prove that both excess risk and generalization error are bounded above by terms that diminish with the square-root of T, the time-interval over which training data is specified.

BIG-bench Machine Learning Learning Theory

Ensemble Inference Methods for Models With Noisy and Expensive Likelihoods

no code implementations7 Apr 2021 Oliver R. A. Dunbar, Andrew B. Duncan, Andrew M. Stuart, Marie-Therese Wolfram

The ensemble Kalman methods are shown to behave favourably in the presence of noise in the parameter-to-data map, whereas Langevin methods are adversely affected.

Iterated Kalman Methodology For Inverse Problems

1 code implementation2 Feb 2021 Daniel Z. Huang, Tapio Schneider, Andrew M. Stuart

In this paper, we work with the ExKI, EKI, and a variant on EKI which we term unscented Kalman inversion (UKI).

Numerical Analysis Numerical Analysis Dynamical Systems

Calibration and Uncertainty Quantification of Convective Parameters in an Idealized GCM

no code implementations24 Dec 2020 Oliver R. A. Dunbar, Alfredo Garbuno-Inigo, Tapio Schneider, Andrew M. Stuart

Here we demonstrate an approach to model calibration and uncertainty quantification that requires only $O(10^2)$ model runs and can accommodate internal climate variability.

Gaussian Processes Statistics Theory Statistics Theory

Posterior Consistency of Semi-Supervised Regression on Graphs

no code implementations25 Jul 2020 Andrea L. Bertozzi, Bamdad Hosseini, Hao Li, Kevin Miller, Andrew M. Stuart

Graph-based semi-supervised regression (SSR) is the problem of estimating the value of a function on a weighted graph from its values (labels) on a small subset of the vertices.

Clustering regression

Consistency of Empirical Bayes And Kernel Flow For Hierarchical Parameter Estimation

no code implementations22 May 2020 Yifan Chen, Houman Owhadi, Andrew M. Stuart

The purpose of this paper is to study two paradigms of learning hierarchical parameters: one is from the probabilistic Bayesian perspective, in particular, the empirical Bayes approach that has been largely used in Bayesian statistics; the other is from the deterministic and approximation theoretic view, and in particular the kernel flow algorithm that was proposed recently in the machine learning literature.

BIG-bench Machine Learning

The Random Feature Model for Input-Output Maps between Banach Spaces

1 code implementation20 May 2020 Nicholas H. Nelsen, Andrew M. Stuart

Well known to the machine learning community, the random feature model is a parametric approximation to kernel interpolation or regression methods.

Model Reduction and Neural Networks for Parametric PDEs

no code implementations7 May 2020 Kaushik Bhattacharya, Bamdad Hosseini, Nikola B. Kovachki, Andrew M. Stuart

We develop a general framework for data-driven approximation of input-output maps between infinite-dimensional spaces.

Spectral Analysis Of Weighted Laplacians Arising In Data Clustering

no code implementations13 Sep 2019 Franca Hoffmann, Bamdad Hosseini, Assad A. Oberai, Andrew M. Stuart

Graph Laplacians computed from weighted adjacency matrices are widely used to identify geometric structure in data, and clusters in particular; their spectral properties play a central role in a number of unsupervised and semi-supervised learning algorithms.

Clustering

Consistency of semi-supervised learning algorithms on graphs: Probit and one-hot methods

no code implementations18 Jun 2019 Franca Hoffmann, Bamdad Hosseini, Zhi Ren, Andrew M. Stuart

Graph-based semi-supervised learning is the problem of propagating labels from a small number of labelled data points to a larger set of unlabelled data.

Binary Classification General Classification +1

Continuous Time Analysis of Momentum Methods

no code implementations10 Jun 2019 Nikola B. Kovachki, Andrew M. Stuart

Firstly we show that standard implementations of fixed momentum methods approximate a time-rescaled gradient descent flow, asymptotically as the learning rate shrinks to zero; this result does not distinguish momentum methods from pure gradient descent, in the limit of vanishing learning rate.

Ensemble Kalman Inversion: A Derivative-Free Technique For Machine Learning Tasks

no code implementations10 Aug 2018 Nikola B. Kovachki, Andrew M. Stuart

The standard probabilistic perspective on machine learning gives rise to empirical risk-minimization tasks that are frequently solved by stochastic gradient descent (SGD) and variants thereof.

BIG-bench Machine Learning

Large Data and Zero Noise Limits of Graph-Based Semi-Supervised Learning Algorithms

no code implementations23 May 2018 Matthew M. Dunlop, Dejan Slepčev, Andrew M. Stuart, Matthew Thorpe

Scalings in which the graph Laplacian approaches a differential operator in the large graph limit are used to develop understanding of a number of algorithms for semi-supervised learning; in particular the extension, to this graph setting, of the probit algorithm, level set and kriging methods, are studied.

Dimension-Robust MCMC in Bayesian Inverse Problems

no code implementations9 Mar 2018 Victor Chen, Matthew M. Dunlop, Omiros Papaspiliopoulos, Andrew M. Stuart

One popular formulation of such problems is as Bayesian inverse problems, where a prior distribution is used to regularize inference on a high-dimensional latent state, typically a function or a field.

Active Learning Efficient Exploration +4

Uncertainty quantification in graph-based classification of high dimensional data

no code implementations26 Mar 2017 Andrea L. Bertozzi, Xiyang Luo, Andrew M. Stuart, Konstantinos C. Zygalakis

In this paper we introduce, develop algorithms for, and investigate the properties of, a variety of Bayesian models for the task of binary classification; via the posterior distribution on the classification labels, these methods automatically give measures of uncertainty.

Binary Classification Classification +3

Posterior Consistency for Gaussian Process Approximations of Bayesian Posterior Distributions

1 code implementation7 Mar 2016 Andrew M. Stuart, Aretha L. Teckentrup

We prove error bounds on the Hellinger distance between the true posterior distribution and various approximations based on the Gaussian process emulator.

Numerical Analysis

Cannot find the paper you are looking for? You can Submit a new open access paper.