no code implementations • 30 Jun 2019 • Alessio Spantini, Ricardo Baptista, Youssef Marzouk
We consider filtering in high-dimensional non-Gaussian state-space models with intractable transition kernels, nonlinear and possibly chaotic dynamics, and sparse observations in space and time.
1 code implementation • NeurIPS 2020 • Michael C. Brennan, Daniele Bigoni, Olivier Zahm, Alessio Spantini, Youssef Marzouk
We prove weak convergence of the generated sequence of distributions to the posterior, and we demonstrate the benefits of the framework on challenging inference problems in machine learning and differential equations, using inverse autoregressive flows and polynomial maps as examples of the underlying density estimators.
1 code implementation • NeurIPS 2018 • Gianluca Detommaso, Tiangang Cui, Alessio Spantini, Youssef Marzouk, Robert Scheichl
Stein variational gradient descent (SVGD) was recently proposed as a general purpose nonparametric variational inference algorithm [Liu & Wang, NIPS 2016]: it minimizes the Kullback-Leibler divergence between the target distribution and its approximation by implementing a form of functional gradient descent on a reproducing kernel Hilbert space.
no code implementations • 17 Mar 2017 • Alessio Spantini, Daniele Bigoni, Youssef Marzouk
In the context of statistics and machine learning, these transformations can be used to couple a tractable "reference" measure (e. g., a standard Gaussian) with a target measure of interest.