Search Results for author: Loucas Pillaud-Vivien

Found 15 papers, 5 papers with code

An Ordering of Divergences for Variational Inference with Factorized Gaussian Approximations

1 code implementation20 Mar 2024 Charles C. Margossian, Loucas Pillaud-Vivien, Lawrence K. Saul

Our analysis covers the KL divergence, the R\'enyi divergences, and a score-based divergence that compares $\nabla\log p$ and $\nabla\log q$.

valid Variational Inference

Computational-Statistical Gaps in Gaussian Single-Index Models

no code implementations8 Mar 2024 Alex Damian, Loucas Pillaud-Vivien, Jason D. Lee, Joan Bruna

Single-Index Models are high-dimensional regression problems with planted structure, whereby labels depend on an unknown one-dimensional projection of the input via a generic, non-linear, and potentially non-deterministic transformation.

Batch and match: black-box variational inference with a score-based divergence

no code implementations22 Feb 2024 Diana Cai, Chirag Modi, Loucas Pillaud-Vivien, Charles C. Margossian, Robert M. Gower, David M. Blei, Lawrence K. Saul

We analyze the convergence of BaM when the target distribution is Gaussian, and we prove that in the limit of infinite batch size the variational parameter updates converge exponentially quickly to the target mean and covariance.

Variational Inference

On Learning Gaussian Multi-index Models with Gradient Flow

no code implementations30 Oct 2023 Alberto Bietti, Joan Bruna, Loucas Pillaud-Vivien

We study gradient flow on the multi-index regression problem for high-dimensional Gaussian data.

On Single Index Models beyond Gaussian Data

no code implementations28 Jul 2023 Joan Bruna, Loucas Pillaud-Vivien, Aaron Zweig

Sparse high-dimensional functions have arisen as a rich framework to study the behavior of gradient-descent methods using shallow neural networks, showcasing their ability to perform feature learning beyond linear models.

Kernelized Diffusion maps

1 code implementation13 Feb 2023 Loucas Pillaud-Vivien, Francis Bach

Spectral clustering and diffusion maps are celebrated dimensionality reduction algorithms built on eigen-elements related to the diffusive structure of the data.

Clustering Dimensionality Reduction

SGD with Large Step Sizes Learns Sparse Features

1 code implementation11 Oct 2022 Maksym Andriushchenko, Aditya Varre, Loucas Pillaud-Vivien, Nicolas Flammarion

We present empirical observations that commonly used large step sizes (i) lead the iterates to jump from one side of a valley to the other causing loss stabilization, and (ii) this stabilization induces a hidden stochastic dynamics orthogonal to the bouncing directions that biases it implicitly toward sparse predictors.

Label noise (stochastic) gradient descent implicitly solves the Lasso for quadratic parametrisation

no code implementations20 Jun 2022 Loucas Pillaud-Vivien, Julien Reygner, Nicolas Flammarion

Understanding the implicit bias of training algorithms is of crucial importance in order to explain the success of overparametrised neural networks.

Gradient flow dynamics of shallow ReLU networks for square loss and orthogonal inputs

1 code implementation2 Jun 2022 Etienne Boursier, Loucas Pillaud-Vivien, Nicolas Flammarion

The training of neural networks by gradient descent methods is a cornerstone of the deep learning revolution.

Implicit Bias of SGD for Diagonal Linear Networks: a Provable Benefit of Stochasticity

no code implementations NeurIPS 2021 Scott Pesme, Loucas Pillaud-Vivien, Nicolas Flammarion

Understanding the implicit bias of training algorithms is of crucial importance in order to explain the success of overparametrised neural networks.

Last iterate convergence of SGD for Least-Squares in the Interpolation regime.

no code implementations NeurIPS 2021 Aditya Vardhan Varre, Loucas Pillaud-Vivien, Nicolas Flammarion

Motivated by the recent successes of neural networks that have the ability to fit the data perfectly \emph{and} generalize well, we study the noiseless model in the fundamental least-squares setup.

Stochastic Optimization

Last iterate convergence of SGD for Least-Squares in the Interpolation regime

no code implementations NeurIPS 2021 Aditya Varre, Loucas Pillaud-Vivien, Nicolas Flammarion

Motivated by the recent successes of neural networks that have the ability to fit the data perfectly and generalize well, we study the noiseless model in the fundamental least-squares setup.

Stochastic Optimization

Overcoming the curse of dimensionality with Laplacian regularization in semi-supervised learning

2 code implementations NeurIPS 2021 Vivien Cabannes, Loucas Pillaud-Vivien, Francis Bach, Alessandro Rudi

As annotations of data can be scarce in large-scale practical problems, leveraging unlabelled examples is one of the most important aspects of machine learning.

Clustering

Exponential convergence of testing error for stochastic gradient methods

no code implementations13 Dec 2017 Loucas Pillaud-Vivien, Alessandro Rudi, Francis Bach

We consider binary classification problems with positive definite kernels and square loss, and study the convergence rates of stochastic gradient methods.

Binary Classification Classification +1

Cannot find the paper you are looking for? You can Submit a new open access paper.