2 code implementations • 28 Oct 2022 • Alexandre Perez-Lebel, Marine Le Morvan, Gaël Varoquaux
Yet calibration is not enough: even a perfectly calibrated classifier with the best possible accuracy can have confidence scores that are far from the true posterior probabilities.
1 code implementation • 17 Feb 2022 • Alexandre Perez-Lebel, Gaël Varoquaux, Marine Le Morvan, Julie Josse, Jean-Baptiste Poline
Using gradient-boosted trees, we compare native support for missing values with simple and state-of-the-art imputation prior to learning.
no code implementations • NeurIPS 2021 • Marine Le Morvan, Julie Josse, Erwan Scornet, Gael Varoquaux
In fact, we show that on perfectly imputed data the best regression function will generally be discontinuous, which makes it hard to learn.
1 code implementation • 1 Jun 2021 • Marine Le Morvan, Julie Josse, Erwan Scornet, Gaël Varoquaux
In fact, we show that on perfectly imputed data the best regression function will generally be discontinuous, which makes it hard to learn.
no code implementations • NeurIPS 2020 • Marine Le Morvan, Julie Josses, Thomas Moreau, Erwan Scornet, Gael Varoquaux
We provide an upper bound on the Bayes risk of NeuMiss networks, and show that they have good predictive accuracy with both a number of parameters and a computational complexity independent of the number of missing data patterns.
no code implementations • 3 Jul 2020 • Marine Le Morvan, Julie Josse, Thomas Moreau, Erwan Scornet, Gaël Varoquaux
We provide an upper bound on the Bayes risk of NeuMiss networks, and show that they have good predictive accuracy with both a number of parameters and a computational complexity independent of the number of missing data patterns.
1 code implementation • 3 Feb 2020 • Marine Le Morvan, Nicolas Prost, Julie Josse, Erwan Scornet, Gaël Varoquaux
In the particular Gaussian case, it can be written as a linear function of multiway interactions between the observed data and the various missing-value indicators.
no code implementations • ICML 2018 • Marine Le Morvan, Jean-Philippe Vert
Learning sparse linear models with two-way interactions is desirable in many application domains such as genomics.
no code implementations • 1 Jun 2017 • Marine Le Morvan, Jean-Philippe Vert
Quantile normalisation is a popular normalisation method for data subject to unwanted variations such as images, speech, or genomic data.