1 code implementation • 3 May 2022 • Thomas Lartigue, Sach Mukherjee
Building on developments at the intersection of unsupervised learning and regularised regression, we propose an approach for heterogeneous data that allows joint learning of (i) explicit multivariate feature distributions, (ii) high-dimensional regression models and (iii) latent group labels, with both (i) and (ii) specific to latent groups and both elements informing (iii).
no code implementations • 11 Apr 2022 • Thomas Lartigue, Sach Mukherjee
Motivated by such applications, in this paper we ask whether linear projections can preserve differences in second order structure between latent groups.
1 code implementation • 19 Jun 2020 • Thomas Lartigue, Stanley Durrleman, Stéphanie Allassonnière
We demonstrate on synthetic and real data how this method fulfils its goal and succeeds in identifying the sub-populations where the Mixtures of GGM are disrupted by the effect of the co-features.
no code implementations • 23 Mar 2020 • Thomas Lartigue, Stanley Durrleman, Stéphanie Allassonnière
In this paper, we introduce a theoretical framework, with state-of-the-art convergence guarantees, for any deterministic approximation of the E step.
no code implementations • 11 Mar 2020 • Thomas Lartigue, Simona Bottani, Stephanie Baron, Olivier Colliot, Stanley Durrleman, Stéphanie Allassonnière
We demonstrate on synthetic data that, when the sample size is small, the two methods produce graphs with either too few or too many edges when compared to the real one.