no code implementations • 15 Apr 2024 • Ashna Jose, Emilie Devijver, Massih-Reza Amini, Noel Jakse, Roberta Poloni
A classification tree constructed on an initial set of labeled samples is considered to decompose the space into low-entropy regions.
1 code implementation • 9 Feb 2024 • Lei Zan, Charles K. Assaad, Emilie Devijver, Eric Gaussier
This paper introduces a new structural causal model tailored for representing threshold-based IT systems and presents a new algorithm designed to rapidly detect root causes of anomalies in such systems.
no code implementations • 23 Oct 2023 • Charles K. Assaad, Emilie Devijver, Eric Gaussier, Gregor Gössler, Anouar Meynaoui
We study the problem of identifiability of the total effect of an intervention from observational time series only given an abstraction of the causal graph of the system.
no code implementations • 2 Oct 2023 • Lies Hadjadj, Emilie Devijver, Remi Molinier, Massih-Reza Amini
Machine learning methods usually rely on large sample size to have good performance, while it is difficult to provide labeled set in many applications.
no code implementations • 28 Jul 2023 • Ali Aït-Bachir, Charles K. Assaad, Christophe de Bignicourt, Emilie Devijver, Simon Ferreira, Eric Gaussier, Hosein Mohanna, Lei Zan
Despite its potential benefits, applying causal discovery algorithms on IT monitoring data poses challenges, due to the complexity of the data.
1 code implementation • 14 Jun 2023 • Daria Bystrova, Charles K. Assaad, Julyan Arbel, Emilie Devijver, Eric Gaussier, Wilfried Thuiller
In the second class, a constraint-based strategy is applied to identify a skeleton, which is then oriented using a noise-based strategy.
no code implementations • 19 May 2022 • Charles K. Assaad, Emilie Devijver, Eric Gaussier
This study addresses the problem of learning an extended summary causal graph on time series.
no code implementations • 24 Feb 2022 • Massih-Reza Amini, Vasilii Feofanov, Loic Pauletto, Lies Hadjadj, Emilie Devijver, Yury Maximov
Semi-supervised algorithms aim to learn prediction functions from a small set of labeled observations and a large set of unlabeled observations.
no code implementations • 29 Sep 2021 • Vasilii Feofanov, Emilie Devijver, Massih-Reza Amini
First, we derive a transductive bound over the risk of the multi-class majority vote classifier.
no code implementations • 21 May 2021 • Charles K. Assaad, Emilie Devijver, Eric Gaussier
This study addresses the problem of learning a summary causal graph on time series with potentially different sampling rates.
no code implementations • NeurIPS 2020 • Sami Alkhoury, Emilie Devijver, Marianne Clausel, Myriam Tami, Eric Gaussier, Georges Oppenheim
We propose here a generalization of regression trees, referred to as Probabilistic Regression (PR) trees, that adapt to the smoothness of the prediction function relating input and output variables while preserving the interpretability of the prediction and being robust to noise.
no code implementations • 12 Nov 2019 • Vasilii Feofanov, Emilie Devijver, Massih-Reza Amini
In this paper, we propose a new wrapper feature selection approach with partially labeled training examples where unlabeled observations are pseudo-labeled using the predictions of an initial classifier trained on the labeled training set.
no code implementations • 27 Oct 2018 • Myriam Tami, Marianne Clausel, Emilie Devijver, Adrien Dulac, Eric Gaussier, Stefan Janaqi, Meriam Chebre
Tree-based ensemble methods, as Random Forests and Gradient Boosted Trees, have been successfully used for regression in many applications and research studies.
1 code implementation • 26 Jan 2017 • Emilie Devijver, Mélina Gallopin, Emeline Perthame
In this paper, we present a novel approach designed to predict quantitative trait from transcriptomic data, taking into account the heterogeneity in biological samples and the hidden gene regulatory networks underlying different biological mechanisms.
no code implementations • 12 Nov 2015 • Emilie Devijver, Mélina Gallopin
Based on the block-diagonal structure of the covariance matrix, the estimation problem is divided into several independent problems: subsequently, the network of dependencies between variables is inferred using the graphical lasso algorithm in each block.