no code implementations • 5 Sep 2023 • Eric Bax, Frédéric Ouimet
In this paper, we develop a non-asymptotic local normal approximation for multinomial probabilities.
no code implementations • 14 Aug 2022 • Eric Bax, John Donald
A data sketch algorithm scans a big data set, collecting a small amount of data -- the sketch, which can be used to statistically infer properties of the big data set.
no code implementations • 23 Apr 2021 • Eric Bax
For a voting ensemble that selects an odd-sized subset of the ensemble classifiers at random for each example, applies them to the example, and returns the majority vote, we show that any number of voters may minimize the error rate over an out-of-sample distribution.
no code implementations • 4 Oct 2016 • Eric Bax, Farshad Kooti
Suppose some classifiers are selected from a set of hypothesis classifiers to form an equally-weighted ensemble that selects a member classifier at random for each input example.
no code implementations • 9 Oct 2015 • Eric Bax, Ya Le
WAG withholds a validation set, trains a holdout classifier on the remaining data, uses the validation data to validate that classifier, then adds the rate of disagreement between the holdout classifier and one trained using all in-sample data, which is an upper bound on the difference in error rates.
no code implementations • 31 Mar 2015 • Eric Bax
Error bounds based on worst likely assignments use permutation tests to validate classifiers.
no code implementations • 31 Oct 2014 • Ya Le, Eric Bax, Nicola Barbieri, David Garcia Soriano, Jitesh Mehta, James Li
We introduce a technique to compute probably approximately correct (PAC) bounds on precision and recall for matching algorithms.
no code implementations • 9 Oct 2014 • Eric Bax, Lingjie Weng, Xu Tian
We introduce the speculate-correct method to derive error bounds for local classifiers.