no code implementations • 26 Sep 2023 • Daniil Ryabko, Angustias Vaca, Prudencio Pazoca
Moreover, after being established in the population, it can sustain costs of over 35\% .
no code implementations • 1 Oct 2020 • Daniil Ryabko
The main subject of this book are such mixture predictors, and the main results demonstrate the universality of this method in a very general probabilistic setting, but also show some of its limitations.
no code implementations • 26 Jun 2019 • Azadeh Khaleghi, Daniil Ryabko
The problem of time-series clustering is considered in the case where each data-point is a sample generated by a piecewise stationary ergodic process.
no code implementations • 30 Mar 2019 • Daniil Ryabko
It is thus a rather attractive assumption to base statistical analysis on, especially for problems for which less general qualitative assumptions, such as independence or finite memory, clearly fail.
no code implementations • 20 Dec 2018 • Daniil Ryabko
The problem of sequential probability forecasting is considered in the most general setting: a model set C is given, and it is required to predict as well as possible if any of the measures (environments) in C is chosen to generate the data.
no code implementations • 10 Aug 2017 • Daniil Ryabko
Drawing on some recent results that provide the formalism necessary to definite stationarity for infinite random graphs, this paper initiates the study of statistical and learning questions pertaining to these objects.
no code implementations • NeurIPS 2017 • Daniil Ryabko
The independence clustering problem is considered in the following formulation: given a set $S$ of random variables, it is required to find the finest partitioning $\{U_1,\dots, U_k\}$ of $S$ into clusters such that the clusters $U_1,\dots, U_k$ are mutually independent.
no code implementations • 26 Oct 2016 • Daniil Ryabko
In this work it is shown that the minimax asymptotic performance is always attainable, and it is attained by a convex combination of a countably many measures from the set C (a Bayesian mixture).
no code implementations • 26 Oct 2016 • Daniil Ryabko
In such a case, if there is a predictor that achieves asymptotically vanishing error for any measure in C, then there is a Bayesian predictor that also has this property, and whose prior is concentrated on (a countable subset of) C.
no code implementations • 9 Aug 2014 • Daniil Ryabko
A sequence x1,..., xn,... of discrete-valued observations is generated according to some unknown probabilistic law (measure) mu.
no code implementations • 12 May 2014 • Ronald Ortner, Odalric-Ambrym Maillard, Daniil Ryabko
We consider a reinforcement learning setting introduced in (Maillard et al., NIPS 2011) where the learner does not have explicit access to the states of the underlying Markov decision process (MDP).
no code implementations • 17 Apr 2013 • Daniil Ryabko
Numerous control and learning problems face the situation where sequences of high-dimensional highly dependent data are available but no or little feedback is provided to the learner, which makes any inference rather challenging.
no code implementations • NeurIPS 2012 • Ronald Ortner, Daniil Ryabko
We derive sublinear regret bounds for undiscounted reinforcement learning in continuous state space.
no code implementations • NeurIPS 2012 • Azadeh Khaleghi, Daniil Ryabko
The problem of multiple change point estimation is considered for sequences with unknown number of change points.
no code implementations • NeurIPS 2012 • Daniil Ryabko, Jérémie Mary
The algorithms that we construct for solving these problems are based on a new metric between time-series distributions, which can be evaluated using binary classification methods.
no code implementations • 7 Mar 2012 • Azadeh Khaleghi, Daniil Ryabko
Given a heterogeneous time-series sample, the objective is to find points in time (called change points) where the probability distribution generating the data has changed.
no code implementations • NeurIPS 2011 • Odalric-Ambrym Maillard, Daniil Ryabko, Rémi Munos
Without knowing neither which of the models is the correct one, nor what are the probabilistic characteristics of the resulting MDP, it is required to obtain as much reward as the optimal policy for the correct model (or for the best of the correct models, if there are several).
no code implementations • 31 May 2010 • Daniil Ryabko
For some of the formalizations we also show that when a solution exists, it can be obtained as a Bayes mixture over a countable subset of $\mathcal C$.
no code implementations • 5 May 2010 • Daniil Ryabko
We show that, for the case of a known number of clusters, consistency can be achieved under the only assumption that the joint distribution of the data is stationary ergodic (no parametric or Markovian assumptions, no assumptions of independence, neither between nor within the samples).