Search Results for author: Yoan Russac

Found 7 papers, 3 papers with code

Efficient Algorithms for Extreme Bandits

1 code implementation21 Mar 2022 Dorian Baudry, Yoan Russac, Emilie Kaufmann

In this paper, we contribute to the Extreme Bandit problem, a variant of Multi-Armed Bandits in which the learner seeks to collect the largest possible reward.

Multi-Armed Bandits

A/B/n Testing with Control in the Presence of Subpopulations

no code implementations NeurIPS 2021 Yoan Russac, Christina Katsimerou, Dennis Bohle, Olivier Cappé, Aurélien Garivier, Wouter Koolen

At every time step, a subpopulation is sampled and an arm is chosen: the resulting observation is an independent draw from the arm conditioned on the subpopulation.

On Limited-Memory Subsampling Strategies for Bandits

1 code implementation21 Jun 2021 Dorian Baudry, Yoan Russac, Olivier Cappé

There has been a recent surge of interest in nonparametric bandit algorithms based on subsampling.

Regret Bounds for Generalized Linear Bandits under Parameter Drift

no code implementations9 Mar 2021 Louis Faury, Yoan Russac, Marc Abeille, Clément Calauzènes

Generalized Linear Bandits (GLBs) are powerful extensions to the Linear Bandit (LB) setting, broadening the benefits of reward parametrization beyond linearity.

Self-Concordant Analysis of Generalized Linear Bandits with Forgetting

no code implementations2 Nov 2020 Yoan Russac, Louis Faury, Olivier Cappé, Aurélien Garivier

Contextual sequential decision problems with categorical or numerical observations are ubiquitous and Generalized Linear Bandits (GLB) offer a solid theoretical framework to address them.

Algorithms for Non-Stationary Generalized Linear Bandits

no code implementations23 Mar 2020 Yoan Russac, Olivier Cappé, Aurélien Garivier

The statistical framework of Generalized Linear Models (GLM) can be applied to sequential problems involving categorical or ordinal rewards associated, for instance, with clicks, likes or ratings.

Weighted Linear Bandits for Non-Stationary Environments

1 code implementation NeurIPS 2019 Yoan Russac, Claire Vernade, Olivier Cappé

To address this problem, we propose D-LinUCB, a novel optimistic algorithm based on discounted linear regression, where exponential weights are used to smoothly forget the past.

regression

Cannot find the paper you are looking for? You can Submit a new open access paper.