1 code implementation • 25 Aug 2021 • Joel Q. L. Chang, Vincent Y. F. Tan
This paper unifies the design and the analysis of risk-averse Thompson sampling algorithms for the multi-armed bandit problem for a class of risk functionals $\rho$ that are continuous and dominant.
no code implementations • 14 May 2021 • Ming Liang Ang, Eloise Y. Y. Lim, Joel Q. L. Chang
The multi-armed bandit (MAB) problem is a ubiquitous decision-making problem that exemplifies exploration-exploitation tradeoff.
no code implementations • 16 Nov 2020 • Joel Q. L. Chang, Qiuyu Zhu, Vincent Y. F. Tan
The multi-armed bandit (MAB) problem is a ubiquitous decision-making problem that exemplifies the exploration-exploitation tradeoff.