Search Results for author: Rohan Deb

Found 6 papers, 0 papers with code

Think Before You Duel: Understanding Complexities of Preference Learning under Constrained Resources

no code implementations28 Dec 2023 Rohan Deb, Aadirupa Saha

We show that due to the relative nature of the feedback, the problem is more difficult than its bandit counterpart and that without further assumptions the problem is not learnable from a regret minimization perspective.

Contextual Bandits with Online Neural Regression

no code implementations12 Dec 2023 Rohan Deb, Yikun Ban, Shiliang Zuo, Jingrui He, Arindam Banerjee

Based on such a perturbed prediction, we show a ${\mathcal{O}}(\log T)$ regret for online regression with both squared loss and KL loss, and subsequently convert these respectively to $\tilde{\mathcal{O}}(\sqrt{KT})$ and $\tilde{\mathcal{O}}(\sqrt{KL^*} + K)$ regret for NeuCB, where $L^*$ is the loss of the best policy.

Multi-Armed Bandits regression

$N$-Timescale Stochastic Approximation: Stability and Convergence

no code implementations7 Dec 2021 Rohan Deb, Shalabh Bhatnagar

This paper presents the first sufficient conditions that guarantee the stability and almost sure convergence of $N$-timescale stochastic approximation (SA) iterates for any $N\geq1$.

Schedule Based Temporal Difference Algorithms

no code implementations23 Nov 2021 Rohan Deb, Meet Gandhi, Shalabh Bhatnagar

However, the weights assigned to different $n$-step returns in TD($\lambda$), controlled by the parameter $\lambda$, decrease exponentially with increasing $n$.

Gradient Temporal Difference with Momentum: Stability and Convergence

no code implementations22 Nov 2021 Rohan Deb, Shalabh Bhatnagar

Here, we consider Gradient TD algorithms with an additional heavy ball momentum term and provide choice of step size and momentum parameter that ensures almost sure convergence of these algorithms asymptotically.

Does Momentum Help? A Sample Complexity Analysis

no code implementations29 Oct 2021 Swetha Ganesh, Rohan Deb, Gugan Thoppe, Amarjit Budhiraja

Stochastic Heavy Ball (SHB) and Nesterov's Accelerated Stochastic Gradient (ASG) are popular momentum methods in stochastic optimization.

Stochastic Optimization

Cannot find the paper you are looking for? You can Submit a new open access paper.