Search Results for author: Ruiquan Huang

Found 10 papers, 0 papers with code

Provable Benefits of Multi-task RL under Non-Markovian Decision Making Processes

no code implementations20 Oct 2023 Ruiquan Huang, Yuan Cheng, Jing Yang, Vincent Tan, Yingbin Liang

To this end, we posit a joint model class for tasks and use the notion of $\eta$-bracketing number to quantify its complexity; this number also serves as a general metric to capture the similarity of tasks and thus determines the benefit of multi-task over single-task RL.

Decision Making Multi-Task Learning +1

Temporal-Distributed Backdoor Attack Against Video Based Action Recognition

no code implementations21 Aug 2023 Xi Li, Songhe Wang, Ruiquan Huang, Mahanth Gowda, George Kesidis

Although there are extensive studies on backdoor attacks against image data, the susceptibility of video-based systems under backdoor attacks remains largely unexplored.

Action Recognition Backdoor Attack +3

Provably Efficient UCB-type Algorithms For Learning Predictive State Representations

no code implementations1 Jul 2023 Ruiquan Huang, Yingbin Liang, Jing Yang

The general sequential decision-making problem, which includes Markov decision processes (MDPs) and partially observable MDPs (POMDPs) as special cases, aims at maximizing a cumulative reward by making a sequence of decisions based on a history of observations and actions over time.

Computational Efficiency Decision Making

Differentially Private Wireless Federated Learning Using Orthogonal Sequences

no code implementations14 Jun 2023 Xizixiang Wei, Tianhao Wang, Ruiquan Huang, Cong Shen, Jing Yang, H. Vincent Poor

A new FL convergence bound is derived which, combined with the privacy guarantees, allows for a smooth tradeoff between the achieved convergence rate and differential privacy levels.

Federated Learning Privacy Preserving

Near-optimal Conservative Exploration in Reinforcement Learning under Episode-wise Constraints

no code implementations9 Jun 2023 Donghao Li, Ruiquan Huang, Cong Shen, Jing Yang

This paper investigates conservative exploration in reinforcement learning where the performance of the learning agent is guaranteed to be above a certain threshold throughout the learning process.

reinforcement-learning

Non-stationary Reinforcement Learning under General Function Approximation

no code implementations1 Jun 2023 Songtao Feng, Ming Yin, Ruiquan Huang, Yu-Xiang Wang, Jing Yang, Yingbin Liang

To the best of our knowledge, this is the first dynamic regret analysis in non-stationary MDPs with general function approximation.

reinforcement-learning Reinforcement Learning (RL)

Improved Sample Complexity for Reward-free Reinforcement Learning under Low-rank MDPs

no code implementations20 Mar 2023 Yuan Cheng, Ruiquan Huang, Jing Yang, Yingbin Liang

In this work, we first provide the first known sample complexity lower bound that holds for any algorithm under low-rank MDPs.

reinforcement-learning Reinforcement Learning (RL) +1

Safe Exploration Incurs Nearly No Additional Sample Complexity for Reward-free RL

no code implementations28 Jun 2022 Ruiquan Huang, Jing Yang, Yingbin Liang

In particular, we consider the scenario where a safe baseline policy is known beforehand, and propose a unified Safe reWard-frEe ExploraTion (SWEET) framework.

Safe Exploration

Federated Linear Contextual Bandits

no code implementations NeurIPS 2021 Ruiquan Huang, Weiqiang Wu, Jing Yang, Cong Shen

This paper presents a novel federated linear contextual bandits model, where individual clients face different $K$-armed stochastic bandits coupled through common global parameters.

Multi-Armed Bandits

Cannot find the paper you are looking for? You can Submit a new open access paper.