Search Results for author: Qi Yi

Found 13 papers, 6 papers with code

Introducing Foundation Models as Surrogate Models: Advancing Towards More Practical Adversarial Attacks

no code implementations13 Jul 2023 Jiaming Zhang, Jitao Sang, Qi Yi, Changsheng Xu

Harnessing the concept of non-robust features, we elaborate on two guiding principles for surrogate model selection to explain why the foundational model is an optimal choice for this role.

Adversarial Attack Attribute +1

Online Prototype Alignment for Few-shot Policy Transfer

1 code implementation12 Jun 2023 Qi Yi, Rui Zhang, Shaohui Peng, Jiaming Guo, Yunkai Gao, Kaizhao Yuan, Ruizhi Chen, Siming Lan, Xing Hu, Zidong Du, Xishan Zhang, Qi Guo, Yunji Chen

Domain adaptation in reinforcement learning (RL) mainly deals with the changes of observation when transferring the policy to a new environment.

Domain Adaptation Reinforcement Learning (RL)

Conceptual Reinforcement Learning for Language-Conditioned Tasks

no code implementations9 Mar 2023 Shaohui Peng, Xing Hu, Rui Zhang, Jiaming Guo, Qi Yi, Ruizhi Chen, Zidong Du, Ling Li, Qi Guo, Yunji Chen

Recently, the language-conditioned policy is proposed to facilitate policy transfer through learning the joint representation of observation and text that catches the compact and invariant information across environments.

reinforcement-learning Reinforcement Learning (RL)

Unlearnable Clusters: Towards Label-agnostic Unlearnable Examples

1 code implementation CVPR 2023 Jiaming Zhang, Xingjun Ma, Qi Yi, Jitao Sang, Yu-Gang Jiang, YaoWei Wang, Changsheng Xu

Furthermore, we propose to leverage VisionandLanguage Pre-trained Models (VLPMs) like CLIP as the surrogate model to improve the transferability of the crafted UCs to diverse domains.

Data Poisoning

Causality-driven Hierarchical Structure Discovery for Reinforcement Learning

no code implementations13 Oct 2022 Shaohui Peng, Xing Hu, Rui Zhang, Ke Tang, Jiaming Guo, Qi Yi, Ruizhi Chen, Xishan Zhang, Zidong Du, Ling Li, Qi Guo, Yunji Chen

To address this issue, we propose CDHRL, a causality-driven hierarchical reinforcement learning framework, leveraging a causality-driven discovery instead of a randomness-driven exploration to effectively build high-quality hierarchical structures in complicated environments.

Hierarchical Reinforcement Learning reinforcement-learning +1

Object-Category Aware Reinforcement Learning

no code implementations13 Oct 2022 Qi Yi, Rui Zhang, Shaohui Peng, Jiaming Guo, Xing Hu, Zidong Du, Xishan Zhang, Qi Guo, Yunji Chen

Object-oriented reinforcement learning (OORL) is a promising way to improve the sample efficiency and generalization ability over standard RL.

Feature Engineering Object +3

Low-Mid Adversarial Perturbation against Unauthorized Face Recognition System

no code implementations19 Jun 2022 Jiaming Zhang, Qi Yi, Dongyuan Lu, Jitao Sang

In light of the growing concerns regarding the unauthorized use of facial recognition systems and its implications on individual privacy, the exploration of adversarial perturbations as a potential countermeasure has gained traction.

Face Recognition

Towards Adversarial Attack on Vision-Language Pre-training Models

1 code implementation19 Jun 2022 Jiaming Zhang, Qi Yi, Jitao Sang

While vision-language pre-training model (VLP) has shown revolutionary improvements on various vision-language (V+L) tasks, the studies regarding its adversarial robustness remain largely unexplored.

Adversarial Attack Adversarial Robustness

Hindsight Value Function for Variance Reduction in Stochastic Dynamic Environment

1 code implementation26 Jul 2021 Jiaming Guo, Rui Zhang, Xishan Zhang, Shaohui Peng, Qi Yi, Zidong Du, Xing Hu, Qi Guo, Yunji Chen

In this paper, we propose to replace the state value function with a novel hindsight value function, which leverages the information from the future to reduce the variance of the gradient estimate for stochastic dynamic environments.

Policy Gradient Methods

Cannot find the paper you are looking for? You can Submit a new open access paper.