Search Results for author: Wenqian Yu

Found 2 papers, 0 papers with code

Red Teaming GPT-4V: Are GPT-4V Safe Against Uni/Multi-Modal Jailbreak Attacks?

no code implementations4 Apr 2024 Shuo Chen, Zhen Han, Bailan He, Zifeng Ding, Wenqian Yu, Philip Torr, Volker Tresp, Jindong Gu

Various jailbreak attacks have been proposed to red-team Large Language Models (LLMs) and revealed the vulnerable safeguards of LLMs.

Reliable Evaluation of Adversarial Transferability

no code implementations14 Jun 2023 Wenqian Yu, Jindong Gu, Zhijiang Li, Philip Torr

Adversarial examples (AEs) with small adversarial perturbations can mislead deep neural networks (DNNs) into wrong predictions.

Cannot find the paper you are looking for? You can Submit a new open access paper.