Search Results for author: Ruoxi Chen

Found 11 papers, 6 papers with code

MLLM-as-a-Judge: Assessing Multimodal LLM-as-a-Judge with Vision-Language Benchmark

1 code implementation7 Feb 2024 Dongping Chen, Ruoxi Chen, Shilin Zhang, Yinuo Liu, Yaochen Wang, Huichi Zhou, Qihui Zhang, Pan Zhou, Yao Wan, Lichao Sun

Multimodal Large Language Models (MLLMs) have gained significant attention recently, showing remarkable potential in artificial general intelligence.

GUARD: Role-playing to Generate Natural-language Jailbreakings to Test Guideline Adherence of Large Language Models

no code implementations5 Feb 2024 Haibo Jin, Ruoxi Chen, Andy Zhou, Jinyin Chen, Yang Zhang, Haohan Wang

Our system of different roles will leverage this knowledge graph to generate new jailbreaks, which have proved effective in inducing LLMs to generate unethical or guideline-violating responses.

Sentence

AdvCheck: Characterizing Adversarial Examples via Local Gradient Checking

no code implementations25 Mar 2023 Ruoxi Chen, Haibo Jin, Jinyin Chen, Haibin Zheng

To address the issues, we introduce the concept of local gradient, and reveal that adversarial examples have a quite larger bound of local gradient than the benign ones.

Is Multi-Modal Necessarily Better? Robustness Evaluation of Multi-modal Fake News Detection

no code implementations17 Jun 2022 Jinyin Chen, Chengyu Jia, Haibin Zheng, Ruoxi Chen, Chenbo Fu

The proliferation of fake news and its serious negative social influence push fake news detection methods to become necessary tools for web managers.

Backdoor Attack Fake News Detection

Unleashing the Potential of Adjacent Snippets for Weakly-supervised Temporal Action Localization

1 code implementation1 May 2022 Qinying Liu, Zilei Wang, Ruoxi Chen, Zhilin Li

C$^3$BN consists of two key ingredients: a micro data augmentation strategy that increases the diversity in-between adjacent snippets by convex combination of adjacent snippets, and a macro-micro consistency regularization that enforces the model to be invariant to the transformations~\textit{w. r. t.}

Data Augmentation Video Classification +3

Excitement Surfeited Turns to Errors: Deep Learning Testing Framework Based on Excitable Neurons

1 code implementation12 Feb 2022 Haibo Jin, Ruoxi Chen, Haibin Zheng, Jinyin Chen, Yao Cheng, Yue Yu, Xianglong Liu

By maximizing the number of excitable neurons concerning various wrong behaviors of models, DeepSensor can generate testing examples that effectively trigger more errors due to adversarial inputs, polluted data and incomplete training.

Image Classification Speaker Recognition

NIP: Neuron-level Inverse Perturbation Against Adversarial Attacks

no code implementations24 Dec 2021 Ruoxi Chen, Haibo Jin, Jinyin Chen, Haibin Zheng, Yue Yu, Shouling Ji

From the perspective of image feature space, some of them cannot reach satisfying results due to the shift of features.

CatchBackdoor: Backdoor Testing by Critical Trojan Neural Path Identification via Differential Fuzzing

no code implementations24 Dec 2021 Haibo Jin, Ruoxi Chen, Jinyin Chen, Yao Cheng, Chong Fu, Ting Wang, Yue Yu, Zhaoyan Ming

Existing DNN testing methods are mainly designed to find incorrect corner case behaviors in adversarial settings but fail to discover the backdoors crafted by strong trojan attacks.

DNN Testing

Salient Feature Extractor for Adversarial Defense on Deep Neural Networks

1 code implementation14 May 2021 Jinyin Chen, Ruoxi Chen, Haibin Zheng, Zhaoyan Ming, Wenrong Jiang, Chen Cui

Motivated by the observation that adversarial examples are due to the non-robust feature learned from the original dataset by models, we propose the concepts of salient feature(SF) and trivial feature(TF).

Adversarial Defense Generative Adversarial Network

Cannot find the paper you are looking for? You can Submit a new open access paper.