Adversarial Attack

596 papers with code • 2 benchmarks • 9 datasets

An Adversarial Attack is a technique to find a perturbation that changes the prediction of a machine learning model. The perturbation can be very small and imperceptible to human eyes.

Source: Recurrent Attention Model with Log-Polar Mapping is Robust against Adversarial Attacks

Libraries

Use these libraries to find Adversarial Attack models and implementations

Counterfactual Explanations for Face Forgery Detection via Adversarial Removal of Artifacts

faceonlive/ai-research 12 Apr 2024

We verify the effectiveness of the proposed explanations from two aspects: (1) Counterfactual Trace Visualization: the enhanced forgery images are useful to reveal artifacts by visually contrasting the original images and two different visualization methods; (2) Transferable Adversarial Attacks: the adversarial forgery images generated by attacking the detection model are able to mislead other detection models, implying the removed artifacts are general.

131
12 Apr 2024

READ: Improving Relation Extraction from an ADversarial Perspective

david-li0406/read 2 Apr 2024

This strategy enables a larger attack budget for entities and coaxes the model to leverage relational patterns embedded in the context.

2
02 Apr 2024

Humanizing Machine-Generated Content: Evading AI-Text Detection through Adversarial Attack

zhouying20/hmgc 2 Apr 2024

While well-trained text detectors have demonstrated promising performance on unseen test data, recent research suggests that these detectors have vulnerabilities when dealing with adversarial attacks such as paraphrasing.

2
02 Apr 2024

Physical 3D Adversarial Attacks against Monocular Depth Estimation in Autonomous Driving

gandolfczjh/3d2fool 26 Mar 2024

Deep learning-based monocular depth estimation (MDE), extensively applied in autonomous driving, is known to be vulnerable to adversarial attacks.

6
26 Mar 2024

$\textit{LinkPrompt}$: Natural and Universal Adversarial Attacks on Prompt-based Language Models

savannahxu79/linkprompt 25 Mar 2024

Prompt-based learning is a new language model training paradigm that adapts the Pre-trained Language Models (PLMs) to downstream tasks, which revitalizes the performance benchmarks across various natural language processing (NLP) tasks.

0
25 Mar 2024

Fast Inference of Removal-Based Node Influence

weikai-li/nora 13 Mar 2024

We propose a new method of evaluating node influence, which measures the prediction change of a trained GNN model caused by removing a node.

2
13 Mar 2024

Hard-label based Small Query Black-box Adversarial Attack

jpark04-qub/sqba 9 Mar 2024

We consider the hard label based black box adversarial attack setting which solely observes predicted classes from the target model.

3
09 Mar 2024

Hide in Thicket: Generating Imperceptible and Rational Adversarial Perturbations on 3D Point Clouds

trlou/hit-adv 8 Mar 2024

We find that concealing deformation perturbations in areas insensitive to human eyes can achieve a better trade-off between imperceptibility and adversarial strength, specifically in parts of the object surface that are complex and exhibit drastic curvature changes.

17
08 Mar 2024

One Prompt Word is Enough to Boost Adversarial Robustness for Pre-trained Vision-Language Models

treelli/apt 4 Mar 2024

This work studies the adversarial robustness of VLMs from the novel perspective of the text prompt instead of the extensively studied model weights (frozen in this work).

12
04 Mar 2024

Beyond Worst-case Attacks: Robust RL with Adaptive Defense via Non-dominated Policies

umd-huang-lab/protected 20 Feb 2024

In light of the burgeoning success of reinforcement learning (RL) in diverse real-world applications, considerable focus has been directed towards ensuring RL policies are robust to adversarial attacks during test time.

0
20 Feb 2024