Data Poisoning

123 papers with code • 0 benchmarks • 0 datasets

Data Poisoning is an adversarial attack that tries to manipulate the training dataset in order to control the prediction behavior of a trained model such that the model will label malicious examples into a desired classes (e.g., labeling spam e-mails as safe).

Source: Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Libraries

Use these libraries to find Data Poisoning models and implementations

Federated Learning Under Attack: Exposing Vulnerabilities through Data Poisoning Attacks in Computer Networks

ehsannowroozi/federatedlearning_poison_lf_fp 5 Mar 2024

In LF, we randomly flipped the labels of benign data and trained the model on the manipulated data.

1
05 Mar 2024

Learning to Poison Large Language Models During Instruction Tuning

rookiezxy/gbtl 21 Feb 2024

The advent of Large Language Models (LLMs) has marked significant achievements in language processing and reasoning capabilities.

2
21 Feb 2024

Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based Agents

lancopku/agent-backdoor-attacks 17 Feb 2024

We first formulate a general framework of agent backdoor attacks, then we present a thorough analysis on the different forms of agent backdoor attacks.

14
17 Feb 2024

The Effect of Data Poisoning on Counterfactual Explanations

andreartelt/datapoisoningcounterfactuals 13 Feb 2024

Counterfactual explanations provide a popular method for analyzing the predictions of black-box systems, and they can offer the opportunity for computational recourse by suggesting actionable changes on how to change the input to obtain a different (i. e. more favorable) system output.

0
13 Feb 2024

Shadowcast: Stealthy Data Poisoning Attacks Against Vision-Language Models

umd-huang-lab/vlm-poisoning 5 Feb 2024

We show that Shadowcast are highly effective in achieving attacker's intentions using as few as 50 poison samples.

15
05 Feb 2024

Game-Theoretic Unlearnable Example Generator

hong-xian/gue 31 Jan 2024

Unlearnable example attacks are data poisoning attacks aiming to degrade the clean test accuracy of deep learning by adding imperceptible perturbations to the training samples, which can be formulated as a bi-level optimization problem.

2
31 Jan 2024

Progressive Poisoned Data Isolation for Training-time Backdoor Defense

rorschachchen/pipd 20 Dec 2023

Extensive experiments on multiple benchmark datasets and DNN models, assessed against nine state-of-the-art backdoor attacks, demonstrate the superior performance of our PIPD method for backdoor defense.

3
20 Dec 2023

FlowMur: A Stealthy and Practical Audio Backdoor Attack with Limited Knowledge

cristinalan/flowmur 15 Dec 2023

Despite the initial success of current audio backdoor attacks, they suffer from the following limitations: (i) Most of them require sufficient knowledge, which limits their widespread adoption.

4
15 Dec 2023

IMMA: Immunizing text-to-image Models against Malicious Adaptation

amberyzheng/imma 30 Nov 2023

Advancements in text-to-image models and fine-tuning methods have led to the increasing risk of malicious adaptation, i. e., fine-tuning to generate harmful unauthorized content.

11
30 Nov 2023

Universal Backdoor Attacks

ben-schneider-code/universal-backdoor-attacks 30 Nov 2023

We demonstrate the effectiveness and robustness of our universal backdoor attacks by controlling models with up to 6, 000 classes while poisoning only 0. 15% of the training dataset.

0
30 Nov 2023