Data Poisoning

124 papers with code • 0 benchmarks • 0 datasets

Data Poisoning is an adversarial attack that tries to manipulate the training dataset in order to control the prediction behavior of a trained model such that the model will label malicious examples into a desired classes (e.g., labeling spam e-mails as safe).

Source: Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Libraries

Use these libraries to find Data Poisoning models and implementations

Universal Backdoor Attacks

ben-schneider-code/universal-backdoor-attacks 30 Nov 2023

We demonstrate the effectiveness and robustness of our universal backdoor attacks by controlling models with up to 6, 000 classes while poisoning only 0. 15% of the training dataset.

0
30 Nov 2023

Transferable Availability Poisoning Attacks

trustmlrg/transpoison 8 Oct 2023

We consider availability data poisoning attacks, where an adversary aims to degrade the overall test accuracy of a machine learning model by crafting small perturbations to its training data.

2
08 Oct 2023

Seeing Is Not Always Believing: Invisible Collision Attack and Defence on Pre-Trained Models

anonymous10240/framework 24 Sep 2023

The typical paradigm is to pre-train a big deep learning model on large-scale data sets, and then fine-tune the model on small task-specific data sets for downstream tasks.

0
24 Sep 2023

HINT: Healthy Influential-Noise based Training to Defend against Data Poisoning Attacks

minhhao97vn/hint 15 Sep 2023

While numerous defense methods have been proposed to prohibit potential poisoning attacks from untrusted data sources, most research works only defend against specific attacks, which leaves many avenues for an adversary to exploit.

1
15 Sep 2023

Vulnerabilities in AI Code Generators: Exploring Targeted Data Poisoning Attacks

dessertlab/targeted-data-poisoning-attacks 4 Aug 2023

To address this threat, this work investigates the security of AI code generators by devising a targeted data poisoning strategy.

1
04 Aug 2023

FedDefender: Backdoor Attack Defense in Federated Learning

warisgill/FedDefender 2 Jul 2023

Federated Learning (FL) is a privacy-preserving distributed machine learning technique that enables individual clients (e. g., user participants, edge devices, or organizations) to train a model on their local data in a secure environment and then share the trained model with an aggregator to build a global model collaboratively.

7
02 Jul 2023

On the Exploitability of Instruction Tuning

azshue/autopoison NeurIPS 2023

In this work, we investigate how an adversary can exploit instruction tuning by injecting specific instruction-following examples into the training data that intentionally changes the model's behavior.

52
28 Jun 2023

DeepfakeArt Challenge: A Benchmark Dataset for Generative AI Art Forgery and Data Poisoning Detection

h-aboutalebi/deepfakeart 2 Jun 2023

Motivated to address these key concerns to encourage responsible generative AI, we introduce the DeepfakeArt Challenge, a large-scale challenge benchmark dataset designed specifically to aid in the building of machine learning algorithms for generative AI art forgery and data poisoning detection.

8
02 Jun 2023

From Shortcuts to Triggers: Backdoor Defense with Denoised PoE

luka-group/dpoe 24 May 2023

Language models are often at risk of diverse backdoor attacks, especially data poisoning.

2
24 May 2023

Differentially-Private Decision Trees and Provable Robustness to Data Poisoning

tudelft-cda-lab/privatree 24 May 2023

By leveraging the better privacy-utility trade-off of PrivaTree we are able to train decision trees with significantly better robustness against backdoor attacks compared to regular decision trees and with meaningful theoretical guarantees.

2
24 May 2023