Adversarial Defense

179 papers with code • 10 benchmarks • 5 datasets

Competitions with currently unpublished results:

Libraries

Use these libraries to find Adversarial Defense models and implementations

Most implemented papers

Learnable Boundary Guided Adversarial Training

fra31/auto-attack ICCV 2021

Previous adversarial training raises model robustness under the compromise of accuracy on natural data.

Safety Verification of Deep Neural Networks

VeriDeep/DLV 21 Oct 2016

Our method works directly with the network code and, in contrast to existing methods, can guarantee that adversarial examples, if they exist, are found for the given region and family of manipulations.

Delving into Transferable Adversarial Examples and Black-box Attacks

sunblaze-ucb/transferability-advdnn-pub 8 Nov 2016

In this work, we are the first to conduct an extensive study of the transferability over large models and a large scale dataset, and we are also the first to study the transferability of targeted adversarial examples with their target labels.

Mitigating Adversarial Effects Through Randomization

cihangxie/NIPS2017_adv_challenge_defense ICLR 2018

Convolutional neural networks have demonstrated high accuracy on various tasks in recent years.

Defense against Adversarial Attacks Using High-Level Representation Guided Denoiser

lfz/Guided-Denoise CVPR 2018

First, with HGD as a defense, the target model is more robust to either white-box or black-box adversarial attacks.

Benchmarking Neural Network Robustness to Common Corruptions and Surface Variations

bethgelab/imagecorruptions ICLR 2019

Then we propose a new dataset called Icons-50 which opens research on a new kind of robustness, surface variation robustness.

Efficient Formal Safety Analysis of Neural Networks

tcwangshiqi-columbia/ReluVal NeurIPS 2018

Our approach can check different safety properties and find concrete counterexamples for networks that are 10$\times$ larger than the ones supported by existing analysis techniques.

Feature Denoising for Improving Adversarial Robustness

facebookresearch/ImageNet-Adversarial-Training CVPR 2019

This study suggests that adversarial perturbations on images lead to noise in the features constructed by these networks.

Wasserstein Adversarial Examples via Projected Sinkhorn Iterations

locuslab/projected_sinkhorn 21 Feb 2019

In this paper, we propose a new threat model for adversarial attacks based on the Wasserstein distance.

Adversarial Examples on Graph Data: Deep Insights into Attack and Defense

stellargraph/stellargraph 5 Mar 2019

Based on this observation, we propose a defense approach which inspects the graph and recovers the potential adversarial perturbations.