Skip to content

LiuWenQingS/Defending-Adversarial-Examples-via-DNN-Bottleneck-Reinforcement

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Defending-Adversarial-Examples-via-DNN-Bottleneck-Reinforcement

Code of Defending Adversarial Examples via DNN Bottleneck Reinforcement.

If you have any questions, please send email to liuwenqing@tongji.edu.cn

1.data

1.1 dataset

We use three classical image dataset to train our model.

MNIST:http://yann.lecun.com/exdb/mnist/

Cifar10:https://www.cs.toronto.edu/~kriz/cifar.html

ImageNet:http://www.image-net.org/

1.2 adversarial examples

We use FGSM SGD to generate adversarial example for adversarial training.

During testing period, we use FGSM BIM C&W DNN to test the robustness of our model.

2.attack

To confuse the CNN, many method proposed to cheat add small perturbation on pixels of images such as FGSM.

During our work, we use FGSM SGD BIM C&W DNN to train our model or test our model.

3.defense

Think of the advantages of the Autoencoder, we try to combine the Classifier and Autoencoder to enable the ablility of

denoising of the source classifier.

About

Code of Defending Adversarial Examples via DNN Bottleneck Reinforcement

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages