Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

why the adversarial perturbations was damaged by saving the adversarial samples with scipy.misc. #3

Closed
shenqixiaojiang opened this issue Jul 26, 2017 · 3 comments

Comments

@shenqixiaojiang
Copy link

shenqixiaojiang commented Jul 26, 2017

Hello, @carlini . Sorry to bother you again. I'm try to defense your attack. But there is a strange thing happened: the adversarial perturbations was damaged by saving the adversarial samples with scipy.misc.
And the image is dealed with ' image = image / 255.0 - 0.5 ' when as the input of pre-trained model
and We find the final output is different with the start.

@shenqixiaojiang shenqixiaojiang changed the title why the adversariay why the adversarial perturbations was damaged by save the adversarial samples with scipy.misc. Jul 26, 2017
@shenqixiaojiang shenqixiaojiang changed the title why the adversarial perturbations was damaged by save the adversarial samples with scipy.misc. why the adversarial perturbations was damaged by saving the adversarial samples with scipy.misc. Jul 26, 2017
@carlini
Copy link
Owner

carlini commented Jul 26, 2017

Discretizing the values from a real-numbered value to one of the 256 points degrades the quality of the adversarial examples. This can easily be fixed by performing a second optimization step on the lattice of discretized images (often only a few iterations is necessary). However, if you don't want to have to do this, you can also just save and load it as float32.

@carlini carlini closed this as completed Jul 26, 2017
@shenqixiaojiang
Copy link
Author

shenqixiaojiang commented Jul 26, 2017

@carlini
Yeah, now the adversarial samples were saved with 'npy' format.
In addition, the no-targeted adversarial samples of MNIST datasets was used to attack the model used in the cleverhans library.
And the accuracy was 0.9 which means the 90% of adversarial samples was failed to attack.
So do you think it is normal and Can I defense your attack with accuracy on the test model of cleverhans library?

Actually, the model used in your 'setup_mnist.py' file was trained based on cleverhans library.
And the result was agree with the above.

@carlini
Copy link
Owner

carlini commented Aug 3, 2017

Sorry, I'm not sure what you're trying to say. Are you attacking and defending using the same model? If they are different, you will need to generate transferable adversarial examples, to do this set the confidence to 3 or 4 on MNIST.

I'm not sure what this has to do with cleverhans.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants