Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Make ImageNet-C module pip-installable #2

Closed
nottombrown opened this issue Sep 20, 2018 · 8 comments
Closed

Make ImageNet-C module pip-installable #2

nottombrown opened this issue Sep 20, 2018 · 8 comments

Comments

@nottombrown
Copy link
Contributor

We're interested in using the ImageNet-C transformations as a baseline attack in the Unrestricted Adversarial Examples Challenge: openphilanthropy/unrestricted-adversarial-examples#40

It would be easier for us to do this if the module included a setup.py and was available on pypi

@hendrycks
Copy link
Owner

I will look into this in a week (after ICLR's submission deadline).

@nottombrown
Copy link
Contributor Author

Thanks! Good luck with the the submission

@hendrycks
Copy link
Owner

hendrycks commented Oct 3, 2018

This package is pip installable.

I have a question about the warm-up challenge. Imagine we pre-specify large-scale content that we would like a network to detect. So consider a network that classifies 1 if the large-scale input contains the content of the pre-specified image, and 0 if the input contains completely different content. The network must only be robust over one image's content, rather than be robust across a broad distribution of images such as birds and bikes. Do you know if current approaches can reliably defend against unrestricted attacks (or stronger attacks such as PGD \ell_\infty with \epsilon=16/255) for a single pre-specified large-scale image? The reason I ask is because some papers suggest adversarial robustness requires the number of training samples to be exponential in the image dimension.

@nottombrown
Copy link
Contributor Author

nottombrown commented Oct 3, 2018 via email

@jmgilmer
Copy link

jmgilmer commented Oct 4, 2018

Adversarial spheres wasn't really intended to be a difficult dataset. So it's not clear to me whether or not we can draw conclusions about perfect models for the sphere dataset to perfect models for image data.

The "semantic ball" around a single image is going to be a very complex object, probably much more so than a sphere. There it's plausible to me that we might need an exponential amount of data (haven't had the time yet to read the paper Dan is referring to).

Regarding the success of adversarial training, https://arxiv.org/abs/1804.11285 found that adversarial training on CIFAR-10 for long enough with a large enough network would yield a model that was 100% robust on the training set, but that this robustness did not generalize to the test set (they were only 50% robust at l_infinity eps=8/255 and 0% robust in larger threat models). So I don't think a solution is as simple as just adversarially train longer with a larger network.

@nottombrown
Copy link
Contributor Author

nottombrown commented Oct 4, 2018 via email

@jmgilmer
Copy link

jmgilmer commented Oct 4, 2018

If the goal is l_p robustness to images in the training set then adversarial training long enough could get you there (or you could alternatively use a NN classifier). But if you want significant l_p robustness to new images in the test set, that has yet to be demonstrated on CIFAR-10 for even modest sized epsilon.

So I guess my agreement depends on the question :).

@nottombrown
Copy link
Contributor Author

Cool, I agree with that.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants