Skip to content
This repository has been archived by the owner on Aug 28, 2021. It is now read-only.

Commit

Permalink
Highlight more that we consider targeted attack.
Browse files Browse the repository at this point in the history
  • Loading branch information
ppwwyyxx committed Feb 11, 2019
1 parent f80f5e7 commit ae5c586
Show file tree
Hide file tree
Showing 2 changed files with 5 additions and 4 deletions.
7 changes: 4 additions & 3 deletions INSTRUCTIONS.md
Expand Up @@ -65,10 +65,11 @@ Note:

1. As mentioned in the paper, the threat model is:

1. targeted PGD attack with one random uniform target label associated with each image
2. maximum perturbation per pixel is 16.
1. __Targeted attack__, with one target label associated with each image. The target lable is
independently generated by uniformly sampling the incorrect labels.
2. Maximum perturbation per pixel is 16.

We do not consider untargeted attack, nor do we let the attacker control the target label,
We do not consider untargeted attacks, nor do we let the attacker control the target labels,
because we think such tasks are not realistic on the ImageNet-1k categories.

2. For each (attacker, model) pair, we provide both the __error rate__ of our model,
Expand Down
2 changes: 1 addition & 1 deletion README.md
Expand Up @@ -13,7 +13,7 @@ By combining large-scale adversarial training and feature-denoising layers,
we developed ImageNet classifiers with strong adversarial robustness.

Trained on __128 GPUs__, our ImageNet classifier has 42.6% accuracy against an extremely strong
__2000-steps white-box__ PGD attacker.
__2000-steps white-box__ PGD targeted attack.
This is a scenario where no previous models have achieved more than 1% accuracy.

On black-box adversarial defense, our method won the __champion of defense track__ in the
Expand Down

0 comments on commit ae5c586

Please sign in to comment.