A pytorch implementation of "Intriguing properties of neural networks"
- The natural basis is not better than a random basis for inspecting the properties of latent vectors.
- there are serveral directions which have the semantic meaning not only individual units.
- We can generate images(adversarial examples) with small perturbations to fooling neural network models.
- Weight Decay or Regularization couldn't help model to defend adversarial examples.
- One adversarial example for a specific model is possible to deceive other models.
- According to spectral analysis of unstability, the deeper models, the more stupid.
- python==3.6
- numpy==1.14.2
- pytorch==1.0.0
In the paper, L-BFGS is used to solve equation with constraints.
However, in this code, backpropagation method is used instead of L-BFGS.
Hence, it doesn't cover "4.2 Exprimental results"
- This Repository won't be updated.
- Please check the package of adversarial attacks in pytorch