Detection by Attack: Detecting Adversarial Samples by Undercover Attack
-
Updated
Feb 13, 2021 - Python
Detection by Attack: Detecting Adversarial Samples by Undercover Attack
Code for "BayesAdapter: Being Bayesian, Inexpensively and Robustly, via Bayeisan Fine-tuning"
Adversarial Detection in ROS Gazebo.
Gaussian process regression-based adversarial image detection
Using Gaussian Processes for Deep Neural Network Predictive Uncertainty Estimation
This work demonstrates an altogether different utility of attention heads. Self-attention heads are characteristic of Transformer models and have been well studied for interpretability and pruning, but here we build a novel adversarial detection model based on them.
This work demonstrates an altogether different utility of attention heads. Self-attention heads are characteristic of Transformer models and have been well studied for interpretability and pruning, but here we build a novel adversarial detection model based on them.
Add a description, image, and links to the adversarial-detection topic page so that developers can more easily learn about it.
To associate your repository with the adversarial-detection topic, visit your repo's landing page and select "manage topics."