A Novel Out-of-Distribution Detection Approach for Spiking Neural Networks: Design, Fusion, Performance Evaluation and Explainability
Repository containing the code and results for the paper A Novel Out-of-Distribution Detection Approach for Spiking Neural Networks: Design, Fusion, Performance Evaluation and Explainability. The paper is currently being evaluated is Information Fusion, but a preprint is available at https://arxiv.org/abs/2210.00894. Obtained results are available in the results folder.
- PyTorch 1.12.1 (vision 0.13.1) [https://github.com/pytorch/pytorch]
- Norse 0.0.7 [https://github.com/norse/norse]
- Other libraries typically included in mainstream Python distributions (NumPy, Pickle, Matplotlib, Pandas, scikit-learn, SciPy etc). There is a requirements.txt with the concrete libraries and versions.
To correctly install Norse locally, first install desired PyTorch version and then install Norse following the instructions at the norse instalation guide and troubleshooting
For the bayesian test and the CD graph, the following resources are used:
- Orange3, version 3.30.0 [https://github.com/biolab/orange3]
- BayesianTestsML repository's bayesiantest.py
The majority of datasets is downloaded automatically, except for:
For reproducing results of black and white datasets and with one FC hidden layer models, run:
python main.py --conf bw --pretrained --arch-selector 1 --f-max 100 --n-time-steps 50
For the case of color datasets
python main.py --conf rgb --pretrained --arch-selector 11 --f-max 200 --n-time-steps 64
For the case of neuromorphic datasets
python main.py --conf rgb --pretrained --arch-selector 11 --f-max 200 --n-time-steps 64 --neuromorphic
For training:
CUDA_VISIBLE_DEVICES=0 python train.py --conf datasets --dataset CIFAR10 --model ConvNet --arch-selector 10 --f-max 200 -b 128 --lr 0.002 --epochs 75 --weight-decay 0.00001 --opt adamw --n-time-steps 64 --save-every 10 --lr-decay-milestones 10 20 25 30 --lr-decay-rate 0.8 --penultimate-layer-neurons 100
The name of the dataset must be the same as in the configuration file
For training neuromorphic
CUDA_VISIBLE_DEVICES=2 python train.py --conf datasets --dataset DVSGesture --model ConvNet --arch-selector 11 --f-max 200 -b 8 --lr 0.0005 --epochs 20 --weight-decay 0.00001 --opt adamw --n-time-steps 64 --save-every 1 --lr-decay-milestones 10 --lr-decay-rate 0.5 --penultimate-layer-neurons 100 --encoder neuromorphic
For extracting metric plots:
python .\benchmark_metrics.py -c bw -r 0.01 -m all-subdirectories -d "path/to/dir" -cd -bs --metric AUROC --models ConvNet
There is also available an older version of the code in the jupyter notebook called Explainable_OoD_detection_on_SNNs, where instructions about how the algorithm works and how to execute it are provided.
If you find a bug, create a GitHub issue, or even better, submit a pull request. Similarly, if you have questions, simply post them as GitHub issues.
Martinez-Seras, A., Del Ser, J., Lobo, J. L., Garcia-Bringas, P., & Kasabov, N. (2023). A novel out-of-distribution detection approach for spiking neural networks: design, fusion, performance evaluation and explainability. Information Fusion, 100, 101943.
@article{martinez2023novel,
title={A novel out-of-distribution detection approach for spiking neural networks: design, fusion, performance evaluation and explainability},
author={Martinez-Seras, Aitor and Del Ser, Javier and Lobo, Jesus L and Garcia-Bringas, Pablo and Kasabov, Nikola},
journal={Information Fusion},
volume={100},
pages={101943},
year={2023},
publisher={Elsevier}
}
Research around Spiking Neural Networks has ignited during the last years due to their advantages when compared to traditional neural networks, including their efficient processing and inherent ability to model complex temporal dynamics. Despite these differences, Spiking Neural Networks face similar issues than other neural computation counterparts when deployed in real-world settings. This work addresses one of the practical circumstances that can hinder the trustworthiness of this family of models: the possibility of querying a trained model with samples far from the distribution of its training data (also referred to as Out-of-Distribution or OoD data). Specifically, this work presents a novel OoD detector that can identify whether test examples input to a Spiking Neural Network belong to the distribution of the data over which it was trained. For this purpose, we characterize the internal activations of the hidden layers of the network in the form of spike count patterns, which lay a basis for determining when the activations induced by a test instance is atypical. Furthermore, a local explanation method is devised to produce attribution maps revealing which parts of the input instance push most towards the detection of an example as an OoD sample. Experimental results are performed over several image classification datasets to compare the proposed detector to other OoD detection schemes from the literature, and to assess whether its fusion with other baseline OoD detection schemes can complement and boost the overall detection capability. As the obtained results clearly show, the proposed detector performs competitively against such alternative schemes, and when fused together, can significantly improve the detection scores of their constituent individual detectors. Furthermore, the explainability technique associated to our proposal is proven to produce relevance attribution maps that conform to expectations for synthetically created OoD instances.