This repo provides the official implementations for the experiments described in the following paper:
Adversarial Attack on Graph Neural Networks as An Influence Maximization Problem
Jiaqi Ma*, Junwei Deng*, and Qiaozhu Mei. WSDM 2022.
(*: equal constribution)
- dgl 0.4.3
- torch 1.4.0
- networkx 2.3
- numpy 1.16.4
This repo is built on top of the code for a NeurIPS 2020 publication: Towards More Practical Adversarial Attacks on Graph Neural Networks.
A major change in the experiment setup between this work and the NeurIPS 2020 publication is the way we construct the node attribute perturbation, as noted in Section 5.2 of our paper. The perturbation construction implemented in this repo is more strictly black-box.
Example command to run the code: python main.py --seed 0 --model GAT --dataset cora --threshold 0.1 --norm_length 10
.
@article{ma2021adversarial,
title={Adversarial Attack on Graph Neural Networks as An Influence Maximization Problem},
author={Ma, Jiaqi and Deng, Junwei and Mei, Qiaozhu},
journal={arXiv preprint arXiv:2106.10785},
year={2021}
}