Graph neural networks (GNNs) which apply the deep neural networks to graph data have achieved significant performance for the task of semi-supervised node classification. However, only few work has addressed the adversarial robustness of GNNs. In this paper, we first present a novel gradient-based attack method that facilitates the difficulty of tackling discrete graph data. When comparing to current adversarial attacks on GNNs, the results show that by only perturbing a small number of edge perturbations, including addition and deletion, our optimization-based attack can lead to a noticeable decrease in classification performance. Moreover, leveraging our gradient-based attack, we propose the first optimization-based adversarial training for GNNs. Our method yields higher robustness against both different gradient based and greedy attack methods without sacrificing classification accuracy on original graph.
CITATION STYLE
Xu, K., Chen, H., Liu, S., Chen, P. Y., Weng, T. W., Hong, M., & Lin, X. (2019). Topology attack and defense for graph neural networks: An optimization perspective. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2019-August, pp. 3961–3967). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2019/550
Mendeley helps you to discover research relevant for your work.