Topology attack and defense for graph neural networks: An optimization perspective

209Citations
Citations of this article
168Readers
Mendeley users who have this article in their library.

Abstract

Graph neural networks (GNNs) which apply the deep neural networks to graph data have achieved significant performance for the task of semi-supervised node classification. However, only few work has addressed the adversarial robustness of GNNs. In this paper, we first present a novel gradient-based attack method that facilitates the difficulty of tackling discrete graph data. When comparing to current adversarial attacks on GNNs, the results show that by only perturbing a small number of edge perturbations, including addition and deletion, our optimization-based attack can lead to a noticeable decrease in classification performance. Moreover, leveraging our gradient-based attack, we propose the first optimization-based adversarial training for GNNs. Our method yields higher robustness against both different gradient based and greedy attack methods without sacrificing classification accuracy on original graph.

Cite

CITATION STYLE

APA

Xu, K., Chen, H., Liu, S., Chen, P. Y., Weng, T. W., Hong, M., & Lin, X. (2019). Topology attack and defense for graph neural networks: An optimization perspective. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2019-August, pp. 3961–3967). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2019/550

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free