Graph Propagation Transformer for Graph Representation Learning

2Citations
Citations of this article
18Readers
Mendeley users who have this article in their library.

Abstract

This paper presents a novel transformer architecture for graph representation learning. The core insight of our method is to fully consider the information propagation among nodes and edges in a graph when building the attention module in the transformer blocks. Specifically, we propose a new attention mechanism called Graph Propagation Attention (GPA). It explicitly passes the information among nodes and edges in three ways, i.e., node-to-node, node-to-edge, and edge-to-node, which is essential for learning graph-structured data. On this basis, we design an effective transformer architecture named Graph Propagation Transformer (GPTrans) to further help learn graph data. We verify the performance of GPTrans in a wide range of graph learning experiments on several benchmark datasets. These results show that our method outperforms many state-of-the-art transformer-based graph models with better performance. The code will be released at https://github.com/czczup/GPTrans.

Cite

CITATION STYLE

APA

Chen, Z., Tan, H., Wang, T., Shen, T., Lu, T., Peng, Q., … Qi, Y. (2023). Graph Propagation Transformer for Graph Representation Learning. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2023-August, pp. 3559–3567). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2023/396

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free