Heterogeneous Graph Transformer
Citations Over TimeTop 1% of 2020 papers
Abstract
Recent years have witnessed the emerging success of graph neural networks (GNNs) for modeling structured data. However, most GNNs are designed for homogeneous graphs, in which all nodes and edges belong to the same types, making it infeasible to represent heterogeneous structures. In this paper, we present the Heterogeneous Graph Transformer (HGT) architecture for modeling Web-scale heterogeneous graphs. To model heterogeneity, we design node- and edge-type dependent parameters to characterize the heterogeneous attention over each edge, empowering HGT to maintain dedicated representations for different types of nodes and edges. To handle Web-scale graph data, we design the heterogeneous mini-batch graph sampling algorithm—HGSampling—for efficient and scalable training. Extensive experiments on the Open Academic Graph of 179 million nodes and 2 billion edges show that the proposed HGT model consistently outperforms all the state-of-the-art GNN baselines by 9–21 on various downstream tasks. The dataset and source code of HGT are publicly available at https://github.com/acbull/pyHGT.
Related Papers
- → Scalability Issues of Blockchain Technology(2020)30 cited
- → INVITED PAPER: Scalability and Performance Issues in Deeply Embedded Sensor Systems(2009)7 cited
- → On the scalability of multistage interconnection networks(2004)3 cited
- → Using Empirical Data for Scalability Analysis of Parallel Applications(2019)1 cited
- RESEARCH ON THE SCALABILITY OF THE LARGE SCALE PARALLEL APPLICATION PROGRAMS(2000)