NGAT: attention in breadth and depth exploration for semi-supervised graph representation learning

Jianke HU, Yin ZHANG

PDF(1019 KB)
PDF(1019 KB)
Front. Inform. Technol. Electron. Eng ›› 2022, Vol. 23 ›› Issue (3) : 409-421. DOI: 10.1631/FITEE.2000657
Orginal Article
Orginal Article

NGAT: attention in breadth and depth exploration for semi-supervised graph representation learning

Author information +
History +

Abstract

Recently, graph neural networks (GNNs) have achieved remarkable performance in representation learning on graph-structured data. However, as the number of network layers increases, GNNs based on the neighborhood aggregation strategy deteriorate due to the problem of oversmoothing, which is the major bottleneck for applying GNNs to real-world graphs. Many efforts have been made to improve the process of feature information aggregation from directly connected nodes, i.e., breadth exploration. However, these models perform the best only in the case of three or fewer layers, and the performance drops rapidly for deep layers. To alleviate oversmoothing, we propose a nested graph attention network (NGAT), which can work in a semi-supervised manner. In addition to breadth exploration, a k-layer NGAT uses a layer-wise aggregation strategy guided by the attention mechanism to selectively leverage feature information from the kth-order neighborhood, i.e., depth exploration. Even with a 10-layer or deeper architecture, NGAT can balance the need for preserving the locality (including root node features and the local structure) and aggregating the information from a large neighborhood. In a number of experiments on standard node classification tasks, NGAT outperforms other novel models and achieves state-of-the-art performance.

Keywords

Graph learning / Semi-supervised learning / Node classification / Attention

Cite this article

Download citation ▾
Jianke HU, Yin ZHANG. NGAT: attention in breadth and depth exploration for semi-supervised graph representation learning. Front. Inform. Technol. Electron. Eng, 2022, 23(3): 409‒421 https://doi.org/10.1631/FITEE.2000657

RIGHTS & PERMISSIONS

2022 Zhejiang University Press
PDF(1019 KB)

Accesses

Citations

Detail

Sections
Recommended

/