Ivysaur-One Stitched Toy Graph Neural Network

Shanzai CL

Published on: 2021-08-12

Abstract

Graph neural networks have shown significant success in the field of graph representation learning. Graph convolution performs neighborhood aggregation and represents one of the most important graph operations. However, one layer of these neighbor aggregation methods only considers direct neighbors, and when several recent studies attributed this performance degradation to an over-refined problem, the problem pointed out that repeated propagation makes it difficult to distinguish between different types of representations. In this work, we believe that the key factor affecting performance is the entanglement of representation conversion and propagation in current graphics convolution operations. After decomposing these two operations, a deeper graph neural network can be used to learn the graph level representation from the above receptive domain. Two simple and effective techniques are used in combination: initial residuals and Identity mapping. Based on our theoretical and empirical analysis, we propose a Deep Adaptive Graph Initial Residual Identity Mapping Neural Network (IVYSAUR) to adaptively merge information from large receiving fields. Each experiment on citations, co-authors, and co-purchase data sets confirms our analysis and insights and proves the superiority of our proposed method.

Keywords

Graph Convolutional Network; Initial Residuals; Identity Mapping; Graph Representation Learning

Introduction

Graphs, representing entities and their relationships, are ubiquitous in the real world, such as social networks, point clouds, traffic networks, knowledge graphs, and molecular structures. Recently, many studies focus on developing deep learning approaches for graph data, leading to rapid development in the field of graph neural networks. Great successes have been achieved for many applications, such as node classification [1-8], graph classification [9-15] and link prediction [16-18] Graph convolutions adopt a neighborhood aggregation (or message passing) scheme to learn node representations by considering the node features and graph topology information together, among which the most representative method is Graph Convolutional Networks (GCNs) [19]. GCN learns representation for a node by aggregating representations of its neighbors iteratively. However, a common challenge faced by GCN and most other graph convolutions is that one layer of graph convolutions only consider immediate neighbors and the performance degrades greatly when we apply multiple layers to leverage large receptive fields. Several recent works attribute this performance degradation to the oversmoothing issue [20-22], which states that representations from different classes become inseparable due to repeated propagation. In this work, we study this performance deterioration systematically and develop new insights towards deeper graph neural networks.