Graph Information Vanishing Phenomenon in Implicit Graph Neural Networks
-
Published:2024-08-27
Issue:17
Volume:12
Page:2659
-
ISSN:2227-7390
-
Container-title:Mathematics
-
language:en
-
Short-container-title:Mathematics
Author:
He Silu1ORCID, Cao Jun1, Yuan Hongyuan1, Chen Zhe1, Gao Shijuan12, Li Haifeng1ORCID
Affiliation:
1. School of Geosciences and Info-Physics, Central South University, Changsha 410083, China 2. Information & Network Center, Central South University, Changsha 410083, China
Abstract
Graph neural networks (GNNs) have been highly successful in graph representation learning. The goal of GNNs is to enrich node representations by aggregating information from neighboring nodes. Much work has attempted to improve the quality of aggregation by introducing a variety of graph information with representational capabilities. The class of GNNs that improves the quality of aggregation by encoding graph information with representational capabilities into the weights of neighboring nodes through different learnable transformation structures (LTSs) are referred to as implicit GNNs. However, we argue that LTSs only transform graph information into the weights of neighboring nodes in the direction that minimizes the loss function during the learning process and does not actually utilize the effective properties of graph information, a phenomenon that we refer to as graph information vanishing (GIV). To validate this point, we perform thousands of experiments on seven node classification benchmark datasets. We first replace the graph information utilized by five implicit GNNs with random values and surprisingly observe that the variation range of accuracies is less than ± 0.3%. Then, we quantitatively characterize the similarity of the weights generated from graph information and random values by cosine similarity, and the cosine similarities are greater than 0.99. The empirical experiments show that graph information is equivalent to initializing the input of LTSs. We believe that graph information as an additional supervised signal to constrain the training of GNNs can effectively solve GIV. Here, we propose GinfoNN, which utilizes both labels and discrete graph curvature as supervised signals to jointly constrain the training of the model. The experimental results show that the classification accuracies of GinfoNN improve by two percentage points over baselines on large and dense datasets.
Funder
National Natural Science Foundation of China Research Foundation of the Department of Natural Resources of Hunan Province Scientific Research Project of the Department of Education of Hunan Province
Reference47 articles.
1. Zhang, M., and Chen, Y. (2018, January 3–8). Link prediction based on graph neural networks. Proceedings of the 32nd International Conference on Neural Information Processing Systems, Red Hook, NY, USA. 2. Bian, T., Xiao, X., Xu, T., Zhao, P., Huang, W., Rong, Y., and Huang, J. (2020, January 7–12). Rumor detection on social media with bi-directional graph convolutional networks. Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA. 3. Yu, B., Yin, H., and Zhu, Z. (2018, January 13–19). Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden. 4. Kosaraju, V., Sadeghian, A., Martín-Martín, R., Reid, I.D., Rezatofighi, H., and Savarese, S. (2019, January 8–14). Social-BiGAT: Multimodal trajectory forecasting using Bicycle-GAN and graph attention networks. Proceedings of the Advances in Neural Information Processing Systems, Neural Information Processing Systems (NIPS), Vancouver, BC, Canada. 5. Duvenaud, D., Maclaurin, D., Aguilera-Iparraguirre, J., Gómez-Bombarelli, R., Hirzel, T., Aspuru-Guzik, A., and Adams, R.P. (2015, January 7–12). Convolutional networks on graphs for learning molecular fingerprints. Proceedings of the 28th International Conference on Neural Information Processing Systems, Cambridge, MA, USA.
|
|