DOI QR코드

DOI QR Code

GBGNN: Gradient Boosted Graph Neural Networks

  • Eunjo Jang (Dept. of Computer Science, Sookmyung Women's University) ;
  • Ki Yong Lee (Dept. of Computer Science, Sookmyung Women's University)
  • Received : 2023.02.23
  • Accepted : 2023.07.23
  • Published : 2024.08.31

Abstract

In recent years, graph neural networks (GNNs) have been extensively used to analyze graph data across various domains because of their powerful capabilities in learning complex graph-structured data. However, recent research has focused on improving the performance of a single GNN with only two or three layers. This is because stacking layers deeply causes the over-smoothing problem of GNNs, which degrades the performance of GNNs significantly. On the other hand, ensemble methods combine individual weak models to obtain better generalization performance. Among them, gradient boosting is a powerful supervised learning algorithm that adds new weak models in the direction of reducing the errors of the previously created weak models. After repeating this process, gradient boosting combines the weak models to produce a strong model with better performance. Until now, most studies on GNNs have focused on improving the performance of a single GNN. In contrast, improving the performance of GNNs using multiple GNNs has not been studied much yet. In this paper, we propose gradient boosted graph neural networks (GBGNN) that combine multiple shallow GNNs with gradient boosting. We use shallow GNNs as weak models and create new weak models using the proposed gradient boosting-based loss function. Our empirical evaluations on three real-world datasets demonstrate that GBGNN performs much better than a single GNN. Specifically, in our experiments using graph convolutional network (GCN) and graph attention network (GAT) as weak models on the Cora dataset, GBGNN achieves performance improvements of 12.3%p and 6.1%p in node classification accuracy compared to a single GCN and a single GAT, respectively.

Keywords

Acknowledgement

This paper is the extended version of "A gradient boosting method for graph neural networks," in the Annual Conference of KIPS (ACK 2022) held in Seoul, Republic of Korea dated November 3-5, 2022.

References

  1. Z. Wu, S. Pan, F. Chen, G. Long, C. Zhang, and S. Y. Philip, "A comprehensive survey on graph neural networks," IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 1, pp. 4-24, 2021. https://doi.org/10.1109/TNNLS.2020.2978386 
  2. Y. Ding, Z. Zhang, X. Zhao, W. Cai, F. He, Y. Cai, and W. W. Cai, "Deep hybrid: multi-graph neural network collaboration for hyperspectral image classification," Defence Technology, vol. 23, pp. 164-176, 2023. https://doi.org/10.1016/j.dt.2022.02.007 
  3. S. Wu, F. Sun, W. Zhang, X. Xie, and B. Cui, "Graph neural networks in recommender systems: a survey," ACM Computing Surveys, vol. 55, no. 5, pp. 1-37, 2022. https://doi.org/10.1145/3535101 
  4. M. Gori, G. Monfardini, and F. Scarselli, "A new model for learning in graph domains," in Proceedings of 2005 IEEE International Joint Conference on Neural Networks, Montreal, Canada, 2005, pp. 729-734. https://doi.org/10.1109/IJCNN.2005.1555942 
  5. W. L. Hamilton, R. Ying, and J. Leskovec, "Representation learning on graphs: methods and applications," 2017 [Online]. Available: https://arxiv.org/abs/1709.05584. 
  6. T. N. Kipf and M. Welling, "Semi-supervised classification with graph convolutional networks," 2016 [Online]. Available: https://arxiv.org/abs/1609.02907. 
  7. P. Velickovic, G. Cucurull, A. Casanova, A. Romero, P. Lio, and Y. Bengio, "Graph attention networks," 2018 [Online]. Available: https://arxiv.org/abs/1710.10903. 
  8. W. Hamilton, Z. Ying, and J. Leskovec, "Inductive representation learning on large graphs," Advances in Neural Information Processing Systems, vol. 30, pp. 1024-1034, 2017. 
  9. Y. Yang, T. Liu, Y. Wang, J. Zhou, Q. Gan, Z. Wei, Z. Zhang, Z. Huang, and D. Wipf, "Graph neural networks inspired by classical iterative algorithms," Proceedings of Machine Learning Research, vol. 139, pp. 11773- 11783, 2021. 
  10. J. H. Friedman, "Greedy function approximation: a gradient boosting machine," Annals of Statistics, vol. 29, no. 5, pp. 1189-1232, 2001.  https://doi.org/10.1214/aos/1013203451
  11. S. Badirli, X. Liu, Z. Xing, A. Bhowmik, K. Doan, and S. S. Keerthi, "Gradient boosting neural networks: GrowNet," 2020 [Online]. Available: https://arxiv.org/abs/2002.07971. 
  12. S. Ivanov and L. Prokhorenkova, "Boost then convolve: gradient boosting meets graph neural networks," 2021 [Online]. Available: https://arxiv.org/abs/2101.08543. 
  13. K. Sun, Z. Zhu, and Z. Lin, "AdaGCN: Adaboosting graph convolutional networks into deep models," 2019 [Online]. Available: https://arxiv.org/abs/1908.05081. 
  14. M. Defferrard, X. Bresson, and P. Vandergheynst, "Convolutional neural networks on graphs with fast localized spectral filtering," Advances in Neural Information Processing Systems, vol. 29, pp. 3837-3845, 2016. 
  15. Cora dataset [Online]. Available: https://paperswithcode.com/dataset/cora. 
  16. Citeseer dataset [Online]. https://paperswithcode.com/dataset/citeseer. 
  17. PubMed dataset [Online]. Avaiable: https://paperswithcode.com/dataset/pubmed.