DOI QR코드

DOI QR Code

A Gradient-Based Explanation Method for Node Classification Using Graph Convolutional Networks

  • Chaehyeon Kim ( Dept. of Computer Science, Sookmyung Women's University) ;
  • Hyewon Ryu ( Dept. of Software Convergence, Sookmyung Women's University) ;
  • Ki Yong Lee ( Dept. of Computer Science, Sookmyung Women's University)
  • Received : 2022.08.16
  • Accepted : 2023.01.21
  • Published : 2023.12.31

Abstract

Explainable artificial intelligence is a method that explains how a complex model (e.g., a deep neural network) yields its output from a given input. Recently, graph-type data have been widely used in various fields, and diverse graph neural networks (GNNs) have been developed for graph-type data. However, methods to explain the behavior of GNNs have not been studied much, and only a limited understanding of GNNs is currently available. Therefore, in this paper, we propose an explanation method for node classification using graph convolutional networks (GCNs), which is a representative type of GNN. The proposed method finds out which features of each node have the greatest influence on the classification of that node using GCN. The proposed method identifies influential features by backtracking the layers of the GCN from the output layer to the input layer using the gradients. The experimental results on both synthetic and real datasets demonstrate that the proposed explanation method accurately identifies the features of each node that have the greatest influence on its classification.

Keywords

Acknowledgement

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. NRF-2021R1A2C1012543). This paper is the extended version of the Annual Spring Conference of KIPS (ASK 2022) held in Seoul, Republic of Korea dated May 19-21, 2022 [11].

References

  1. A. B. Arrieta, N. Diaz-Rodriguez, J. Del Ser, A. Bennetot, S. Tabik, A. Barbado, et al., "Explainable Artificial Intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI," Information Fusion, vol. 58, pp. 82-115, 2020. https://doi.org/10.1016/j.inffus.2019.12.012
  2. S. Lee, S. Lee, D. Chang, M. H. Song, J. Y. Kim, and S. Lee, "Explainable machine learning based a packed red blood cell transfusion prediction and evaluation for major internal medical condition," Journal of Information Processing Systems, vol. 18, no. 3, pp. 302-310, 2022. https://doi.org/10.3745/JIPS.04.0243
  3. T. N. Kipf and M. Welling, "Semi-supervised classification with graph convolutional networks," 2017 [Online]. Available: https://arxiv.org/abs/1609.02907.
  4. Z. Wu, S. Pan, F. Chen, G. Long, C. Zhang, and P. S. Yu, "A comprehensive survey on graph neural networks," IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 1, pp. 4-24, 2021. https://doi.org/10.1109/TNNLS.2020.2978386
  5. B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, and A. Torralba, "Learning deep features for discriminative localization," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 2016, pp. 2921-2929. https://doi.org/10.1109/CVPR.2016.319
  6. R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, "Grad-CAM: visual explanations from deep networks via gradient-based localization," in Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 2017, pp. 618-626. https://doi.org/10.1109/ICCV.2017.74
  7. S. Bach, A. Binder, G. Montavon, F. Klauschen, K. R. Muller, and W. Samek, "On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation," PLoS One, vol. 10, no. 7, article no. e0130140, 2015. https://doi.org/10.1371/journal.pone.0130140
  8. M. T. Ribeiro, S. Singh, and C. Guestrin, "Why should I trust you? Explaining the predictions of any classifier," in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 2016, pp. 1135-1144. https://doi.org/10.1145/2939672.2939778
  9. E. Tjoa and C. Guan, "A survey on explainable artificial intelligence (XAI): toward medical XAI," IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 11, pp. 4793-4813, 2021. https://doi.org/10.1109/TNNLS.2020.3027314
  10. J. Hu, T. Li, and S. Dong, "GCN-LRP explanation: exploring latent attention of graph convolutional networks," in Proceedings of 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 2020, pp. 1-8. https://doi.org/10.1109/IJCNN48605.2020.9207639
  11. C. Kim and K. Y. Lee, "A gradient-based explanation method for graph convolutional neural networks," in Proceedings of Annual Spring Conference of KIPS (ASK 2022), 2022, pp. 670-673.
  12. P. Sen, G. Namata, M. Bilgic, L. Getoor, B. Galligher, and T. Eliassi-Rad, "Collective classification in network data," AI Magazine, vol. 29, no. 3, pp. 93-93, 2008. https://doi.org/10.1609/aimag.v29i3.2157
  13. G. Vilone and L. Longo, "Notions of explainability and evaluation approaches for explainable artificial intelligence," Information Fusion, vol. 76, pp. 89-106, 2021. https://doi.org/10.1016/j.inffus.2021.05.009