Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3340531.3411872acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article

Investigating and Mitigating Degree-Related Biases in Graph Convoltuional Networks

Published: 19 October 2020 Publication History
  • Get Citation Alerts
  • Abstract

    Graph Convolutional Networks (GCNs) show promising results for semi-supervised learning tasks on graphs, thus become favorable comparing with other approaches. Despite the remarkable success of GCNs, it is difficult to train GCNs with insufficient supervision. When labeled data are limited, the performance of GCNs becomes unsatisfying for low-degree nodes. While some prior work analyze successes and failures of GCNs on the entire model level, profiling GCNs on individual node level is still underexplored.
    In this paper, we analyze GCNs in regard to the node degree distribution. From empirical observation to theoretical proof, we confirm that GCNs are biased towards nodes with larger degrees with higher accuracy on them, even if high-degree nodes are underrepresented in most graphs. We further develop a novel Self-Supervised-Learning Degree-Specific GCN (SL-DSGCN) that mitigate the degree-related biases of GCNs from model and data aspects. Firstly, we propose a degree-specific GCN layer that captures both discrepancies and similarities of nodes with different degrees, which reduces the inner model-aspect biases of GCNs caused by sharing the same parameters with all nodes. Secondly, we design a self-supervised-learning algorithm that creates pseudo labels with uncertainty scores on unlabeled nodes with a Bayesian neural network. Pseudo labels increase the chance of connecting to labeled neighbors for low-degree nodes, thus reducing the biases of GCNs from the data perspective. Uncertainty scores are further exploited to weight pseudo labels dynamically in the stochastic gradient descent for SL-DSGCN. Experiments on three benchmark datasets show SL-DSGCN not only outperforms state-of-the-art self-training/self-supervised-learning GCN methods, but also improves GCN accuracy dramatically for low-degree nodes.

    Supplementary Material

    MP4 File (3340531.3411872.mp4)
    Graph Convolutional Networks (GCNs) show promising results for semi-supervised learning tasks, but are limited with insufficient supervision. When labeled data are limited, the performance of GCNs becomes unsatisfying for low-degree nodes. We analyze GCNs regarding the node degree distribution. From empirical observation to theoretical proof, we confirm that GCNs are biased towards nodes with larger degrees with higher accuracy on them, even if high-degree nodes are underrepresented in most graphs. We further develop a novel Self-Supervised-Learning Degree-Specific GCN that mitigates the degree-related biases of GCNs from model and data aspects. Firstly, we propose a degree-specific GCN layer for nodes with different degrees. Secondly, we design a self-supervised-learning algorithm that creates pseudo labels with uncertainty scores on unlabeled nodes with a Bayesian neural network. Experiments on three benchmark datasets show significant improvements over baseline methods.

    References

    [1]
    Réka Albert and Albert-László Barabási. 2002. Statistical mechanics of complex networks. Reviews of modern physics (2002).
    [2]
    James Atwood and Don Towsley. 2016. Diffusion-convolutional neural networks. In Advances in neural information processing systems.
    [3]
    Michael M Bronstein, Joan Bruna, Yann LeCun, Arthur Szlam, and Pierre Vandergheynst. 2017. Geometric deep learning: going beyond euclidean data. IEEE Signal Processing Magazine (2017).
    [4]
    Joan Bruna, Wojciech Zaremba, Arthur Szlam, and Yann LeCun. 2013. Spectral networks and locally connected networks on graphs. arXiv:1312.6203 (2013).
    [5]
    Mathilde Caron, Piotr Bojanowski, Armand Joulin, and Matthijs Douze. 2018. Deep clustering for unsupervised learning of visual features. In ECCV.
    [6]
    Aaron Clauset, Cosma Rohilla Shalizi, and Mark EJ Newman. 2009. Power-law distributions in empirical data. SIAM review (2009).
    [7]
    Hanjun Dai, Hui Li, Tian Tian, Xin Huang, Lin Wang, Jun Zhu, and Le Song. 2019. Adversarial attack on graph structured data. In ICML.
    [8]
    Michaël Defferrard, Xavier Bresson, and Pierre Vandergheynst. 2016. Convolutional neural networks on graphs with fast localized spectral filtering. In NeurIPS.
    [9]
    Michalis Faloutsos, Petros Faloutsos, and Christos Faloutsos. 1999. On power-law relationships of the internet topology. ACM SIGCOMM computer communication review (1999).
    [10]
    Will Hamilton, Zhitao Ying, and Jure Leskovec. 2017. Inductive representation learning on large graphs. In NeurIPS.
    [11]
    Chao Huang, Xian Wu, Xuchao Zhang, Chuxu Zhang, Jiashu Zhao, Dawei Yin, and Nitesh V Chawla. 2019. Online purchase prediction via multi-scale modeling of behavior dynamics. In KDD.
    [12]
    Wei Jin, Tyler Derr, Haochen Liu, Yiqi Wang, Suhang Wang, Zitao Liu, and Jiliang Tang. 2020 a. Self-supervised Learning on Graphs: Deep Insights and New Direction. arXiv preprint arXiv:2006.10141 (2020).
    [13]
    Wei Jin, Yao Ma, Xiaorui Liu, Xianfeng Tang, Suhang Wang, and Jiliang Tang. 2020 b. Graph Structure Learning for Robust Graph Neural Networks. arXiv:2005.10203 (2020).
    [14]
    Longlong Jing and Yingli Tian. 2020. Self-supervised visual feature learning with deep neural networks: A survey. T-PAMI (2020).
    [15]
    Thomas N Kipf and Max Welling. 2017. Semi-Supervised Classification with Graph Convolutional Networks. In ICLR.
    [16]
    Pang Wei Koh and Percy Liang. 2017. Understanding black-box predictions via influence functions. In ICML.
    [17]
    Qimai Li, Zhichao Han, and Xiao-Ming Wu. 2018. Deeper insights into graph convolutional networks for semi-supervised learning. In AAAI.
    [18]
    Ruirui Li, Xian Wu, Xian Wu, and Wei Wang. 2020. Few-Shot Learning for New User Recommendation in Location-based Social Networks. In WWW.
    [19]
    Alan Mislove, Massimiliano Marcon, Krishna P Gummadi, Peter Druschel, and Bobby Bhattacharjee. 2007. Measurement and analysis of online social networks. In SIGCOMM.
    [20]
    Mehdi Noroozi and Paolo Favaro. 2016. Unsupervised learning of visual representations by solving jigsaw puzzles. In ECCV.
    [21]
    Deepak Pathak, Philipp Krahenbuhl, Jeff Donahue, Trevor Darrell, and Alexei A Efros. 2016. Context encoders: Feature learning by inpainting. In CVPR.
    [22]
    Prithviraj Sen, Galileo Namata, Mustafa Bilgic, Lise Getoor, Brian Galligher, and Tina Eliassi-Rad. 2008. Collective classification in network data. AI magazine (2008).
    [23]
    Chuan Shi, Yitong Li, Jiawei Zhang, Yizhou Sun, and S Yu Philip. 2016. A survey of heterogeneous information network analysis. TKDE (2016).
    [24]
    Ke Sun, Zhouchen Lin, and Zhanxing Zhu. 2019. Multi-Stage Self-Supervised Learning for Graph Convolutional Networks on Graphs with Few Labels. arXiv:1902.11038 (2019).
    [25]
    Yiwei Sun, Suhang Wang, Xianfeng Tang, Tsung-Yu Hsieh, and Vasant Honavar. 2020. Adversarial Attacks on Graph Neural Networks via Node Injections: A Hierarchical Reinforcement Learning Approach. In WWW. 673--683.
    [26]
    Xianfeng Tang, Yandong Li, Yiwei Sun, Huaxiu Yao, Prasenjit Mitra, and Suhang Wang. 2020 a. Transferring Robustness for Graph Neural Network Against Poisoning Attacks. In WSDM.
    [27]
    Xianfeng Tang, Yozen Liu, Neil Shah, Xiaolin Shi, Prasenjit Mitra, and Suhang Wang. 2020 b. Knowing your FATE: Friendship, Action and Temporal Explanations for User Engagement Prediction on Social Apps. arXiv:2006.06427 (2020).
    [28]
    Petar Velivc ković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. 2017. Graph attention networks. arXiv:1710.10903 (2017).
    [29]
    Daixin Wang, Jianbin Lin, Peng Cui, Quanhui Jia, Zhen Wang, Yanming Fang, Quan Yu, Jun Zhou, Shuang Yang, and Yuan Qi. 2019. A Semi-supervised Graph Attentive Network for Financial Fraud Detection. In ICDM.
    [30]
    Xiaoyang Wang, Yao Ma, Yiqi Wang, Wei Jin, Xin Wang, Jiliang Tang, Caiyan Jia, and Jian Yu. 2020. Traffic Flow Prediction via Spatial Temporal Graph Neural Network. In WWW.
    [31]
    Jun Wu, Jingrui He, and Jiejun Xu. 2019. Demo-net: Degree-specific graph neural networks for node and graph classification. In KDD.
    [32]
    Xiao-Ming Wu, Zhenguo Li, Anthony M So, John Wright, and Shih-Fu Chang. 2012. Learning with partially absorbing random walks. In NeurIPS.
    [33]
    Zonghan Wu, Shirui Pan, Fengwen Chen, Guodong Long, Chengqi Zhang, and S Yu Philip. 2020. A comprehensive survey on graph neural networks. IEEE Neural Networks and Learning Systems (2020).
    [34]
    Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. 2018a. How powerful are graph neural networks? arXiv:1810.00826 (2018).
    [35]
    Keyulu Xu, Chengtao Li, Yonglong Tian, Tomohiro Sonobe, Ken-ichi Kawarabayashi, and Stefanie Jegelka. 2018b. Representation learning on graphs with jumping knowledge networks. arXiv:1806.03536 (2018).
    [36]
    Rex Ying, Ruining He, Kaifeng Chen, Pong Eksombatchai, William L Hamilton, and Jure Leskovec. 2018. Graph convolutional neural networks for web-scale recommender systems. In KDD.
    [37]
    Jiaxuan You, Bowen Liu, Zhitao Ying, Vijay Pande, and Jure Leskovec. 2018. Graph convolutional policy network for goal-directed molecular graph generation. In NeurIPS.
    [38]
    Bing Yu, Haoteng Yin, and Zhanxing Zhu. 2017. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. arXiv:1709.04875 (2017).
    [39]
    Jiani Zhang, Xingjian Shi, Junyuan Xie, Hao Ma, Irwin King, and Dit-Yan Yeung. 2018. Gaan: Gated attention networks for learning on large and spatiotemporal graphs. arXiv:1803.07294 (2018).
    [40]
    Dengyong Zhou, Olivier Bousquet, Thomas N Lal, Jason Weston, and Bernhard Schölkopf. 2004. Learning with local and global consistency. In NeurIPS.
    [41]
    Xiaojin Zhu and Zoubin Ghahramani. 2002. Learning from labeled and unlabeled data with label propagation. (2002).
    [42]
    Daniel Zügner, Amir Akbarnejad, and Stephan Günnemann. 2018. Adversarial attacks on neural networks for graph data. In KDD.

    Cited By

    View all
    • (2024)"Pinocchio had a Nose, You have a Network!": On Characterizing Fake News Spreaders on Arabic Social MediaProceedings of the ACM on Human-Computer Interaction10.1145/36537028:CSCW1(1-20)Online publication date: 26-Apr-2024
    • (2024)Fairness-Aware Graph Neural Networks: A SurveyACM Transactions on Knowledge Discovery from Data10.1145/364914218:6(1-23)Online publication date: 12-Apr-2024
    • (2024)Bayesian Graph Local Extrema Convolution with Long-tail Strategy for Misinformation DetectionACM Transactions on Knowledge Discovery from Data10.1145/363940818:4(1-21)Online publication date: 12-Feb-2024
    • Show More Cited By

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    CIKM '20: Proceedings of the 29th ACM International Conference on Information & Knowledge Management
    October 2020
    3619 pages
    ISBN:9781450368599
    DOI:10.1145/3340531
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 19 October 2020

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. deep learning
    2. graph neural network
    3. self-supervised learning

    Qualifiers

    • Research-article

    Funding Sources

    • National Science Foundation
    • Global Research Outreach program of Samsung Advanced Institute of Technology

    Conference

    CIKM '20
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 1,861 of 8,427 submissions, 22%

    Upcoming Conference

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)145
    • Downloads (Last 6 weeks)15
    Reflects downloads up to 26 Jul 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)"Pinocchio had a Nose, You have a Network!": On Characterizing Fake News Spreaders on Arabic Social MediaProceedings of the ACM on Human-Computer Interaction10.1145/36537028:CSCW1(1-20)Online publication date: 26-Apr-2024
    • (2024)Fairness-Aware Graph Neural Networks: A SurveyACM Transactions on Knowledge Discovery from Data10.1145/364914218:6(1-23)Online publication date: 12-Apr-2024
    • (2024)Bayesian Graph Local Extrema Convolution with Long-tail Strategy for Misinformation DetectionACM Transactions on Knowledge Discovery from Data10.1145/363940818:4(1-21)Online publication date: 12-Feb-2024
    • (2024)Pitfalls in Link Prediction with Graph Neural Networks: Understanding the Impact of Target-link Inclusion & Better PracticesProceedings of the 17th ACM International Conference on Web Search and Data Mining10.1145/3616855.3635786(994-1002)Online publication date: 4-Mar-2024
    • (2024)Knowledge Graph-based Session Recommendation with Session-Adaptive PropagationCompanion Proceedings of the ACM on Web Conference 202410.1145/3589335.3648324(264-273)Online publication date: 13-May-2024
    • (2024)Brave the Wind and the Waves: Discovering Robust and Generalizable Graph Lottery TicketsIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2023.334218446:5(3388-3405)Online publication date: May-2024
    • (2024)Tackling Long-tailed Distribution Issue in Graph Neural Networks via NormalizationIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2023.3315284(1-11)Online publication date: 2024
    • (2024)Locality-Aware Tail Node Embeddings on Homogeneous and Heterogeneous NetworksIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2023.3313355(1-16)Online publication date: 2024
    • (2024)FairSample: Training Fair and Accurate Graph Convolutional Neural Networks EfficientlyIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2023.330637836:4(1537-1551)Online publication date: Apr-2024
    • (2024)DeltaGNN: Accelerating Graph Neural Networks on Dynamic Graphs With Delta UpdatingIEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems10.1109/TCAD.2023.333515343:4(1163-1176)Online publication date: Apr-2024
    • Show More Cited By

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media