Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3447548.3467221acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

ROD: Reception-aware Online Distillation for Sparse Graphs

Published: 14 August 2021 Publication History

Abstract

Graph neural networks (GNNs) have been widely used in many graph-based tasks such as node classification, link prediction, and node clustering. However, GNNs gain their performance benefits mainly from performing the feature propagation and smoothing across the edges of the graph, thus requiring sufficient connectivity and label information for effective propagation. Unfortunately, many real-world networks are sparse in terms of both edges and labels, leading to sub-optimal performance of GNNs. Recent interest in this sparse problem has focused on the self-training approach, which expands supervised signals with pseudo labels. Nevertheless, the self-training approach inherently cannot realize the full potential of refining the learning performance on sparse graphs due to the unsatisfactory quality and quantity of pseudo labels.
In this paper, we propose ROD, a novel reception-aware online knowledge distillation approach for sparse graph learning. We design three supervision signals for ROD: multi-scale reception-aware graph knowledge, task-based supervision, and rich distilled knowledge, allowing online knowledge transfer in a peer-teaching manner. To extract knowledge concealed in the multi-scale reception fields, ROD explicitly requires individual student models to preserve different levels of locality information. For a given task, each student would predict based on its reception-scale knowledge, while simultaneously a strong teacher is established on-the-fly by combining multi-scale knowledge. Our approach has been extensively evaluated on 9 datasets and a variety of graph-based tasks, including node classification, link prediction, and node clustering. The result demonstrates that ROD achieves state-of-art performance and is more robust for the graph sparsity.

Supplementary Material

MP4 File (ROD.mp4)
Presentation video

References

[1]
Luciano Baresi and Sam Guinea. 2010. Self-supervising bpel processes. IEEE Transactions on Software Engineering, Vol. 37, 2 (2010), 247--263.
[2]
Deyu Bo, Xiao Wang, Chuan Shi, Meiqi Zhu, Emiao Lu, and Peng Cui. 2020. Structural deep clustering network. In WWWW 2020. 1400--1410.
[3]
Defang Chen, Jian-Ping Mei, Can Wang, Yan Feng, and Chun Chen. 2020. Online knowledge distillation with diverse peers. In AAAI, Vol. 34. 3430--3437.
[4]
Jie Chen, Tengfei Ma, and Cao Xiao. 2018. FastGCN: Fast Learning with Graph Convolutional Networks via Importance Sampling. In ICLR.
[5]
Wei-Lin Chiang, Xuanqing Liu, Si Si, Yang Li, Samy Bengio, and Cho-Jui Hsieh. 2019. Cluster-gcn: An efficient algorithm for training deep and large graph convolutional networks. In SIGKDD. 257--266.
[6]
Ganqu Cui, Jie Zhou, Cheng Yang, and Zhiyuan Liu. 2020. Adaptive Graph Encoder for Attributed Graph Embedding. In SIGKDD. 976--985.
[7]
Victor Garcia and Joan Bruna. 2017. Few-shot learning with graph neural networks. arXiv preprint arXiv:1711.04043 (2017).
[8]
William L Hamilton, Rex Ying, and Jure Leskovec. 2017. Inductive representation learning on large graphs. In NeurIPS. 1025--1035.
[9]
John A Hartigan and Manchek A Wong. 1979. Algorithm AS 136: A k-means clustering algorithm. Journal of the royal statistical society. series c (applied statistics), Vol. 28, 1 (1979), 100--108.
[10]
Xiangnan He, Kuan Deng, Xiang Wang, Yan Li, Yongdong Zhang, and Meng Wang. 2020. LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation. arXiv preprint arXiv:2002.02126 (2020).
[11]
Thomas N Kipf and Max Welling. 2016a. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016).
[12]
Thomas N Kipf and Max Welling. 2016b. Variational graph auto-encoders. arXiv preprint arXiv:1611.07308 (2016).
[13]
Johannes Klicpera, Aleksandar Bojchevski, and Stephan Günnemann. 2018. Predict then propagate: Graph neural networks meet personalized pagerank. arXiv preprint arXiv:1810.05997 (2018).
[14]
Johannes Klicpera, Stefan Weißenberger, and Stephan Günnemann. 2019. Diffusion improves graph learning. arXiv preprint arXiv:1911.05485 (2019).
[15]
Michihiro Kuramochi and George Karypis. 2005. Finding frequent patterns in a large sparse graph. Data mining and knowledge discovery, Vol. 11, 3 (2005), 243--271.
[16]
Qimai Li, Zhichao Han, and Xiao-Ming Wu. 2018. Deeper insights into graph convolutional networks for semi-supervised learning. In AAAI.
[17]
Andrew Ng, Michael Jordan, and Yair Weiss. 2001. On spectral clustering: Analysis and an algorithm. NeurIPS, Vol. 14 (2001), 849--856.
[18]
Shirui Pan, Ruiqi Hu, Guodong Long, Jing Jiang, Lina Yao, and Chengqi Zhang. 2018. Adversarially regularized graph autoencoder for graph embedding. In IJCAI. 2609--2615.
[19]
Bryan Perozzi, Rami Al-Rfou, and Steven Skiena. 2014. Deepwalk: Online learning of social representations. In SIGKDD. 701--710.
[20]
Emanuele Rossi, Fabrizio Frasca, Ben Chamberlain, Davide Eynard, Michael Bronstein, and Federico Monti. 2020. SIGN: Scalable Inception Graph Neural Networks. arXiv preprint arXiv:2004.11198 (2020).
[21]
Oleksandr Shchur, Maximilian Mumme, Aleksandar Bojchevski, and Stephan Günnemann. 2018. Pitfalls of graph neural network evaluation. arXiv preprint arXiv:1811.05868 (2018).
[22]
Indro Spinelli, Simone Scardapane, and Aurelio Uncini. 2020. Adaptive propagation graph convolutional network. IEEE Transactions on Neural Networks and Learning Systems (2020).
[23]
Ke Sun, Zhouchen Lin, and Zhanxing Zhu. 2020. Multi-stage self-supervised learning for graph convolutional networks on graphs with few labeled nodes. In AAAI, Vol. 34. 5892--5899.
[24]
Petar Velivc ković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. 2017. Graph attention networks. arXiv preprint arXiv:1710.10903 (2017).
[25]
Chun Wang, Shirui Pan, Ruiqi Hu, Guodong Long, Jing Jiang, and Chengqi Zhang. 2019 a. Attributed Graph Clustering: a Deep Attentional Embedding approach. In IJCAI. AAAI, 3670--3676.
[26]
Chun Wang, Shirui Pan, Guodong Long, Xingquan Zhu, and Jing Jiang. 2017. Mgae: Marginalized graph autoencoder for graph clustering. In CIKM. 889--898.
[27]
Yishu Wang, Ye Yuan, Yuliang Ma, and Guoren Wang. 2019 b. Time-dependent graphs: Definitions, applications, and algorithms. Data Science and Engineering, Vol. 4, 4 (2019), 352--366.
[28]
Felix Wu, Tianyi Zhang, Amauri Holanda de Souza Jr, Christopher Fifty, Tao Yu, and Kilian Q Weinberger. 2019. Simplifying graph convolutional networks. arXiv preprint arXiv:1902.07153 (2019).
[29]
Shiwen Wu, Yuanxing Zhang, Chengliang Gao, Kaigui Bian, and Bin Cui. 2020. GARG: Anonymous Recommendation of Point-of-Interest in Mobile Networks by Graph Convolution Network. Data Science and Engineering, Vol. 5, 4 (2020), 433--447.
[30]
Keyulu Xu, Chengtao Li, Yonglong Tian, Tomohiro Sonobe, Ken-ichi Kawarabayashi, and Stefanie Jegelka. 2018. Representation learning on graphs with jumping knowledge networks. In ICML. PMLR, 5453--5462.
[31]
Yuning You, Tianlong Chen, Zhangyang Wang, and Yang Shen. 2020. When does self-supervision help graph convolutional networks?. In ICML. PMLR, 10871--10880.
[32]
Hanqing Zeng, Hongkuan Zhou, Ajitesh Srivastava, Rajgopal Kannan, and Viktor Prasanna. 2019. GraphSAINT: Graph Sampling Based Inductive Learning Method. In ICLR.
[33]
Wentao Zhang, Jiawei Jiang, Yingxia Shao, and Bin Cui. 2020 a. Efficient Diversity-Driven Ensemble for Deep Neural Networks. In 2020 IEEE 36th International Conference on Data Engineering (ICDE). IEEE, 73--84.
[34]
Wentao Zhang, Jiawei Jiang, Yingxia Shao, and Bin Cui. 2020 b. Snapshot boosting: a fast ensemble framework for deep neural networks. Sci. China Inf. Sci., Vol. 63, 1 (2020), 112102.
[35]
Wentao Zhang, Xupeng Miao, Yingxia Shao, Jiawei Jiang, Lei Chen, Olivier Ruas, and Bin Cui. 2020 c. Reliable Data Distillation on Graph Convolutional Network. In Proceedings of the 2020 ACM SIGMOD International Conference on Management of Data. 1399--1414.
[36]
Xiaotong Zhang, Han Liu, Qimai Li, and Xiao Ming Wu. 2019. Attributed graph clustering via adaptive graph convolution. In IJCAI. IJCAI, 4327--4333.
[37]
Ying Zhang, Tao Xiang, Timothy M Hospedales, and Huchuan Lu. 2018. Deep mutual learning. In ICCV. 4320--4328.
[38]
Yang Zou, Zhiding Yu, Xiaofeng Liu, BVK Kumar, and Jinsong Wang. 2019. Confidence regularized self-training. In ICCV. 5982--5991.

Cited By

View all
  • (2025)Knowledge Distillation on Graphs: A SurveyACM Computing Surveys10.1145/3711121Online publication date: 30-Jan-2025
  • (2025)Revisiting multi-dimensional classification from a dimension-wise perspectiveFrontiers of Computer Science: Selected Publications from Chinese Universities10.1007/s11704-023-3272-919:1Online publication date: 1-Jan-2025
  • (2024)Self-Attention Enhanced Auto-Encoder for Link Weight Prediction With Graph CompressionIEEE Transactions on Network Science and Engineering10.1109/TNSE.2023.329253511:1(89-99)Online publication date: Jan-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
KDD '21: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining
August 2021
4259 pages
ISBN:9781450383325
DOI:10.1145/3447548
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 14 August 2021

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. graph neural networks
  2. online distillation
  3. reception field

Qualifiers

  • Research-article

Funding Sources

  • National Key Research and Develop-ment Program of China
  • NSFC

Conference

KDD '21
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

Upcoming Conference

KDD '25

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)40
  • Downloads (Last 6 weeks)3
Reflects downloads up to 08 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2025)Knowledge Distillation on Graphs: A SurveyACM Computing Surveys10.1145/3711121Online publication date: 30-Jan-2025
  • (2025)Revisiting multi-dimensional classification from a dimension-wise perspectiveFrontiers of Computer Science: Selected Publications from Chinese Universities10.1007/s11704-023-3272-919:1Online publication date: 1-Jan-2025
  • (2024)Self-Attention Enhanced Auto-Encoder for Link Weight Prediction With Graph CompressionIEEE Transactions on Network Science and Engineering10.1109/TNSE.2023.329253511:1(89-99)Online publication date: Jan-2024
  • (2024)Accelerating Scalable Graph Neural Network Inference with Node-Adaptive Propagation2024 IEEE 40th International Conference on Data Engineering (ICDE)10.1109/ICDE60146.2024.00236(3042-3055)Online publication date: 13-May-2024
  • (2024)BIM: Improving Graph Neural Networks with Balanced Influence Maximization2024 IEEE 40th International Conference on Data Engineering (ICDE)10.1109/ICDE60146.2024.00228(2931-2944)Online publication date: 13-May-2024
  • (2024)Breaking the Entanglement of Homophily and Heterophily in Semi-supervised Node Classification2024 IEEE 40th International Conference on Data Engineering (ICDE)10.1109/ICDE60146.2024.00188(2379-2392)Online publication date: 13-May-2024
  • (2022)K-core decomposition on super large graphs with limited resourcesProceedings of the 37th ACM/SIGAPP Symposium on Applied Computing10.1145/3477314.3507064(413-422)Online publication date: 25-Apr-2022
  • (2022)Zoomer: Boosting Retrieval on Web-scale Graphs by Regions of Interest2022 IEEE 38th International Conference on Data Engineering (ICDE)10.1109/ICDE53745.2022.00212(2224-2236)Online publication date: May-2022
  • (2022)A geometric deep learning framework for drug repositioning over heterogeneous information networksBriefings in Bioinformatics10.1093/bib/bbac38423:6Online publication date: 19-Sep-2022
  • (2021)Node dependent local smoothing for scalable graph learningProceedings of the 35th International Conference on Neural Information Processing Systems10.5555/3540261.3541815(20321-20332)Online publication date: 6-Dec-2021

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media