Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3340531.3411983acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article

A Feature-Importance-Aware and Robust Aggregator for GCN

Published: 19 October 2020 Publication History
  • Get Citation Alerts
  • Abstract

    Neighborhood aggregation is a key step in Graph Convolutional Networks (GCNs) for graph representation learning. Two commonly used aggregators, sum and mean, are designed with the homophily assumption that connected nodes are likely to share the same label. However, real-world graphs are noisy and adjacent nodes do not necessarily imply similarity.Learnable aggregators are proposed in Graph Attention Network (GAT) and Learnable Graph Convolutional Layer (LGCL). However, GAT considers node importance but not the importance of different features. The convolution aggregator in LGCL considers feature importance but it can not directly operate on graphs due to the irregular connectivity and lack of orderliness. In this paper, we firstly unify the current learnable aggregators in a framework: Learnable Aggregator for GCN (LA-GCN) by introducing a shared auxiliary model that provides a customized schema in neighborhood aggregation. Under this framework, we propose a new model called LA-GCNMask consisting of a new aggregator function,mask aggregator. The auxiliary model learns a specific mask for each neighbor of a given node, allowing both node-level and feature-level attention. This mechanism learns to assign different importance to both nodes and features for prediction, which provides interpretable explanations for prediction and increases the model robustness. Experiments on seven graphs for node classification and graph classification tasks show that LA-GCNMask outperforms the state-of-the-art methods. Moreover, our aggregator can identify both the important nodes and node features simultaneously, which provides a quantified understanding of the relationship between input nodes and the prediction. We further conduct experiments on noisy graphs to evaluate the robustness of our model. Experiments show that LA-GCNMask consistently outperforms the state-of-the-art methods, with up to 15% improvements in terms of accuracy compared to the second best.

    Supplementary Material

    MP4 File (3340531.3411983.mp4)
    Neighborhood aggregation is a key step in GCNs for graph representation learning. Two commonly used aggregators, sum and mean, are designed with the homophily assumption that connected nodes are likely to share the same label. However, real-world graphs are noisy and adjacent nodes do not necessarily imply similarity. In this paper, we firstly unify the current learn- able aggregators in a framework: Learnable Aggregator for GCN (LA-GCN) by introducing a shared auxiliary model that provides a customized schema in neighborhood aggregation. Under this framework, we propose a new model called LA-GCN Mask consisting of a new aggregator function, mask aggregator. The auxiliary model learns a specific mask for each neighbor of a given node, allowing both node-level and feature-level attention. This mechanism learns to assign different importance to both nodes and features for prediction, which provides interpretable explanations for prediction and increases the model robustness.

    References

    [1]
    Sami Abu-El-Haija, Bryan Perozzi, Amol Kapoor, Hrayr Harutyunyan, Nazanin Alipourfard, Kristina Lerman, Greg Ver Steeg, and Aram Galstyan. 2019. MixHop: Higher-Order Graph Convolutional Architectures via Sparsified Neighborhood Mixing. In ICML.
    [2]
    James Atwood and Don Towsley. 2016. Diffusion-convolutional Neural Networks. In NeurIPS.
    [3]
    Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural Machine Translation by Jointly Learning to Align and Translate. In ICLR.
    [4]
    Peter W Battaglia, Jessica B Hamrick, Victor Bapst, Alvaro Sanchez-Gonzalez, Vinicius Zambaldi, Mateusz Malinowski, Andrea Tacchetti, David Raposo, Adam Santoro, Ryan Faulkner, et al. 2018. Relational inductive biases, deep learning, and graph networks. arXiv preprint arXiv:1806.01261 (2018).
    [5]
    Mikhail Belkin, Partha Niyogi, and Vikas Sindhwani. 2006. Manifold regularization: A geometric framework for learning from labeled and unlabeled examples. Journal of machine learning research (2006).
    [6]
    Joan Bruna, Wojciech Zaremba, Arthur Szlam, and Yann LeCun. 2014. Spectral networks and locally connected networks on graphs. In ICLR.
    [7]
    Jie Chen, Tengfei Ma, and Cao Xiao. 2018. FastGCN: Fast Learning with Graph Convolutional Networks via Importance Sampling. In ICLR.
    [8]
    Hanjun Dai, Hui Li, Tian Tian, Xin Huang, Lin Wang, Jun Zhu, and Le Song. 2019. Adversarial Attack on Graph Structured Data. In International Conference on Machine Learning (ICML).
    [9]
    Michaël Defferrard, Xavier Bresson, and Pierre Vandergheynst. 2016. Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering. In NeurIPS.
    [10]
    Hongyang Gao, Zhengyang Wang, and Shuiwang Ji. 2018. Large-Scale Learnable Graph Convolutional Networks. In SIGKDD.
    [11]
    Justin Gilmer, Samuel S Schoenholz, Patrick F Riley, Oriol Vinyals, and George E Dahl. 2017. Neural message passing for quantum chemistry. In ICML.
    [12]
    Michele Gori, Gabriele Monfardini, and Franco Scarselli. 2005. A new model for learning in graph domains. Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005. 2 (2005), 729--734.
    [13]
    Aditya Grover and Jure Leskovec. 2016. node2vec: Scalable feature learning for networks. In SIGKDD.
    [14]
    Will Hamilton, Zhitao Ying, and Jure Leskovec. 2017. Inductive representation learning on large graphs. In NeurIPS.
    [15]
    William L. Hamilton, Rex Ying, and Jure Leskovec. 2017. Representation Learning on Graphs: Methods and Applications. IEEE Data Eng. Bull. (2017).
    [16]
    Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long Short-Term Memory. Neural Comput. (1997).
    [17]
    Kurt Hornik. 1991. Approximation capabilities of multilayer feedforward networks. Neural networks (1991).
    [18]
    Kurt Hornik, Maxwell Stinchcombe, and Halbert White. 1989. Multilayer feedforward networks are universal approximators. Neural networks (1989).
    [19]
    Sergey Ivanov and Evgeny Burnaev. 2018. Anonymous Walk Embeddings. In ICML.
    [20]
    Diederik P Kingma and Jimmy Ba. 2015. Adam: A method for stochastic optimization. In ICLR.
    [21]
    Thomas N Kipf and MaxWelling. 2017. Semi-supervised classification with graph convolutional networks. In ICLR.
    [22]
    Johannes Klicpera, Stefan Weißenberger, and Stephan Günnemann. 2019. Diffusion Improves Graph Learning. In NeurIPS.
    [23]
    Yann LeCun, Léon Bottou, Yoshua Bengio, Patrick Haffner, et al. 1998. Gradientbased learning applied to document recognition. Proc. IEEE (1998).
    [24]
    John Boaz Lee, Ryan A Rossi, Sungchul Kim, Nesreen K Ahmed, and Eunyee Koh. 2019. Attention models in graphs: A survey. ACM Transactions on Knowledge Discovery from Data (TKDD) 13, 6 (2019), 1--25.
    [25]
    Andrew Kachites McCallum, Kamal Nigam, Jason Rennie, and Kristie Seymore. 2000. Automating the construction of internet portals with machine learning. Information Retrieval 3, 2 (2000), 127--163.
    [26]
    Ryan L. Murphy, Balasubramaniam Srinivasan, Vinayak Rao, and Bruno Ribeiro. 2019. Janossy Pooling: Learning Deep Permutation-Invariant Functions for Variable-Size Inputs. In ICLR.
    [27]
    Mathias Niepert, Mohamed Ahmed, and Konstantin Kutzkov. 2016. Learning convolutional neural networks for graphs. In ICML.
    [28]
    Meng Qu, Yoshua Bengio, and Jian Tang. 2019. GMNN: Graph Markov Neural Networks. In ICML.
    [29]
    Franco Scarselli, Marco Gori, Ah Chung Tsoi, Markus Hagenbuchner, and Gabriele Monfardini. 2009. The Graph Neural Network Model. IEEE Transactions on Neural Networks (2009).
    [30]
    Prithviraj Sen, Galileo Namata, Mustafa Bilgic, Lise Getoor, Brian Galligher, and Tina Eliassi-Rad. 2008. Collective classification in network data. AI magazine (2008).
    [31]
    Otilia Stretcu, Krishnamurthy Viswanathan, Dana Movshovitz-Attias, Emmanouil Platanios, Sujith Ravi, and Andrew Tomkins. 2019. Graph Agreement Models for Semi-Supervised Learning. In NeurIPS.
    [32]
    Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. In NeurIPS.
    [33]
    Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. 2018. Graph attention networks. In ICLR.
    [34]
    Edward Wagstaff, Fabian B. Fuchs, Martin Engelcke, Ingmar Posner, and Michael A. Osborne. 2019. On the Limitations of Representing Functions on Sets. In ICML.
    [35]
    B. Yu. Weisfeiler and A. A. Leman. 1968. Reduction of a graph to a canonical form and an algebra arising during this reduction.
    [36]
    Jason Weston, Frédéric Ratle, Hossein Mobahi, and Ronan Collobert. 2012. Deep learning via semi-supervised embedding. In Neural networks: Tricks of the trade. Springer, 639--655.
    [37]
    Keyulu Xu,Weihua Hu, Jure Leskovec, and Stefanie Jegelka. 2019. How Powerful are Graph Neural Networks?. In ICLR.
    [38]
    Keyulu Xu, Chengtao Li, Yonglong Tian, Tomohiro Sonobe, Ken-ichi Kawarabayashi, and Stefanie Jegelka. 2018. Representation Learning on Graphs with Jumping Knowledge Networks. In ICML.
    [39]
    Pinar Yanardag and SVN Vishwanathan. 2015. Deep graph kernels. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 1365--1374.
    [40]
    Rex Ying, Dylan Bourgeois, Jiaxuan You, Marinka Zitnik, and Jure Leskovec. 2019. GNN Explainer: A Tool for Post-hoc Explanation of Graph Neural Networks. In NeurIPS.
    [41]
    Manzil Zaheer, Satwik Kottur, Siamak Ravanbakhsh, Barnabás Póczos, Ruslan Salakhutdinov, and Alexander J. Smola. 2017. Deep Sets. In NeurIPS.
    [42]
    Li Zhang, Heda Song, and Haiping Lu. 2018. Graph node-feature convolution for representation learning. arXiv preprint arXiv:1812.00086 (2018).
    [43]
    Muhan Zhang, Zhicheng Cui, Marion Neumann, and Yixin Chen. 2018. An End-to-End Deep Learning Architecture for Graph Classification. In AAAI.
    [44]
    Ying Zhang, Tao Xiang, Timothy M Hospedales, and Huchuan Lu. 2018. Deep mutual learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 4320--4328.
    [45]
    Xiaojin Zhu, Zoubin Ghahramani, and John D Lafferty. 2003. Semi-supervised learning using gaussian fields and harmonic functions. In ICML.
    [46]
    Daniel Zügner, Amir Akbarnejad, and Stephan Günnemann. 2018. Adversarial attacks on neural networks for graph data. In SIGKDD.
    [47]
    Daniel Zügner and Stephan Günnemann. 2019. Certifiable robustness and robust training for graph convolutional networks. In SIGKDD.

    Cited By

    View all
    • (2024)On the Use of Spatial Graphs for Performance Degradation Root-Cause Analysis Toward Self-Healing Mobile NetworksIEEE Access10.1109/ACCESS.2024.336128412(20490-20508)Online publication date: 2024
    • (2024)KGIE: Knowledge graph convolutional network for recommender system with interactive embeddingKnowledge-Based Systems10.1016/j.knosys.2024.111813295(111813)Online publication date: Jul-2024
    • (2023)Robust Mid-Pass Filtering Graph Convolutional NetworksProceedings of the ACM Web Conference 202310.1145/3543507.3583335(328-338)Online publication date: 30-Apr-2023
    • Show More Cited By

    Index Terms

    1. A Feature-Importance-Aware and Robust Aggregator for GCN

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      CIKM '20: Proceedings of the 29th ACM International Conference on Information & Knowledge Management
      October 2020
      3619 pages
      ISBN:9781450368599
      DOI:10.1145/3340531
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 19 October 2020

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. feauture-level attention
      2. graph convolutional networks
      3. mask aggregator

      Qualifiers

      • Research-article

      Funding Sources

      • UK Engineering and Physical Sciences Research Council (EPSRC)
      • China Scholarship Council (CSC)

      Conference

      CIKM '20
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 1,861 of 8,427 submissions, 22%

      Upcoming Conference

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)99
      • Downloads (Last 6 weeks)17

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)On the Use of Spatial Graphs for Performance Degradation Root-Cause Analysis Toward Self-Healing Mobile NetworksIEEE Access10.1109/ACCESS.2024.336128412(20490-20508)Online publication date: 2024
      • (2024)KGIE: Knowledge graph convolutional network for recommender system with interactive embeddingKnowledge-Based Systems10.1016/j.knosys.2024.111813295(111813)Online publication date: Jul-2024
      • (2023)Robust Mid-Pass Filtering Graph Convolutional NetworksProceedings of the ACM Web Conference 202310.1145/3543507.3583335(328-338)Online publication date: 30-Apr-2023
      • (2023)Data Representation via Attribute Selection- Propagation Neural NetworkIEEE Transactions on Signal and Information Processing over Networks10.1109/TSIPN.2023.32669499(258-267)Online publication date: 2023
      • (2023)DPGNNKnowledge-Based Systems10.1016/j.knosys.2023.110377268:COnline publication date: 23-May-2023
      • (2022)Are defenses for graph neural networks robust?Proceedings of the 36th International Conference on Neural Information Processing Systems10.5555/3600270.3600921(8954-8968)Online publication date: 28-Nov-2022
      • (2022)UD-GNNProceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining10.1145/3534678.3539483(1131-1140)Online publication date: 14-Aug-2022
      • (2022)A GNN-based Multi-task Learning Framework for Personalized Video SearchProceedings of the Fifteenth ACM International Conference on Web Search and Data Mining10.1145/3488560.3498507(1386-1394)Online publication date: 11-Feb-2022
      • (2022)Learning Privacy-Preserving Graph Convolutional Network with Partially Observed Sensitive AttributesProceedings of the ACM Web Conference 202210.1145/3485447.3511975(3552-3561)Online publication date: 25-Apr-2022
      • (2022)EvAnGCN: Evolving Graph Deep Neural Network Based Anomaly Detection in BlockchainAdvanced Data Mining and Applications10.1007/978-3-031-22064-7_32(444-456)Online publication date: 30-Nov-2022
      • Show More Cited By

      View Options

      Get Access

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media