Abstract
Graph neural networks (GNN) have seen significant growth recently for modeling temporal evolution in dynamic networks. Representation of complex networks in the form of graph data structures has enabled researchers to study how entities within these networks interact with each other. These interactions evolve over time. Developing a generic methodology for modeling this temporal evolution in complex networks for tracking evolving relationships has been a significant challenge. Most of the existing methods fail to extract contextual representations of historical neighborhood interactions for future link prediction. To address these challenges, this paper presents a novel method for modeling temporal evolution in complex networks using GNNs. A Context-Aware Graph Temporal Neural Network (CATGNN) method that uses dynamic neighborhood selection based on common neighbors for a given node is presented. The method uses dynamic neighborhood selection using contextual embeddings extracted from the historical interactions of the down-sampled set of neighbors of a central node based on a common neighborhood. Fixed-sized contextual memory modules are constructed for each node that store the historical interactions of its neighbors and are updated based on the recency and significance of interactions. The proposed method has been evaluated using six real-world datasets and has comparable performance against state-of-the-art methods, both in terms of accuracy and efficiency. It shows an improvement of 7.52 to 0.05% over the baselines in terms of average precision. The results demonstrate that the proposed CATGNN model can capture complex patterns of change that are difficult to identify using traditional techniques by propagating information over the graph structure. The model can be applied in various fields involving complex systems.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data Availability
The datasets used in this study are publicly available and can be accessed through the links and references provided in relevant sections in the paper. The project number HU24K14859 Hosei University Japan is acknowledged.
References
Chen Q, Jiang F, Guo X, Chen J, Sha K, Wang Y. Combine temporal information in session-based recommendation with graph neural networks. Expert Syst Appl. 2024;238:121969. https://doi.org/10.1016/j.eswa.2023.121969.
del Mondo G, Stell JG, Claramunt C, Thibaud R. A graph model for spatio-temporal evolution. J Univers Comput Sci. 2010;16(11):1452–77.
Li Z, Wang C, Wang X, Chen Z, Li J. HJE: joint convolutional representation learning for knowledge hypergraph completion. IEEE Trans Knowl Data Eng. 2024;36(8):3879–92. https://doi.org/10.1109/TKDE.2024.3365727.
Gong Y, Lv X, Yuan Z, You X, Hu F, Chen Y. GNN-based multimodal named entity recognition. Comput J. 2024;67(8):2622–32. https://doi.org/10.1093/comjnl/bxae030.
Li H, Li C, Feng K, Yuan Y, Wang G, Zha H. Robust knowledge adaptation for dynamic graph neural networks. IEEE Trans Knowl Data Eng. 2024;36(11):6920–33. https://doi.org/10.1109/TKDE.2024.3388453.
Shabani N, et al. A comprehensive survey on graph summarization with graph neural networks. IEEE Trans Artif Intell. 2024;5(8):3780–800. https://doi.org/10.1109/TAI.2024.3350545.
Cai H, Zheng VW, Chang KC-C. A comprehensive survey of graph embedding: problems, techniques, and applications. IEEE Trans Knowl Data Eng. 2018;30(9):1616–37. https://doi.org/10.1109/TKDE.2018.2807452.
Liang F, Qian C, Yu W, Griffith D, Golmie N. Survey of graph neural networks and applications. Wirel Commun Mob Comput. 2022;2022:1–18. https://doi.org/10.1155/2022/9261537.
Barros CDT, Mendonça MRF, Vieira AB, Ziviani A. A survey on embedding dynamic graphs. ACM Comput Surv. 2021;55(1):1–37. https://doi.org/10.1145/3483595.
Thomas JM, Moallemy-Oureh A, Beddar-Wiesing S, Holzhüter C. Graph neural networks designed for different graph types: A survey. Trans Mach Learn Res 2023. https://doi.org/10.48550/arXiv.2204.03080
AlBadani B, Shi R, Dong J, Al-Sabri R, Moctard OB. Transformer-based graph convolutional network for sentiment analysis. Appl Sci. 2022;12(3):1316. https://doi.org/10.3390/app12031316.
Zhang F, Zheng W, Yang Y. Graph convolutional network with syntactic dependency for aspect-based sentiment analysis. Int J Comput Intell Syst. 2024;17(1):37. https://doi.org/10.1007/s44196-024-00419-6.
Allamanis M. Graph neural networks in program analysis. In: Wu L, Cui P, Pei J, Zhao L, editors. Graph neural networks: foundations, frontiers, and applications. Singapore: Springer Nature Singapore; 2022. p. 483–497. https://doi.org/10.1007/978-981-16-6054-2_22.
Carbonell M, Riba P, Villegas M, Fornes A, Llados J. Named entity recognition and relation extraction with graph neural networks in semi structured documents. In 25th International Conference on Pattern Recognition (ICPR). Milan, Italy: IEEE; 2021, p. 9622–9627. https://doi.org/10.1109/ICPR48806.2021.9412669.
Fan W, et al. Graph neural networks for social recommendation,” in The World Wide Web Conference. In WWW ’19. New York, NY, USA: Association for Computing Machinery; 2019, p. 417–426. https://doi.org/10.1145/3308558.3313488.
Rossi E, Chamberlain B, Frasca F, Eynard D, Monti F, Bronstein M. Temporal graph networks for deep learning on dynamic graphs. In 37th International Conference on Machine Learning Workshop on Graph Representation Learning and Beyond. Vienna, Austria; 2020.
Chakraborty A. Aspect based sentiment analysis using spectral temporal graph neural network. In Proceedings of the 18th International Conference on Natural Language Processing (ICON), National Institute of Technology Silchar. Silchar, India: NLP Association of India (NLPAI); 2021, p. 508–518.
Lu J, Shi L, Liu G, Zhan X. Dual-channel edge-featured graph attention networks for aspect-based sentiment analysis. Electronics. 2023;12(3):624. https://doi.org/10.3390/electronics12030624.
Diao Z, Wang X, Zhang D, Liu Y, Xie K, He S. Dynamic spatial-temporal graph convolutional neural networks for traffic forecasting. Proc AAAI Conf Artif Intell. 2019;33(01):890–7. https://doi.org/10.1609/aaai.v33i01.3301890.
Li J, et al. Evaluating graph neural networks for link prediction: current pitfalls and new benchmarking. In: Oh A, Naumann T, Globerson A, Saenko A, Hardt M, Levine S, editors. Advances in neural information processing systems. Curran Associates, Inc.; 2023, p. 3853–3866. [Online]. Available: https://proceedings.neurips.cc/paper_files/paper/2023/file/0be50b4590f1c5fdf4c8feddd63c4f67-Paper-Datasets_and_Benchmarks.pdf
Bai T, Zhang Y, Wu B, Nie JY. Temporal graph neural networks for social recommendation. In IEEE International Conference on Big Data. Atlanta, GA, USA: IEEE; 2020, p. 898–903. https://doi.org/10.1109/BigData50022.2020.9378444.
Fan Z, Liu Z, Zhang J, Xiong J, Zheng L, Yu PS. Continuous-time sequential recommendation with temporal graph collaborative transformer. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management. Virtual Event Queensland Australia: ACM; 2021, p. 433–442. https://doi.org/10.1145/3459637.3482242.
Guo J, Han Z, Zhou S, Li J, Tresp V, Wang Y. Continuous temporal graph networks for event-based graph data. In Proceedings of the 2nd Workshop on Deep Learning on Graphs for Natural Language Processing. Seattle, Washington: Association for Computational Linguistics; 2022, p. 22–29. https://doi.org/10.18653/v1/2022.dlg4nlp-1.3.
Bhattacharya U, Mittal T, Chandra R, Randhavane T, Bera A, Manocha D. STEP: spatial temporal graph convolutional networks for emotion perception from gaits. Proc AAAI Conf Artif Intell. Apr.2020;34(02):1342–50. https://doi.org/10.1609/aaai.v34i02.5490.
Liu Z, Wang Y, Wang S, Zhao X, Wang H, Yin H. Heterogeneous graphs neural networks based on neighbor relationship filtering. Expert Syst Appl. 2024;239:122489. https://doi.org/10.1016/j.eswa.2023.122489.
Cao D, et al. Spectral temporal graph neural network for multivariate time-series forecasting. In Proceedings of the 34th International Conference on Neural Information Processing Systems, in NIPS’20. Red Hook, NY, USA: Curran Associates Inc.; 2020.
Peng H, Zhang R, Dou Y, Yang R, Zhang J, Yu PS. Reinforced neighborhood selection guided multi-relational graph neural networks. ACM Trans Inf Syst. 2021;40(4):1–46. https://doi.org/10.1145/3490181.
Zang X, Tang B. Self-supervised Dynamic Graph Embedding with evolutionary neighborhood and community. Expert Syst Appl. 2023;228:120409. https://doi.org/10.1016/j.eswa.2023.120409.
Zhu W, Ruan K, Huang J, Xiao J, Yu W. Dynamic graph representation based on temporal and contextual contrasting. In Proceedings of the 2022 5th International Conference on Algorithms, Computing and Artificial Intelligence. Sanya China: ACM; 2022, p. 1–8. https://doi.org/10.1145/3579654.3579771.
Zhang M, Wu S, Gao M, Jiang X, Xu K, Wang L. Personalized graph neural networks with attention mechanism for session-aware recommendation. IEEE Trans Knowl Data Eng. 2022;34(8):3946–57. https://doi.org/10.1109/TKDE.2020.3031329.
Li Y, Wu Y, Sun M, Yang B, Wang Y. Learning continuous dynamic network representation with transformer-based temporal graph neural network. Inf Sci. 2023;649:119596. https://doi.org/10.1016/j.ins.2023.119596.
Wang F, Hou D, Yan H. Detecting community evolution by utilizing individual temporal semantics in social networks. IEEE Access. 2023;11:120362–71. https://doi.org/10.1109/ACCESS.2023.3328616.
Mu Z, Zhuang Y, Tang S. Contrastive Hawkes graph neural networks with dynamic sampling for event prediction. Neurocomputing. 2024;575:127265. https://doi.org/10.1016/j.neucom.2024.127265.
Gao S, Li Y, Shen Y, Shao Y, Chen L. ETC: efficient training of temporal graph neural networks over large-scale dynamic graphs. Proc VLDB Endow. 2024;17(5):1060–72. https://doi.org/10.14778/3641204.3641215.
Gao S, Li Y, Zhang X, Shen Y, Shao Y, Chen L. SIMPLE: efficient temporal graph neural network training at scale with dynamic data placement. Proc ACM Manag Data. 2024;2(3):1–25. https://doi.org/10.1145/3654977.
Zhang H, Han X, Xiao X, Bai J. Time-aware graph structure learning via sequence prediction on temporal graphs. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management. Birmingham United Kingdom: ACM; 2023, p. 3288–3297. https://doi.org/10.1145/3583780.3615081.
Wu Y, Fang Y, Liao L. On the feasibility of simple transformer for dynamic graph modeling. In Proceedings of the ACM Web Conference 2024. Singapore Singapore: ACM; 2024, p. 870–880. https://doi.org/10.1145/3589334.3645622.
Yu L, Sun L, Du B, Lv W. Towards better dynamic graph learning: new architecture and unified library. In Proceedings of the 37th International Conference on Neural Information Processing Systems, in NIPS ’23. Red Hook, NY, USA: Curran Associates Inc., 2024.
Cong W, et al. DyFormer : a scalable dynamic graph transformer with provable benefits on generalization ability. In Proceedings of the 2023 SIAM International Conference on Data Mining (SDM). 2023, p. 442–450. https://doi.org/10.1137/1.9781611977653.ch50.
Peng H, Zhang R, Dou Y, Yang R, Zhang J, Yu PS. Reinforced neighborhood selection guided multi-relational graph neural networks. ACM Trans Inf Syst. 2022;40(4):1–46. https://doi.org/10.1145/3490181.
Xie Y, Li S, Yang C, Wong RCW, Han J. When do GNNs work: understanding and improving neighborhood aggregation. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, Yokohama, Japan: International Joint Conferences on Artificial Intelligence Organization. 2020, p. 1303–1309. https://doi.org/10.24963/ijcai.2020/181.
Li Y, Shen Y, Chen L, Yuan M. Zebra: when temporal graph neural networks meet temporal personalized PageRank. Proc VLDB Endow. 2023;16(6):1332–45. https://doi.org/10.14778/3583140.3583150.
Ghafouri S, Khasteh SH, Azarkasb SO. Influence maximization (IM) in complex networks with limited visibility using statistical methods. J Supercomput. 2024;80(5):6809–54. https://doi.org/10.1007/s11227-023-05695-1.
Finkelshtein B, Huang X, Bronstein M, Ceylan II. Cooperative graph neural networks. In: Proceedings of fourty-first International Conference on Machine Learning (ICML). 2024. [Online]. Available: https://arxiv.org/abs/2310.01267. Accessed 10 Oct 2024.
Kovanen L, Karsai M, Kaski K, Kertész J, Saramäki J. Temporal motifs in time-dependent networks. J Stat Mech Theory Exp. 2011;2011(11):P11005. https://doi.org/10.1088/1742-5468/2011/11/P11005.
Longa A, et al. Graph neural networks for temporal graphs: state of the art, open challenges, and opportunities. Trans Mach Learn Res. 2023.
Kumar S, Zhang X, Leskovec J. Predicting dynamic embedding trajectory in temporal interaction networks. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. Anchorage AK USA: ACM; 2019, p. 1269–1278. https://doi.org/10.1145/3292500.3330895.
Poursafaei F, Huang S, Pelrine K, Rabbany R. Towards better evaluation for dynamic link prediction. In Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track. 2022, p. 32928–32941.
Shetty J, Adibi J. The Enron email dataset database schema and brief statistical report. Information sciences institute technical report. University of Southern California; 2004. [Online]. Available: http://www.cs.cmu.edu/~enron/. Accessed 26 Aug 2024.
Veličković P, Cucurull G, Casanova A, Romero A, Liò P, Bengio Y. Graph attention networks. In: Internation Conference on Learning and Representation. 2018. [Online]. Available: http://arxiv.org/abs/1710.10903. Accessed 17 Nov 2023.
Hamilton WL, Ying R, Leskovec J. Inductive representation learning on large graphs. In 31st Conference on Neural Information Processing Systems. CA, USA: Long Beach; 2017.
Xu D, Ruan C, Korpeoglu E, Kumar S, Achan K. Inductive representation learning on temporal graphs,” Feb. 18, 2020, arXiv: arXiv:2002.07962. [Online]. Available: http://arxiv.org/abs/2002.07962. Accessed 14 May 2023.
Wang L, et al. TCL: Transformer-based dynamic graph modelling via contrastive learning. May 17, 2021, arXiv: arXiv:2105.07944. [Online]. Available: http://arxiv.org/abs/2105.07944. Accessed 30 Aug 2024.
Cong W, et al. Do we really need complicated model architectures for temporal networks? In: The Eleventh International Conference on Learning Representations. 2023. [Online]. Available: https://openreview.net/forum?id=ayPPc0SyLv1. Accessed 30 Aug 2024.
Author information
Authors and Affiliations
Contributions
M. Ali.Zeb; conceptualization, implementation, experimentation, writing the manuscript. M. Irfan Uddin; conceptualization, visualization, critical analysis, mathematical formulation, writing manuscript. A. Abdulsalam Alarood; Mathematical analysis, supervision, reviewing manuscript. M. Shafiq.; Experimentation, resources, software, supervision, reviewing manuscript. S. Habibullah.; Visualization, Analysis, Resources, supervision, manuscript correction. A. A. Alsulami.; Conceptualization, visualization, Mathematical Analysis, Review manuscript.
Corresponding author
Ethics declarations
Conflict of Interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Zeb, M.A., Uddin, M.I., Alarood, A.A. et al. Dynamic Neighborhood Selection for Context Aware Temporal Evolution Using Graph Neural Networks. Cogn Comput 17, 22 (2025). https://doi.org/10.1007/s12559-024-10359-0
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s12559-024-10359-0