Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3488560.3498474acmconferencesArticle/Chapter ViewAbstractPublication PageswsdmConference Proceedingsconference-collections
research-article

Understanding and Improvement of Adversarial Training for Network Embedding from an Optimization Perspective

Published: 15 February 2022 Publication History

Abstract

Network Embedding aims to learn a function mapping the nodes to Euclidean space contribute to multiple learning analysis tasks on networks. However, both the noisy information behind the real-world networks and the overfitting problem negatively impact the quality of embedding vectors. To tackle these problems, researchers utilize Adversarial Perturbations on Parameters (APP) and achieve state-of-the-art performance. Unlike the mainstream methods introducing perturbations on the network structure or the data feature, Adversarial Training for Network Embedding (AdvTNE) adopts APP to directly perturb the model parameters, thus providing a new chance to understand the mechanism behind it. In this paper, we explain APP theoretically from an optimization perspective. Considering the Power-law property of networks and the optimization objective, we analyze the reason for its remarkable results on network embedding. Based on the above analysis and the Sigmoid saturation region problem, we propose a new Sine-base activation to enhance the performance of AdvTNE. We conduct extensive experiments on four real networks to validate the effectiveness of our method in node classification and link prediction. The results demonstrate that our method is competitive with state-of-the-art methods.

Supplementary Material

MP4 File (wsdmfp522.mp4)
Adversarial training on parameter space of network embedding method (AdvTNE) has achieved impressive performance. In this work we give an in-depth analysis of this method. We firstly observe that the optimal embedding similarities of positive node pairs concentrate on the saturation region of Sigmoid, which will damage the quality of learned node embeddings. We then formulate AdvTNE as a Momentum-like optimization strategy. Theoretically and experimental results both show the effectiveness of additional Momentum information in AdvTNE. Finally, we propose a Sine-base activation to tackle the saturation region problem of Sigmoid and the newly designed loss helps model achieve SOTA performance.

References

[1]
Amr Ahmed, Nino Shervashidze, Shravan Narayanamurthy, Vanja Josifovski, and Alexander J Smola. 2013. Distributed large-scale natural graph factorization. In Proceedings of the 22nd international conference on World Wide Web. 37--48.
[2]
James Bergstra and Yoshua Bengio. 2012. Random search for hyper-parameter optimization. The Journal of Machine Learning Research (2012).
[3]
Shaosheng Cao, Wei Lu, and Qiongkai Xu. 2015. GraRep: Learning Graph Representations with Global Structural Information. In ACM International Conference on Information and Knowledge Management . 891--900.
[4]
Xu Chen, Lun Du, Mengyuan Chen, Yun Wang, Qingqing Long, and Kunqing Xie. 2021 a. Fast Hierarchy Preserving Graph Embedding via Subspace Constraints. In IEEE International Conference on Acoustics, Speech and Signal Processing .
[5]
Xu Chen, Junshan Wang, and Kunqing Xie. 2021 b. TrafficStream: A Streaming Traffic Flow Forecasting Framework Based on Graph Neural Networks and Continual Learning. arXiv preprint arXiv:2106.06273 (2021).
[6]
Xu Chen, Yuanxing Zhang, Lun Du, Zheng Fang, Yi Ren, Kaigui Bian, and Kunqing Xie. 2020. Tssrgcn: Temporal spectral spatial retrieval graph convolutional network for traffic flow forecasting. In 2020 IEEE International Conference on Data Mining (ICDM). IEEE, 954--959.
[7]
Aaron Clauset, Cosma Rohilla Shalizi, and Mark EJ Newman. 2009. Power-law distributions in empirical data. SIAM review, Vol. 51, 4 (2009), 661--703.
[8]
Diane J Cook and Lawrence B Holder. 2006. Mining graph data .John Wiley & Sons.
[9]
Hanjun Dai, Hui Li, Tian Tian, Xin Huang, Lin Wang, Jun Zhu, and Le Song. 2018. Adversarial attack on graph structured data. arXiv preprint arXiv:1806.02371 (2018).
[10]
Quanyu Dai, Xiao Shen, Liang Zhang, Qiang Li, and Dan Wang. 2019. Adversarial Training Methods for Network Embedding. In The World Wide Web Conference . ACM, 329--339.
[11]
Zhijie Deng, Yinpeng Dong, and Jun Zhu. 2019. Batch virtual adversarial training for graph convolutional networks. arXiv preprint arXiv:1902.09192 (2019).
[12]
Jennifer Drexler and James Glass. 2018. Combining end-to-end and adversarial training for low-resource speech recognition. In 2018 IEEE Spoken Language Technology Workshop (SLT). IEEE, 361--368.
[13]
Lun Du, Fei Gao, Xu Chen, Ran Jia, Junshan Wang, Jiang Zhang, Shi Han, and Dongmei Zhang. 2021. TabularNet: A Neural Network Architecture for Understanding Semantic Structures of Tabular Data. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining .
[14]
Lun Du, Guojie Song, Yiming Wang, Jipeng Huang, Mengfei Ruan, and Zhanyuan Yu. 2018. Traffic Events Oriented Dynamic Traffic Assignment Model for Expressway Network: A Network Flow Approach. IEEE Intelligent Transportation Systems Magazine (2018).
[15]
Fuli Feng, Xiangnan He, Jie Tang, and Tat-Seng Chua. 2019. Graph adversarial training: Dynamically regularizing based on graph structure. IEEE Transactions on Knowledge and Data Engineering (2019).
[16]
Pierre Foret, Ariel Kleiner, Hossein Mobahi, and Behnam Neyshabur. 2020. Sharpness-Aware Minimization for Efficiently Improving Generalization. arXiv preprint arXiv:2010.01412 (2020).
[17]
Hongchang Gao, Jian Pei, and Heng Huang. 2019. Progan: Network embedding via proximity generative adversarial network. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining .
[18]
Anna Lisa Gentile, Petar Ristoski, Steffen Eckel, Dominique Ritze, and Heiko Paulheim. 2017. Entity Matching on Web Tables: a Table Embeddings approach for Blocking. In EDBT . 510--513.
[19]
Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy. 2014. Explaining and harnessing adversarial examples. In International Conference on Learning Representations .
[20]
Aditya Grover and Jure Leskovec. 2016. node2vec: Scalable feature learning for networks. In ACM Knowledge Discovery and Data Mining. ACM, 855--864.
[21]
Will Hamilton, Zhitao Ying, and Jure Leskovec. 2017. Inductive representation learning on large graphs. In Advances in Neural Information Processing Systems. 1025--1035.
[22]
Huiting Hong, Xin Li, and Mingzhong Wang. 2019. Gane: A generative adversarial network embedding. IEEE transactions on neural networks and learning systems (2019).
[23]
Hongwei Jin and Xinhua Zhang. 2019. Latent adversarial training of graph convolution networks. In ICML Workshop on Learning and Reasoning with Graph-Structured Representations .
[24]
Ivan Kryven. 2017. General expression for the component size distribution in infinite configuration networks. Phys. Rev. E, Vol. 95 (May 2017), 052303. Issue 5. https://doi.org/10.1103/PhysRevE.95.052303
[25]
Omer Levy and Yoav Goldberg. 2014a. Linguistic regularities in sparse and explicit word representations. In Proceedings of the eighteenth conference on computational natural language learning . 171--180.
[26]
Omer Levy and Yoav Goldberg. 2014b. Neural word embedding as implicit matrix factorization. In Advances in neural information processing systems. 2177--2185.
[27]
Alexander H Liu, Hung-yi Lee, and Lin-shan Lee. 2019. Adversarial training of end-to-end speech recognition using a criticizing language model. In ICASSP 2019--2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 6176--6180.
[28]
Xiaodong Liu, Hao Cheng, Pengcheng He, Weizhu Chen, Yu Wang, Hoifung Poon, and Jianfeng Gao. 2020. Adversarial training for large neural language models. arXiv preprint arXiv:2004.08994 (2020).
[29]
Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. 2017. Towards deep learning models resistant to adversarial attacks. arXiv preprint arXiv:1706.06083 (2017).
[30]
Tomas Mikolov, Ilya Sutskever, Kai Chen, Gregory S Corrado, and Jeffrey Dean. 2013. Distributed Representations of Words and Phrases and their Compositionality. Neural Information Processing Systems (2013), 3111--3119.
[31]
Takeru Miyato, Shin-ichi Maeda, Masanori Koyama, and Shin Ishii. 2018. Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE transactions on pattern analysis and machine intelligence (2018).
[32]
Takeru Miyato, Shin-ichi Maeda, Masanori Koyama, Ken Nakae, and Shin Ishii. 2015. Distributional smoothing with virtual adversarial training. In International Conference on Learning Representations .
[33]
Bryan Perozzi, Rami Alrfou, and Steven Skiena. 2014. DeepWalk: online learning of social representations. In ACM Knowledge Discovery and Data Mining. 701--710.
[34]
Sebastian Ruder. 2016. An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016).
[35]
Barbara Rychalska, Piotr Bka bel, Konrad Go?uchowski, Andrzej Micha?owski, and Jacek Dka browski. 2021. Cleora: A Simple, Strong and Scalable Graph Embedding Scheme. arXiv preprint arXiv:2102.02302 (2021).
[36]
Prithviraj Sen, Galileo Namata, Mustafa Bilgic, Lise Getoor, Brian Galligher, and Tina Eliassi-Rad. 2008. Collective classification in network data. AI magazine (2008).
[37]
Ali Shafahi, Mahyar Najibi, Amin Ghiasi, Zheng Xu, John Dickerson, Christoph Studer, Larry S Davis, Gavin Taylor, and Tom Goldstein. 2019. Adversarial training for free! arXiv preprint arXiv:1904.12843 (2019).
[38]
Uri Shaham, Yutaro Yamada, and Sahand Negahban. 2018. Understanding adversarial training: Increasing local stability of supervised models through robust optimization. Neurocomputing, Vol. 307 (2018), 195--204.
[39]
Vincent Sitzmann, Julien NP Martel, Alexander W Bergman, David B Lindell, and Gordon Wetzstein. 2020. Implicit neural representations with periodic activation functions. arXiv preprint arXiv:2006.09661 (2020).
[40]
Guojie Song, Yuanhao Li, Junshan Wang, and Lun Du. 2020. Inferring explicit and implicit social ties simultaneously in mobile social networks. Science China Information Sciences (2020).
[41]
Guojie Song, Yun Wang, Lun Du, Yi Li, and Junshan Wang. 2021. Network Embedding on Hierarchical Community Structure Network. ACM Trans. Knowl. Discov. Data (2021).
[42]
Sining Sun, Ching-Feng Yeh, Mei-Yuh Hwang, Mari Ostendorf, and Lei Xie. 2018. Domain adversarial training for accented speech recognition. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 4854--4858.
[43]
Jian Tang, Meng Qu, Mingzhe Wang, Ming Zhang, Jun Yan, and Qiaozhu Mei. 2015. LINE: Large-scale Information Network Embedding. In International Conference on World Wide Web . 1067--1077.
[44]
Hongwei Wang, Jia Wang, Jialin Wang, Miao Zhao, Weinan Zhang, Fuzheng Zhang, Xing Xie, and Minyi Guo. 2018. Graphgan: Graph representation learning with generative adversarial nets. In Proceedings of the AAAI conference on artificial intelligence, Vol. 32.
[45]
Junshan Wang, Zhicong Lu, Guojia Song, Yue Fan, Lun Du, and Wei Lin. 2019 c. Tag2vec: Learning tag representations in tag networks. In The World Wide Web Conference . 3314--3320.
[46]
Xiaoyun Wang, Xuanqing Liu, and Cho-Jui Hsieh. 2019 b. GraphDefense: Towards Robust Graph Convolutional Networks. arXiv preprint arXiv:1911.04429 (2019).
[47]
Yun Wang, Lun Du, Guojie Song, Xiaojun Ma, Lichen Jin, Wei Lin, and Fei Sun. 2019 a. Tag2Gauss: Learning Tag Representations via Gaussian Distribution in Tagged Networks. In IJCAI. 3799--3805.
[48]
Dongxian Wu, Shu-Tao Xia, and Yisen Wang. 2020. Adversarial weight perturbation helps robust generalization. Advances in Neural Information Processing Systems, Vol. 33 (2020).
[49]
Kaidi Xu, Hongge Chen, Sijia Liu, Pin-Yu Chen, Tsui-Wei Weng, Mingyi Hong, and Xue Lin. 2019. Topology attack and defense for graph neural networks: An optimization perspective. arXiv preprint arXiv:1906.04214 (2019).
[50]
Cheng Yang, Zhiyuan Liu, Deli Zhao, Maosong Sun, and Edward Y Chang. 2015. Network representation learning with rich text information. In IJCAI . 2111--2117.
[51]
Zhilin Yang, William Cohen, and Ruslan Salakhudinov. 2016. Revisiting semi-supervised learning with graph embeddings. In International conference on machine learning. PMLR, 40--48.
[52]
Li Zhang, Shuo Zhang, and Krisztian Balog. 2019. Table2Vec: Neural Word and Entity Embeddings for Table Population and Retrieval. In ACM SIGIR Conference on Research and Development in Information Retrieval . 1029--1032.
[53]
Yaowei Zheng, Richong Zhang, and Yongyi Mao. 2020. Regularizing Neural Networks via Adversarial Model Perturbation. arXiv preprint arXiv:2010.04925 (2020).
[54]
Kai Zhou, Tomasz P Michalak, and Yevgeniy Vorobeychik. 2019. Adversarial robustness of similarity-based link prediction. In 2019 IEEE International Conference on Data Mining (ICDM). IEEE, 926--935.
[55]
Chen Zhu, Yu Cheng, Zhe Gan, Siqi Sun, Tom Goldstein, and Jingjing Liu. 2019. Freelb: Enhanced adversarial training for language understanding. (2019).

Cited By

View all
  • (2024)TAROT: A Hierarchical Framework with Multitask co-pretraining on Semi-Structured Data Towards Effective Person-Job fitICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)10.1109/ICASSP48485.2024.10447647(12046-12050)Online publication date: 14-Apr-2024
  • (2023)Parallel Prediction Method of Knowledge Proficiency Based on Bloom’s Cognitive TheoryMathematics10.3390/math1124500211:24(5002)Online publication date: 18-Dec-2023
  • (2023)Enhancing Dropout Prediction in Distributed Educational Data Using Learning Pattern Awareness: A Federated Learning ApproachMathematics10.3390/math1124497711:24(4977)Online publication date: 16-Dec-2023
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
WSDM '22: Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining
February 2022
1690 pages
ISBN:9781450391320
DOI:10.1145/3488560
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 15 February 2022

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. adversarial training
  2. network embedding
  3. optimization method
  4. saturation region problem

Qualifiers

  • Research-article

Conference

WSDM '22

Acceptance Rates

Overall Acceptance Rate 498 of 2,863 submissions, 17%

Upcoming Conference

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)17
  • Downloads (Last 6 weeks)4
Reflects downloads up to 23 Dec 2024

Other Metrics

Citations

Cited By

View all
  • (2024)TAROT: A Hierarchical Framework with Multitask co-pretraining on Semi-Structured Data Towards Effective Person-Job fitICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)10.1109/ICASSP48485.2024.10447647(12046-12050)Online publication date: 14-Apr-2024
  • (2023)Parallel Prediction Method of Knowledge Proficiency Based on Bloom’s Cognitive TheoryMathematics10.3390/math1124500211:24(5002)Online publication date: 18-Dec-2023
  • (2023)Enhancing Dropout Prediction in Distributed Educational Data Using Learning Pattern Awareness: A Federated Learning ApproachMathematics10.3390/math1124497711:24(4977)Online publication date: 16-Dec-2023
  • (2023)Adversarial Collaborative Filtering for FreeProceedings of the 17th ACM Conference on Recommender Systems10.1145/3604915.3608771(245-255)Online publication date: 14-Sep-2023
  • (2023)A subspace constraint based approach for fast hierarchical graph embeddingWorld Wide Web10.1007/s11280-023-01177-926:5(3691-3705)Online publication date: 16-Aug-2023
  • (2022)Learning Rate PerturbationProceedings of the 31st ACM International Conference on Information & Knowledge Management10.1145/3511808.3557626(4234-4238)Online publication date: 17-Oct-2022

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media