Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3477495.3531906acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
short-paper

MuchSUM: Multi-channel Graph Neural Network for Extractive Summarization

Published: 07 July 2022 Publication History
  • Get Citation Alerts
  • Abstract

    Recent studies of extractive text summarization have leveraged BERT for document encoding with breakthrough performance. However, when using a pre-trained BERT-based encoder, existing approaches for selecting representative sentences for text summarization are inadequate since the encoder is not explicitly trained for representing sentences. Simply providing the BERT-initialized sentences to cross-sentential graph-based neural networks (GNNs) to encode semantic features of the sentences is not ideal because doing so fail to integrate other summary-worthy features like sentence importance and positions. This paper presents MuchSUM, a better approach for extractive text summarization. MuchSUM is a multi-channel graph convolutional network designed to explicitly incorporate multiple salient summary-worthy features. Specifically, we introduce three specific graph channels to encode the node textual features, node centrality features, and node position features, respectively, under bipartite word-sentence heterogeneous graphs. Then, a cross-channel convolution operation is designed to distill the common graph representations shared by different channels. Finally, the sentence representations of each channel are fused for extractive summarization. We also investigate three weighted graphs in each channel to infuse edge features for graph-based summarization modeling. Experimental results demonstrate our model can achieve considerable performance compared with some BERT-initialized graph-based extractive summarization systems.

    Supplementary Material

    MP4 File (sigir_sp2120.mp4)
    Presentation video.

    References

    [1]
    Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT. Association for Computational Linguistics, 4171--4186.
    [2]
    Gü nes Erkan and Dragomir R. Radev. 2004. LexRank: Graph-based Lexical Centrality as Salience in Text Summarization. J. Artif. Intell. Res., Vol. 22 (2004), 457--479.
    [3]
    Yong Guan, Shaoru Guo, Ru Li, Xiaoli Li, and Hongye Tan. 2021. Frame Semantic-Enhanced Sentence Modeling for Sentence-level Extractive Text Summarization. In EMNLP. Association for Computational Linguistics, 4045--4052.
    [4]
    Karl Moritz Hermann, Tomá s Kociský, Edward Grefenstette, Lasse Espeholt, Will Kay, Mustafa Suleyman, and Phil Blunsom. 2015. Teaching Machines to Read and Comprehend. In NIPs. 1693--1701.
    [5]
    Ruipeng Jia, Yanan Cao, Hengzhu Tang, Fang Fang, Cong Cao, and Shi Wang. 2020. Neural Extractive Summarization with Hierarchical Attentive Heterogeneous Graph Network. In EMNLP. Association for Computational Linguistics, 3622--3631.
    [6]
    Baoyu Jing, Zeyu You, Tao Yang, Wei Fan, and Hanghang Tong. 2021. Multiplex Graph Neural Network for Extractive Text Summarization. In EMNLP. Association for Computational Linguistics, 133--139.
    [7]
    Guolin Ke, Di He, and Tie-Yan Liu. 2021. Rethinking Positional Encoding in Language Pre-training. In ICLR. OpenReview.net.
    [8]
    Thomas N. Kipf and Max Welling. 2017. Semi-Supervised Classification with Graph Convolutional Networks. In ICLR. OpenReview.net.
    [9]
    Logan Lebanoff, Kaiqiang Song, Franck Dernoncourt, Doo Soon Kim, Seokhwan Kim, Walter Chang, and Fei Liu. 2019. Scoring Sentence Singletons and Pairs for Abstractive Summarization. In ACL, Volume 1: Long Papers. Association for Computational Linguistics, 2175--2189.
    [10]
    Bohan Li, Hao Zhou, Junxian He, Mingxuan Wang, Yiming Yang, and Lei Li. 2020. On the Sentence Embeddings from Pre-trained Language Models. In EMNLP. Association for Computational Linguistics, 9119--9130.
    [11]
    Yang Liu and Mirella Lapata. 2019. Text Summarization with Pretrained Encoders. In EMNLP-IJCNLP. Association for Computational Linguistics, 3728--3738.
    [12]
    Christopher D. Manning, Mihai Surdeanu, John Bauer, Jenny Rose Finkel, Steven Bethard, and David McClosky. 2014. The Stanford CoreNLP Natural Language Processing Toolkit. In ACL, System Demonstrations . The Association for Computer Linguistics, 55--60.
    [13]
    Rada Mihalcea and Paul Tarau. 2004. TextRank: Bringing Order into Text. In EMNLP. Association for Computational Linguistics, 404--411.
    [14]
    Ramesh Nallapati, Feifei Zhai, and Bowen Zhou. 2017. SummaRuNNer: A Recurrent Neural Network Based Sequence Model for Extractive Summarization of Documents. In AAAI. AAAI Press, 3075--3081.
    [15]
    Shashi Narayan, Shay B. Cohen, and Mirella Lapata. 2018. Ranking Sentences for Extractive Summarization with Reinforcement Learning. In NAACL-HLT, Volume 1 (Long Papers). Association for Computational Linguistics, 1747--1759.
    [16]
    Lawrence Page, Sergey Brin, Rajeev Motwani, and Terry Winograd. 1999. The PageRank Citation Ranking: Bringing Order to the Web. Technical Report. Stanford InfoLab.
    [17]
    Hao Peng, Jianxin Li, Qiran Gong, Yangqiu Song, Yuanxing Ning, Kunfeng Lai, and Philip S. Yu. 2019. Fine-grained Event Categorization with Heterogeneous Graph Convolutional Networks. In IJCAI, Sarit Kraus (Ed.). ijcai.org, 3238--3245.
    [18]
    Jeffrey Pennington, Richard Socher, and Christopher D. Manning. 2014. Glove: Global Vectors for Word Representation. In EMNLP. ACL, 1532--1543.
    [19]
    Nils Reimers and Iryna Gurevych. 2019. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In EMNLP-IJCNLP, Kentaro Inui, Jing Jiang, Vincent Ng, and Xiaojun Wan (Eds.). Association for Computational Linguistics, 3980--3990.
    [20]
    Abigail See, Peter J. Liu, and Christopher D. Manning. 2017. Get To The Point: Summarization with Pointer-Generator Networks. In ACL, Volume 1: Long Papers. Association for Computational Linguistics, 1073--1083.
    [21]
    Peter Shaw, Jakob Uszkoreit, and Ashish Vaswani. 2018. Self-Attention with Relative Position Representations. In NAACL-HLT, Volume 2, Short Papers, Marilyn A. Walker, Heng Ji, and Amanda Stent (Eds.). Association for Computational Linguistics, 464--468.
    [22]
    Le Song, Alexander J. Smola, Arthur Gretton, Karsten M. Borgwardt, and Justin Bedo. 2007. Supervised feature selection via dependence estimation. In (ICML (ACM International Conference Proceeding Series), Vol. 227. ACM, 823--830.
    [23]
    Hinton G Van der Maaten L. 2008. Visualizing Data using T-SNE., Vol. 9, 11 (2008).
    [24]
    Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. In NIPs. 5998--6008.
    [25]
    Danqing Wang, Pengfei Liu, Yining Zheng, Xipeng Qiu, and Xuanjing Huang. 2020 a. Heterogeneous Graph Neural Networks for Extractive Document Summarization. In ACL. Association for Computational Linguistics, 6209--6219.
    [26]
    Danqing Wang, Pengfei Liu, Ming Zhong, Jie Fu, Xipeng Qiu, and Xuanjing Huang. 2019. Exploring Domain Shift in Extractive Text Summarization. CoRR, Vol. abs/1908.11664 (2019).
    [27]
    Xiao Wang, Meiqi Zhu, Deyu Bo, Peng Cui, Chuan Shi, and Jian Pei. 2020 b. AM-GCN: Adaptive Multi-channel Graph Convolutional Networks. In KDD. ACM, 1243--1253.
    [28]
    Jiacheng Xu, Zhe Gan, Yu Cheng, and Jingjing Liu. 2019. Discourse-Aware Neural Extractive Model for Text Summarization. CoRR, Vol. abs/1910.14142 (2019).
    [29]
    Jiacheng Xu, Zhe Gan, Yu Cheng, and Jingjing Liu. 2020 a. Discourse-Aware Neural Extractive Text Summarization. In ACL. Association for Computational Linguistics, 5021--5031.
    [30]
    Shusheng Xu, Xingxing Zhang, Yi Wu, Furu Wei, and Ming Zhou. 2020 b. Unsupervised Extractive Summarization by Pre-training Hierarchical Transformers. In EMNLP (Findings of ACL ), Vol. EMNLP 2020. Association for Computational Linguistics, 1784--1795.
    [31]
    Yuanmeng Yan, Rumei Li, Sirui Wang, Fuzheng Zhang, Wei Wu, and Weiran Xu. 2021. ConSERT: A Contrastive Framework for Self-Supervised Sentence Representation Transfer. In ACL/IJCNLP,Volume 1: Long Papers . Association for Computational Linguistics, 5065--5075.
    [32]
    Michihiro Yasunaga, Rui Zhang, Kshitijh Meelu, Ayush Pareek, Krishnan Srinivasan, and Dragomir R. Radev. 2017. Graph-based Neural Multi-Document Summarization. In CoNLL. Association for Computational Linguistics, 452--462.
    [33]
    Xingxing Zhang, Furu Wei, and Ming Zhou. 2019. HIBERT: Document Level Pre-training of Hierarchical Bidirectional Transformers for Document Summarization. In ACL, Volume 1: Long Papers . Association for Computational Linguistics, 5059--5069.
    [34]
    Hao Zheng and Mirella Lapata. 2019. Sentence Centrality Revisited for Unsupervised Summarization. In ACL, Volume 1: Long Papers . Association for Computational Linguistics, 6236--6247.
    [35]
    Ming Zhong, Pengfei Liu, Yiran Chen, Danqing Wang, Xipeng Qiu, and Xuanjing Huang. 2020. Extractive Summarization as Text Matching. In ACL. Association for Computational Linguistics, 6197--6208.
    [36]
    Ming Zhong, Pengfei Liu, Danqing Wang, Xipeng Qiu, and Xuanjing Huang. 2019. Searching for Effective Neural Extractive Summarization: What Works and What's Next. In ACL, Volume 1: Long Papers . Association for Computational Linguistics, 1049--1058.
    [37]
    Qingyu Zhou, Nan Yang, Furu Wei, Shaohan Huang, Ming Zhou, and Tiejun Zhao. 2018. Neural Document Summarization by Jointly Learning to Score and Select Sentences. In ACL, Volume 1: Long Papers. Association for Computational Linguistics, 654--663.

    Cited By

    View all
    • (2024)FuzzyTP-BERT: Enhancing extractive text summarization with fuzzy topic modeling and transformer networksJournal of King Saud University - Computer and Information Sciences10.1016/j.jksuci.2024.10208036:6(102080)Online publication date: Jul-2024

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    SIGIR '22: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval
    July 2022
    3569 pages
    ISBN:9781450387323
    DOI:10.1145/3477495
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 07 July 2022

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. bipartite word-sentence heterogeneous graph
    2. multi-channel graph
    3. text summarization

    Qualifiers

    • Short-paper

    Funding Sources

    Conference

    SIGIR '22
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 792 of 3,983 submissions, 20%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)71
    • Downloads (Last 6 weeks)3

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)FuzzyTP-BERT: Enhancing extractive text summarization with fuzzy topic modeling and transformer networksJournal of King Saud University - Computer and Information Sciences10.1016/j.jksuci.2024.10208036:6(102080)Online publication date: Jul-2024

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media