Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

From coarse to fine: : Enhancing multi-document summarization with multi-granularity relationship-based extractor

Published: 01 May 2024 Publication History

Abstract

Multi-Document Summarization (MDS) is a challenging task due to the fact that multiple documents not only have extremely long inputs but may also be overlapping, complementary, or contradictory to each other. In this paper, we propose to capture complex cross-document interactions to handle lengthy inputs for better multi-document summarization. Specifically, we present MDS-MGRE, a coarse-to-fine MDS framework that introduces Multi-Granularity Relationships into an Extract-then-summarize pipeline. In the coarse-grained stage, multi-granularity embedding, heterogeneous graph construction, and MGRExtractor work together to convert redundant multi-documents into compact meta-documents. We first utilize pre-trained language model BERT to obtain semantically rich embeddings for documents at different granularities, including documents, paragraphs, sentence-sets, and sentences. Then, we construct a heterogeneous graph with 4 types of nodes (document nodes, paragraph nodes, sentence-set nodes, and sentence nodes) and corresponding connecting edges to model rich document relationships. Furthermore, we propose a novel Multi-Granularity Relationship-based Extractor (MGRExtractor) to produce meta-documents by efficiently pruning heterogeneous graphs. More precisely, it consists of 4 main modules: noise removal, redundancy removal, multi-granularity scoring, and sentence-set selection. In the fine-grained stage, we employ the large configuration of BART as our abstractive summarizer to generate system summaries from the extracted meta-documents. Experimental results on two benchmark datasets show that our framework significantly outperforms strong baselines with comparable parameters, and slightly underperforms methods with a maximum encoding length of 16,384 tokens. For Multi-News and WCEP, automatic evaluation results show that MDS-MGRE achieves an average performance improvement of 1.75% and 8.77% compared to the state-of-the-art systems with comparable parameters, respectively. Such positive results demonstrate the benefits of generating high-quality meta-documents to enhance MDS by modeling rich document relationships.

References

[1]
Agarwal R., Chatterjee N., Improvements in multi-document abstractive summarization using multi sentence compression with word graph and node alignment, Expert Systems with Applications 190 (2022).
[2]
Aletras N., Stevenson M., Measuring the similarity between automatically generated topics, in: Proceedings of the 14th conference of the European chapter of the association for computational linguistics, EACL 2014, April 26-30, 2014, Gothenburg, Sweden, The Association for Computer Linguistics, 2014, pp. 22–27.
[3]
Antognini, D., & Faltings, B. (2019). Learning to Create Sentence Semantic Relation Graphs for Multi-Document Summarization. In Proceedings of the 2nd workshop on new frontiers in summarization (pp. 32–41).
[4]
Bae, S., Kim, T., Kim, J., & Lee, S.-g. (2019). Summary Level Training of Sentence Rewriting for Abstractive Summarization. In Proceedings of the 2nd workshop on new frontiers in summarization (pp. 10–20).
[5]
Banerjee S., Mukherjee S., Bandyopadhyay S., Pakray P., An extract-then-abstract based method to generate disaster-news headlines using a DNN extractor followed by a transformer abstractor, Information Processing and Management 60 (3) (2023).
[6]
Bani-Almarjeh M., Kurdy M., Arabic abstractive text summarization using RNN-based and transformer-based architectures, Information Processing and Management 60 (2) (2023).
[7]
Beltagy I., Peters M.E., Cohan A., Longformer: The long-document transformer, 2020, CoRR abs/2004.05150.
[8]
Belwal R.C., Rai S., Gupta A., Text summarization using topic-based vector space model and semantic measure, Information Processing and Management 58 (3) (2021).
[9]
Bidoki M., Moosavi M.R., Fakhrahmad S.M., A semantic approach to extractive multi-document summarization: Applying sentence expansion for tuning of conceptual densities, Information Processing and Management 57 (6) (2020).
[10]
Chen M., Li W., Liu J., Xiao X., Wu H., Wang H., Sgsum: Transforming multi-document summarization into sub-graph selection, in: Proceedings of the 2021 conference on empirical methods in natural language processing, EMNLP 2021, virtual event / Punta Cana, Dominican Republic, 7-11 November, 2021, Association for Computational Linguistics, 2021, pp. 4063–4074.
[11]
Cheng H., Wu J., Li T., Cao B., Fan J., Improving abstractive multi-document summarization with predicate-argument structure extraction, in: PRICAI 2022: trends in artificial intelligence - 19th Pacific rim international conference on artificial intelligence, PRICAI 2022, Shanghai, China, November 10-13, 2022, proceedings, part II, in: Lecture notes in computer science, Vol. 13630, Springer, 2022, pp. 268–282.
[12]
Dong L., Satpute M.N., Wu W., Du D., Two-phase multidocument summarization through content-attention-based subtopic detection, IEEE Transactions on Computational Social Systems 8 (6) (2021) 1379–1392.
[13]
Fabbri A.R., Li I., She T., Li S., Radev D.R., Multi-news: A large-scale multi-document summarization dataset and abstractive hierarchical model, in: Proceedings of the 57th conference of the association for computational linguistics, ACL 2019, Florence, Italy, July 28- August 2, 2019, volume 1: long papers, Association for Computational Linguistics, 2019, pp. 1074–1084.
[14]
Fan A., Gardent C., Braud C., Bordes A., Using local knowledge graph construction to scale Seq2Seq models to multi-document inputs, in: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3-7, 2019, Association for Computational Linguistics, 2019, pp. 4184–4194.
[15]
Gehrmann S., Deng Y., Rush A.M., Bottom-up abstractive summarization, in: Proceedings of the 2018 conference on empirical methods in natural language processing, Brussels, Belgium, October 31 - November 4, 2018, Association for Computational Linguistics, 2018, pp. 4098–4109.
[16]
Ghadimi A., Beigy H., Hybrid multi-document summarization using pre-trained language models, Expert Systems with Applications 192 (2022).
[17]
Ghalandari D.G., Hokamp C., Pham N.T., Glover J., Ifrim G., A large-scale multi-document summarization dataset from the wikipedia current events portal, in: Proceedings of the 58th annual meeting of the association for computational linguistics, ACL 2020, online, July 5-10, 2020, Association for Computational Linguistics, 2020, pp. 1302–1308.
[18]
Guo M., Ainslie J., Uthus D.C., Ontañón S., Ni J., Sung Y., Yang Y., LongT5: Efficient text-to-text transformer for long sequences, in: Findings of the association for computational linguistics: NAACL 2022, Seattle, WA, United States, July 10-15, 2022, Association for Computational Linguistics, 2022, pp. 724–736.
[19]
Guo J., Liu J., Liu X., Wan Y., Li L., Summarizing source code with heterogeneous syntax graph and dual position, Information Processing and Management 60 (5) (2023).
[20]
Hendrickx I., Daelemans W., Marsi E., Krahmer E., Reducing redundancy in multi-document summarization using lexical semantic similarity, in: Proceedings of the 2009 workshop on language generation and summarisation, 2009, pp. 63–66.
[21]
Jin H., Wang T., Wan X., Multi-granularity interaction network for extractive and abstractive multi-document summarization, in: Proceedings of the 58th annual meeting of the association for computational linguistics, ACL 2020, online, July 5-10, 2020, Association for Computational Linguistics, 2020, pp. 6244–6254.
[22]
Kundi F.M., Asghar M.Z., Zahra S.R., Ahmad S., Khan A., A review of text summarization, Language 6 (7) (2014) 8.
[23]
Lamsiyah S., Mahdaouy A.E., Espinasse B., Ouatik S.E.A., An unsupervised method for extractive multi-document summarization based on centroid approach and sentence embeddings, Expert Systems with Applications 167 (2021).
[24]
Lebanoff L., Song K., Dernoncourt F., Kim D.S., Kim S., Chang W., Liu F., Scoring sentence singletons and pairs for abstractive summarization, in: Proceedings of the 57th conference of the association for computational linguistics, ACL 2019, Florence, Italy, July 28- August 2, 2019, volume 1: long papers, Association for Computational Linguistics, 2019, pp. 2175–2189.
[25]
Lewis M., Liu Y., Goyal N., Ghazvininejad M., Mohamed A., Levy O., Stoyanov V., Zettlemoyer L., BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension, in: Proceedings of the 58th annual meeting of the association for computational linguistics, ACL 2020, online, July 5-10, 2020, Association for Computational Linguistics, 2020, pp. 7871–7880.
[26]
Li M., Qi J., Lau J.H., Compressed heterogeneous graph for abstractive multi-document summarization, in: Thirty-seventh AAAI conference on artificial intelligence, AAAI 2023, thirty-fifth conference on innovative applications of artificial intelligence, IAAI 2023, thirteenth symposium on educational advances in artificial intelligence, EAAI 2023, Washington, DC, USA, February 7-14, 2023, AAAI Press, 2023, pp. 13085–13093.
[27]
Li W., Xiao X., Liu J., Wu H., Wang H., Du J., Leveraging graph to improve abstractive multi-document summarization, in: Proceedings of the 58th annual meeting of the association for computational linguistics, ACL 2020, online, July 5-10, 2020, Association for Computational Linguistics, 2020, pp. 6232–6243.
[28]
Li S., Xu J., Hiermds: a hierarchical multi-document summarization model with global-local document dependencies, Neural Computing and Applications 35 (25) (2023) 18553–18570.
[29]
Li S., Xu J., MRC-sum: An MRC framework for extractive summarization of academic articles in natural sciences and medicine, Information Processing and Management 60 (5) (2023).
[30]
Li W., Zhuge H., Abstractive multi-document summarization based on semantic link network, IEEE Transactions on Knowledge and Data Engineering 33 (1) (2021) 43–54.
[31]
Lin C.-Y., ROUGE: A package for automatic evaluation of summaries, in: Text summarization branches out, Association for Computational Linguistics, Barcelona, Spain, 2004, pp. 74–81.
[32]
Liu S., Cao J., Yang R., Wen Z., Key phrase aware transformer for abstractive summarization, Information Processing and Management 59 (3) (2022).
[33]
Liu Y., Lapata M., Hierarchical transformers for multi-document summarization, in: Proceedings of the 57th conference of the association for computational linguistics, ACL 2019, Florence, Italy, July 28- August 2, 2019, volume 1: long papers, Association for Computational Linguistics, 2019, pp. 5070–5081.
[34]
Liu P.J., Saleh M., Pot E., Goodrich B., Sepassi R., Kaiser L., Shazeer N., Generating wikipedia by summarizing long sequences, in: 6th international conference on learning representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, conference track proceedings, OpenReview.net, 2018.
[35]
Lu M., Liang L., Liu G., Parallel relationship graph to improve multi-document summarization, in: Artificial neural networks and machine learning - ICANN 2022 - 31st international conference on artificial neural networks, Bristol, UK, September 6-9, 2022, proceedings, part II, in: Lecture notes in computer science, Vol. 13530, Springer, 2022, pp. 630–642.
[36]
Ma C., Zhang W.E., Guo M., Wang H., Sheng Q.Z., Multi-document summarization via deep learning techniques: A survey, ACM Computing Surveys 55 (5) (2022) 1–37.
[37]
Ma C., Zhang W.E., Guo M., Wang H., Sheng Q.Z., Multi-document summarization via deep learning techniques: A survey, ACM Computing Surveys 55 (5) (2023) 102:1–102:37.
[38]
Mao Y., Qu Y., Xie Y., Ren X., Han J., Multi-document summarization with maximal marginal relevance-guided reinforcement learning, in: Proceedings of the 2020 conference on empirical methods in natural language processing, EMNLP 2020, online, November 16-20, 2020, Association for Computational Linguistics, 2020, pp. 1737–1751.
[39]
Mao Z., Wu C.H., Ni A., Zhang Y., Zhang R., Yu T., Deb B., Zhu C., Awadallah A.H., Radev D.R., DYLE: dynamic latent extraction for abstractive long-input summarization, in: Proceedings of the 60th annual meeting of the association for computational linguistics (volume 1: long papers), ACL 2022, Dublin, Ireland, May 22-27, 2022, Association for Computational Linguistics, 2022, pp. 1687–1698.
[40]
Mendes A., Narayan S., Miranda S., Marinho Z., Martins A.F.T., Cohen S.B., Jointly extracting and compressing documents with summary state representations, in: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, volume 1 (long and short papers), Association for Computational Linguistics, 2019, pp. 3955–3966.
[41]
Nayeem M.T., Fuad T.A., Chali Y., Abstractive unsupervised multi-document summarization using paraphrastic sentence fusion, in: Proceedings of the 27th international conference on computational linguistics, COLING 2018, Santa Fe, New Mexico, USA, August 20-26, 2018, Association for Computational Linguistics, 2018, pp. 1191–1204.
[42]
Nguyen M., Cuong T.V., Nguyen X.H., Nguyen L., Web document summarization by exploiting social context with matrix co-factorization, Information Processing and Management 56 (3) (2019) 495–515.
[43]
Nguyen V., Mai S.T., Nguyen M., Learning to summarize multi-documents with local and global information, Progress in Artificial Intelligence 12 (3) (2023) 275–286.
[44]
Parnell J., Unanue I.J., Piccardi M., A multi-document coverage reward for relaxed multi-document summarization, in: Proceedings of the 60th annual meeting of the association for computational linguistics (volume 1: long papers), ACL 2022, Dublin, Ireland, May 22-27, 2022, Association for Computational Linguistics, 2022, pp. 5112–5128.
[45]
Pasunuru R., Liu M., Bansal M., Ravi S., Dreyer M., Efficiently summarizing text and graph encodings of multi-document clusters, in: Proceedings of the 2021 conference of the North American chapter of the association for computational linguistics: human language technologies, NAACL-HLT 2021, online, June 6-11, 2021, Association for Computational Linguistics, 2021, pp. 4768–4779.
[46]
Pilault J., Li R., Subramanian S., Pal C., On extractive and abstractive neural document summarization with transformer language models, in: Proceedings of the 2020 conference on empirical methods in natural language processing, EMNLP 2020, online, November 16-20, 2020, Association for Computational Linguistics, 2020, pp. 9308–9319.
[47]
Puduppully R.S., Jain P., Chen N., Steedman M., Multi-document summarization with centroid-based pretraining, in: Proceedings of the 61st annual meeting of the association for computational linguistics (volume 2: short papers), ACL 2023, Toronto, Canada, July 9-14, 2023, Association for Computational Linguistics, 2023, pp. 128–138.
[48]
Qi W., Yan Y., Gong Y., Liu D., Duan N., Chen J., Zhang R., Zhou M., ProphetNet: Predicting future N-gram for sequence-to-sequence pre-training, in: Findings of the association for computational linguistics: EMNLP 2020, online event, 16-20 November 2020, in: Findings of ACL, Vol. EMNLP 2020, Association for Computational Linguistics, 2020, pp. 2401–2410.
[49]
See A., Liu P.J., Manning C.D., Get to the point: Summarization with pointer-generator networks, in: Proceedings of the 55th annual meeting of the association for computational linguistics, ACL 2017, Vancouver, Canada, July 30 - August 4, volume 1: long papers, Association for Computational Linguistics, 2017, pp. 1073–1083.
[50]
Shen C., Cheng L., Nguyen X., You Y., Bing L., A hierarchical encoding-decoding scheme for abstractive multi-document summarization, 2023, CoRR abs/2305.08503.
[51]
Song Y., Chen Y., Shuai H., Improving multi-document summarization through referenced flexible extraction with credit-awareness, in: Proceedings of the 2022 conference of the North American chapter of the association for computational linguistics: human language technologies, NAACL 2022, Seattle, WA, United states, July 10-15, 2022, Association for Computational Linguistics, 2022, pp. 1667–1681.
[52]
Tu S., Yu J., Zhu F., Li J., Hou L., Nie J., UPER: boosting multi-document summarization with an unsupervised prompt-based extractor, in: Proceedings of the 29th international conference on computational linguistics, COLING 2022, Gyeongju, Republic of Korea, October 12-17, 2022, International Committee on Computational Linguistics, 2022, pp. 6315–6326.
[53]
Vaswani A., Shazeer N., Parmar N., Uszkoreit J., Jones L., Gomez A.N., Kaiser L., Polosukhin I., Attention is all you need, in: Advances in neural information processing systems 30: annual conference on neural information processing systems 2017, December 4-9, 2017, Long Beach, CA, USA, 2017, pp. 5998–6008.
[54]
Verma P., Om H., MCRMR: maximum coverage and relevancy with minimal redundancy based multi-document summarization, Expert Systems with Applications 120 (2019) 43–56.
[55]
Wang H., Chang J., Huang J., User intention-based document summarization on heterogeneous sentence networks, in: Database systems for advanced applications - 24th international conference, DASFAA 2019, Chiang Mai, Thailand, April 22-25, 2019, proceedings, part II, in: Lecture notes in computer science, Vol. 11447, Springer, 2019, pp. 572–587.
[56]
Wang P., Li S., Liu S., Tang J., Wang T., Plan and generate: Explicit and implicit variational augmentation for multi-document summarization of scientific articles, Information Processing and Management 60 (4) (2023).
[57]
Wang D., Liu P., Zheng Y., Qiu X., Huang X., Heterogeneous graph neural networks for extractive document summarization, in: Proceedings of the 58th annual meeting of the association for computational linguistics, ACL 2020, online, July 5-10, 2020, Association for Computational Linguistics, 2020, pp. 6209–6219.
[58]
Wu Y., Hu B., Learning to extract coherent summary via deep reinforcement learning, in: Proceedings of the thirty-second AAAI conference on artificial intelligence, (AAAI-18), the 30th innovative applications of artificial intelligence (IAAI-18), and the 8th AAAI symposium on educational advances in artificial intelligence (EAAI-18), New Orleans, Louisiana, USA, February 2-7, 2018, AAAI Press, 2018, pp. 5602–5609.
[59]
Xiao W., Beltagy I., Carenini G., Cohan A., PRIMERA: pyramid-based masked sentence pre-training for multi-document summarization, in: Proceedings of the 60th annual meeting of the association for computational linguistics (volume 1: long papers), ACL 2022, dublin, Ireland, May 22-27, 2022, Association for Computational Linguistics, 2022, pp. 5245–5263.
[60]
Xu S., Li H., Yuan P., Wu Y., He X., Zhou B., Self-attention guided copy mechanism for abstractive summarization, in: Proceedings of the 58th annual meeting of the association for computational linguistics, ACL 2020, online, July 5-10, 2020, Association for Computational Linguistics, 2020, pp. 1355–1362,.
[61]
Yasunaga M., Zhang R., Meelu K., Pareek A., Srinivasan K., Radev D.R., Graph-based neural multi-document summarization, in: Proceedings of the 21st conference on computational natural language learning (coNLL 2017), Cancouver, Canada, August 3-4, 2017, Association for Computational Linguistics, 2017, pp. 452–462.
[62]
Yin Y., Song L., Su J., Zeng J., Zhou C., Luo J., Graph-based neural sentence ordering, in: Proceedings of the twenty-eighth international joint conference on artificial intelligence, IJCAI 2019, Macao, China, August 10-16, 2019, ijcai.org, 2019, pp. 5387–5393.
[63]
Zhang Z., Elfardy H., Dreyer M., Small K., Ji H., Bansal M., Enhancing multi-document summarization with cross-document graph-based information extraction, in: Proceedings of the 17th conference of the European chapter of the association for computational linguistics, EACL 2023, Dubrovnik, Croatia, May 2-6, 2023, Association for Computational Linguistics, 2023, pp. 1688–1699.
[64]
Zhang M., Li C., Wan M., Zhang X., Zhao Q., ROUGE-SEM: Better evaluation of summarization using ROUGE combined with semantics, Expert Systems with Applications 237 (2024).
[65]
Zhang Y., Ni A., Mao Z., Wu C.H., Zhu C., Deb B., Awadallah A.H., Radev D.R., Zhang R., Summn: A multi-stage summarization framework for long input dialogues and documents, in: Proceedings of the 60th annual meeting of the association for computational linguistics (volume 1: long papers), ACL 2022, dublin, Ireland, May 22-27, 2022, Association for Computational Linguistics, 2022, pp. 1592–1604.
[66]
Zhang J., Tan J., Wan X., Adapting neural single-document summarization model for abstractive multi-document summarization: A pilot study, in: Proceedings of the 11th international conference on natural language generation, Tilburg University, the Netherlands, November 5-8, 2018, Association for Computational Linguistics, 2018, pp. 381–390.
[67]
Zhang J., Zhao Y., Saleh M., Liu P., PEGASUS: Pre-training with extracted gap-sentences for abstractive summarization, in: Proceedings of the 37th international conference on machine learning, in: Proceedings of machine learning research, Vol. 119, PMLR, 2020, pp. 11328–11339.
[68]
Zhang M., Zhou G., Huang N., He P., Yu W., Liu W., AsU-OSum: Aspect-augmented unsupervised opinion summarization, Information Processing and Management 60 (1) (2023).
[69]
Zhang M., Zhou G., Yu W., Liu W., FAR-ASS: fact-aware reinforced abstractive sentence summarization, Information Processing and Management 58 (3) (2021).
[70]
Zhao C., Huang T., Chowdhury S.B.R., Chandrasekaran M.K., McKeown K.R., Chaturvedi S., Read top news first: A document reordering approach for multi-document news summarization, in: Findings of the association for computational linguistics: ACL 2022, Dublin, Ireland, May 22-27, 2022, Association for Computational Linguistics, 2022, pp. 613–621.
[71]
Zhong M., Liu P., Chen Y., Wang D., Qiu X., Huang X., Extractive summarization as text matching, in: Proceedings of the 58th annual meeting of the association for computational linguistics, ACL 2020, online, July 5-10, 2020, Association for Computational Linguistics, 2020, pp. 6197–6208.
[72]
Zhu C., Xu R., Zeng M., Huang X., A hierarchical network for abstractive meeting summarization with cross-domain pretraining, in: Findings of the association for computational linguistics: EMNLP 2020, online event, 16-20 November 2020, in: Findings of ACL, Vol. EMNLP 2020, Association for Computational Linguistics, 2020, pp. 194–203.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image Information Processing and Management: an International Journal
Information Processing and Management: an International Journal  Volume 61, Issue 3
May 2024
1388 pages

Publisher

Pergamon Press, Inc.

United States

Publication History

Published: 01 May 2024

Author Tags

  1. Natural language processing
  2. Multi-document summarization
  3. Graph-based summarization
  4. Rich document relationships
  5. Extract-then-summarize pipeline

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 0
    Total Downloads
  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 26 Jan 2025

Other Metrics

Citations

View Options

View options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media