Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Enhanced Language Modeling with Proximity and Sentence Relatedness Information for Extractive Broadcast News Summarization

Published: 07 February 2020 Publication History

Abstract

The primary task of extractive summarization is to automatically select a set of representative sentences from a text or spoken document that can concisely express the most important theme of the original document. Recently, language modeling (LM) has been proven to be a promising modeling framework for performing this task in an unsupervised manner. However, there still remain three fundamental challenges facing the existing LM-based methods, which we set out to tackle in this article. The first one is how to construct a more accurate sentence model in this framework without resorting to external sources of information. The second is how to take into account sentence-level structural relationships, in addition to word-level information within a document, for important sentence selection. The last one is how to exploit the proximity cues inherent in sentences to obtain a more accurate estimation of respective sentence models. Specifically, for the first and second challenges, we explore a novel, principled approach that generates overlapped clusters to extract sentence relatedness information from the document to be summarized, which can be used not only to enhance the estimation of various sentence models but also to render sentence-level structural relationships within the document, leading to better summarization effectiveness. For the third challenge, we investigate several formulations of proximity cues for use in sentence modeling involved in the LM-based summarization framework, free of the strict bag-of-words assumption. Furthermore, we also present various ensemble methods that seamlessly integrate proximity and sentence relatedness information into sentence modeling. Extensive experiments conducted on a Mandarin broadcast news summarization task show that such integration of proximity and sentence relatedness information is indeed beneficial for speech summarization. Our proposed summarization methods can significantly boost the performance of an LM-based strong baseline (e.g., with a maximum ROUGE-2 improvement of 26.7% relative) and also outperform several state-of-the-art unsupervised methods compared in the article.

References

[1]
S. Furui, L. Deng, M. Gales, H. Ney, and K. Tokuda. 2012. Fundamental technologies in modern speech recognition. IEEE Sign. Process. Mag. 29, 6 (2012), 16--17.
[2]
R. Blanco, G. D. F. Morales, and F. Silvestri. 2016. IntoNews: Online news retrieval using closed captions. Inf. Process. Manage. 51, 1 (2016), 148--162.
[3]
M. Ostendorf. 2008. Speech technology and information access. IEEE Sign. Process. Mag. 25, 3 (2008), 150--152.
[4]
L.-S. Lee and B. Chen. 2005. Spoken document understanding and organization. IEEE Sign. Process. Mag. 22, 5 (2005), 42--60.
[5]
D. Yu and L. Deng. 2015. Automatic Speech Recognition: A Deep Learning Approach. Springer.
[6]
Y. Liu and D. Hakkani-Tur. 2011. Speech Summarization. Wiley, New York, NY.
[7]
G. Penn and X. Zhu. 2008. A critical reassessment of evaluation baselines for speech summarization. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL’08), 470--478.
[8]
A. Nenkova and K. McKeown. 2011. Automatic summarization, Found. Trends Inf. Retriev. 5, 2--3 (2011), 103--233.
[9]
I. Mani and M. T. Maybury. (Eds.) 1999. Advances in Automatic Text Summarization. MIT Press, Cambridge, MA.
[10]
J.-M. Torres-Moreno (Ed.). 2014. Automatic Text Summarization. Wiley-ISTE, New York, NY.
[11]
X. Zhu, G. Penn, and F. Rudzicz. 2009. Summarizing multiple spoken documents: Finding evidence from untranscribed audio. In Proceedings of the Joint Conference of Association for Computational Linguistics and International Joint Conference on Natural Language Processing (ACL-IJCNLP’09), 549--557.
[12]
S.-H. Liu, H.-S. Lee, H.-T. Hung, K.-Y. Chen, B. Chen, H.-M. Wang, H.-C. Yen, and W.-L. Hsu. 2015. Incorporating proximity information in relevance language modeling for extractive speech summarization. In Proceedings of the APSIPA Annual Summit and Conference. 401--407.
[13]
A. Celikyilmaz and D. Hakkani-Tur. 2010. A hybrid hierarchical model for multi-document summarization. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL’10), 815--824.
[14]
S.-H. Lin, Y.-M Yeh, and B. Chen. 2011. Leveraging Kullback-Leibler divergence measures and information-rich cues for speech summarization. IEEE Trans. Audio Speech Lang. Process. 19, 4 (2011), 871--882.
[15]
S.-H. Liu, K.-Y. Chen, B. Chen, H.-M. Wang, H.-C. Yen, and W.-L Hsu. 2015. Combining relevance language modeling and clarity measure for extractive speech summarization. IEEE/ACM Trans. Audio Speech Lang. Process. 23, 6 (2015), pp. 957--969.
[16]
P. B. Baxendale. 1958. Machine-made index for technical literature-an experiment. IBM J. 2, 3 (1958), 354--361.
[17]
Y. Gong and X. Liu. 2001. Generic text summarization using relevance measure and latent semantic analysis. In Proceedings of the ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’01), 19--25.
[18]
M. Kagebäck, O. Mogren, N. Tahmasebi, and D. Dubhashi. 2014. Extractive summarization using continuous vector space models. In Proceedings of the Workshop on Continuous Vector Space Models and Their Composition, 31--39.
[19]
X. Wan and J. Yang. 2008. Multi-document summarization using cluster-based link analysis. In Proceedings of the ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’08), 299--306.
[20]
S. Xiong and D. Ji. 2016. Query-focused multi-document summarization using hypergraph-based ranking. Inf Process. Manage. 52, 4 (2016), 670--681.
[21]
J. Carbonell and J. Goldstein. 1998. The use of MMR, diversity based reranking for reordering documents and producing summaries. In Proceedings of the ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’98), 335--336.
[22]
S. Furui, T. Kikuchi, Y. Shinnaka, and C. Hori. 2004. Speech-to-text and speech-to-speech summarization of spontaneous speech. IEEE Trans. Speech Audio Process. 12, 4 (2004), 401--408.
[23]
G. Erkan and D. R. Radev. 2004. LexRank: Graph-based lexical centrality as salience in text summarization. J. Artif. Intell. Res. 22, 1 (2004), 457--479.
[24]
H. Lin and J. Bilmes. 2010. Multi-document summarization via budgeted maximization of submodular functions. In Proceedings of Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL’10), 912--920.
[25]
K. Riedhammer, B. Favrec, and D. Hakkani-Tür. 2010. Long story short - Global unsupervised models for keyphrase based meeting summarization. Speech Commun. 52, 10 (2010), 801--815.
[26]
M. A. Fattah and F. Ren. 2009. GA, MR, FFNN, PNN and GMM based models for automatic text summarization. Comput. Speech Lang. 23, 1 (2009), 126--144.
[27]
J. Kupie, J. Pedersen, and F. Chen. 1995. A trainable document summarizer. In Proceedings of the ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’95), 68--73.
[28]
J. Zhang and P. Fung. 2007. Speech summarization without lexical features for Mandarin broadcast news. In Proceedings of the Annual Conference of the North AMERICAN Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL’07), 213--216.
[29]
M. Galley. 2006. Skip-chain conditional random field for ranking meeting utterances by importance. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP’06), 364--372.
[30]
R. Baeza-Yates and B. Ribeiro-Neto. 2011. Modern Information Retrieval: The Concepts and Technology behind Search. ACM Press.
[31]
C.-X. Zhai. 2008. Statistical language models for information retrieval: A critical review. Found. Trends Inf. Retriev. 2, 3 (2008), 137--213.
[32]
Y. Lv and C.-X. Zhai. 2015. Negative query generation: Bridging the gap between query likelihood retrieval models and relevance. Inf. Retriev. J. 18, 4 (2015), 359--378.
[33]
R. Rahimi, A. Shakery, and I. King. 2016. Extracting translations from comparable corpora for Cross-Language Information Retrieval using the language modeling framework. Inf. Process. Manage. 52, 2 (2016), 299--318.
[34]
S. Liang and M. Rijke. 2016. Formal language models for finding groups of experts. Inf. Process. Manage. 52, 4 (2016), 529--549.
[35]
A. Haghighi and L. Vanderwende. 2009. Exploring content models for multi-document summarization. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL HLT’09), 362--370.
[36]
S.-H. Liu, K.-Y. Chen, Y.-L. Hsieh, B. Chen, H.-M. Wang, H.-C. Yen, and W.-L. Hsu. 2014. Enhanced language modeling for extractive speech summarization with sentence relatedness information. In Proceedings of the Annual Conference of the International Speech Communication Association (Interspeech’14), 1865--1869.
[37]
S.-H Liu, K.-Y. Chen, Y.-L. Hsieh, B. Chen, H.-M. Wang, H.-C. Yen, and W.-L. Hsu. 2017. A position-aware language modeling framework for extractive broadcast news speech summarization, ACM Trans. Asian Low-Res. Lang. Inf. Process. 16, 4 (2017), 27:1--27:13.
[38]
C.-X. Zhai and J. Lafferty. 2001. Model-based feedback in the language modeling approach to information retrieval. In Proceedings of Conference on Information and Knowledge Management (CIKM’01), 403--410.
[39]
V. Lavrenko and B. Croft. 2001. Relevance-based language models. In Proceedings of the ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’01), 120--127.
[40]
B. Chen and K.-Y. Chen. 2013. Leveraging relevance cues for language modeling in speech recognition. Inf. Process. Manage. 49, 4 (2013), 807--816.
[41]
J. Frederick. 1999. Statistical Methods for Speech Recognition. The MIT Press.
[42]
R. A. A. K. Lund and C. Burgess. 1995. Semantic and associative priming in high-dimensional semantic space. In Proceedings of the Annual Conference of the Cognitive Science Society, 660--665.
[43]
J. Miao, J. X. Huang, and Z. Ye. 2012. Proximity-based Rocchio's model for pseudo relevance feedback. In Proceedings of the ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’12), 535--544.
[44]
D. Metzler and W. B. Croft. 2005. A markov random field model for term dependencies. In Proceedings of the ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’05), 472--479.
[45]
Y. Lv and C.-X. Zhai. 2010. Positional relevance model for pseudo relevance feedback. In Proceedings of the ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’10), 579--586.
[46]
X. Liu and W. B. Croft. 2004. Cluster-based retrieval using language models. In Proceedings of the ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’04), 186--193.
[47]
Y. Huang, L. Sun and J. Y. Nie. 2009. Smoothing document language model with local word graph. In Proceedings of the ACM Conference on Information and Knowledge Management (CIKM’09), 1943--1946.
[48]
K. S. Lee and W. B. Croft 2013. A deterministic resampling method using overlapping document clusters for pseudo-relevance feedback. Inf. Process. Manage. 49, 4 (2013), 792--806.
[49]
H.-M. Wang, B. Chen, J.-W. Kuo, and S.-S. Cheng. 2005. MATBN: A Mandarin Chinese broadcast news corpus. Int. J. Comput. Linguist Chin. Lang. Process. 10, 2 (2005), 219--236.
[50]
J.-T. Chien. (2015). Hierarchical Pitman-Yor-Dirichlet language model. IEEE/ACM Trans. Audio Speech Lang. Process. 23, 8 (2015), 1259--1272.
[51]
C.-L Huang and C.-H. Wu. 2007. Spoken document retrieval using multi-level knowledge and semantic verification. IEEE Trans. Audio Speech Lang. Process. 15, 8 (2007), 2551--2560.
[52]
C.-Y. Lin. 2003. ROUGE: Recall-oriented understudy for gisting evaluation. Retrieved from http://haydn.isi.edu/ROUGE/.
[53]
B. Chen, J.-W. Kuo, and W.-H. Tsai. 2004. Lightly supervised and data-driven approaches to Mandarin broadcast news transcription. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP’04), 777--780.
[54]
B. Chen, S.-H. Lin, Y.-M. Chang, and J.-W. Liu. 2013. Extractive speech summarization using evaluation metric-related training criteria. Inf. Process. Manage. 49, 1 (2013), 1--12.
[55]
J. Pennington, R. Socher, and C. D. Manning. 2014. Glove: Global vectors for word representation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP’14), 1532--1543.
[56]
Q. Le and T. Mikolov. 2014. Distributed representations of sentences and documents. In Proceedings of the 31st International Conference on International Conference on Machine Learning (ICML’14), 1188--1196.
[57]
M. Sundermeyer, H. Ney, and R. Schluter. 2015. From feedforward to recurrent LSTM neural networks for language modeling. IEEE/ACM Trans. Audio Speech Lang. Process. 23, 3 (2015), 517--529.
[58]
H. Palangi, L. Deng, Y. Shen, J. Gao, X. He, J. Chen, X. Song, and R. Ward. 2016. Deep sentence embedding using the long short term memory network: Analysis and application to information retrieval. IEEE/ACM Trans. Audio Speech Lang. Process. 24, 4 (2016), 694--707.
[59]
K.-Y. Chen, S.-H. Liu, B. Chen, H.-M. Wang, E.-E. Jan, W.-L. Hsu, and H.-H. Chen. 2015. Extractive broadcast news summarization leveraging recurrent neural network language modeling techniques. IEEE/ACM Trans. Audio Speech Lang. Process. 23, 8 (2015), 1322--1334.
[60]
S. Yan and X. Wan. 2014. SRRank: Leveraging semantic roles for extractive multi-document summarization. IEEE/ACM Trans. Audio Speech Lang. Process. 22, 12 (2014), 2048--2058.
[61]
X. Wang, Y. Yoshida, T. Hirao, K. Sudoh, and M. Nagata. 2015. Summarization based on task-oriented discourse parsing. IEEE/ACM Trans. Audio Speech Lang. Process. 23, 8 (2015), 1358--1367.
[62]
G. Shang, W. Ding, Z. Zhang, A. J.-P. Tixier, P. Meladianos, M. Vazirgiannis, and J.-P. Lorré. 2018. Unsupervised abstractive meeting summarization with multi-sentence compression and budgeted submodular maximization. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (ACL’18), 664--674.
[63]
T. Hasan, M. Abdelwahab, S. Parthasarathy, C. Busso, and Y. Liu. 2016. Automatic composition of broadcast news summaries using rank classifiers trained with acoustic and lexical features. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP’16), 6080--6084.
[64]
C.-I Tsai, H.-T. Hung, K.-Y. Chen, and B. Chen. 2016. Extractive speech summarization leveraging convolutional neural network techniques. In Proceedings of the IEEE Workshop on Spoken Language Technology (SLT’16), 158--164
[65]
G. Murray, S. Renals, J. Carletta, and J. Moore. 2005. Evaluating automatic summaries of meeting recordings. In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, 33--40.
[66]
Z. Cao, F. Wei, S. Li, W. Li, M. Zhou, and H. Wang. 2015. Learning summary prior representation for extractive summarization. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing. 829--833.
[67]
T. Tao and C.-X. Zhai. 2006. Regularized estimation of mixture models for robust pseudo-relevance feedback. In Proceedings of the ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’06). 162--169
[68]
T.-E. Liu, S.-H. Liu, and B. Chen. 2019. A hierarchical neural summarization framework for spoken documents. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP’19).
[69]
J. Clarke and M. Lapata. 2010. Discourse constraints for document compression. J. Mach. Learn. Res. 36, 3, 411--441.
[70]
S. Narayan, S. B. Cohen, and M. Lapata. 2018. Ranking sentences for extractive summarization with reinforcement learning. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT’18). 1747--1759.

Cited By

View all
  • (2024)NLP TRANSFORMERS: ANALYSIS OF LLMS AND TRADITIONAL APPROACHES FOR ENHANCED TEXT SUMMARIZATIONEskişehir Osmangazi Üniversitesi Mühendislik ve Mimarlık Fakültesi Dergisi10.31796/ogummf.1303569Online publication date: 12-Mar-2024
  • (2024)Analyzing the Effects of Transcription Errors on Summary Generation of Bengali Spoken DocumentsACM Transactions on Asian and Low-Resource Language Information Processing10.1145/367800523:9(1-28)Online publication date: 16-Aug-2024
  • (2023)GA-SCS: Graph-Augmented Source Code SummarizationACM Transactions on Asian and Low-Resource Language Information Processing10.1145/355482022:2(1-19)Online publication date: 21-Feb-2023
  • Show More Cited By

Index Terms

  1. Enhanced Language Modeling with Proximity and Sentence Relatedness Information for Extractive Broadcast News Summarization

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Asian and Low-Resource Language Information Processing
    ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 19, Issue 3
    May 2020
    228 pages
    ISSN:2375-4699
    EISSN:2375-4702
    DOI:10.1145/3378675
    Issue’s Table of Contents
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 07 February 2020
    Accepted: 01 December 2019
    Revised: 01 September 2019
    Received: 01 December 2018
    Published in TALLIP Volume 19, Issue 3

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Extractive summarization
    2. language modeling
    3. overlapped clustering
    4. proximity information

    Qualifiers

    • Research-article
    • Research
    • Refereed

    Funding Sources

    • Pervasive Artificial Intelligence Research (PAIR) Labs, Taiwan
    • National Science Council, Taiwan

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)5
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 10 Nov 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)NLP TRANSFORMERS: ANALYSIS OF LLMS AND TRADITIONAL APPROACHES FOR ENHANCED TEXT SUMMARIZATIONEskişehir Osmangazi Üniversitesi Mühendislik ve Mimarlık Fakültesi Dergisi10.31796/ogummf.1303569Online publication date: 12-Mar-2024
    • (2024)Analyzing the Effects of Transcription Errors on Summary Generation of Bengali Spoken DocumentsACM Transactions on Asian and Low-Resource Language Information Processing10.1145/367800523:9(1-28)Online publication date: 16-Aug-2024
    • (2023)GA-SCS: Graph-Augmented Source Code SummarizationACM Transactions on Asian and Low-Resource Language Information Processing10.1145/355482022:2(1-19)Online publication date: 21-Feb-2023
    • (2023)Unsupervised Broadcast News Summarization; a Comparative Study on Maximal Marginal Relevance (MMR) and Latent Semantic Analysis (LSA)2023 28th International Computer Conference, Computer Society of Iran (CSICC)10.1109/CSICC58665.2023.10105403(1-7)Online publication date: 25-Jan-2023
    • (2021)Innovative Bert-Based Reranking Language Models for Speech Recognition2021 IEEE Spoken Language Technology Workshop (SLT)10.1109/SLT48900.2021.9383557(266-271)Online publication date: 19-Jan-2021
    • (2021)Cross-sentence Neural Language Models for Conversational Speech Recognition2021 International Joint Conference on Neural Networks (IJCNN)10.1109/IJCNN52387.2021.9533796(1-7)Online publication date: 2021

    View Options

    Get Access

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media