Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Abstractive Summarization of Text Document in Malayalam Language: Enhancing Attention Model Using POS Tagging Feature

Published: 23 March 2023 Publication History

Editorial Notes

The authors have requested minor, non-substantive changes to the VoR and, in accordance with ACM policies, a Corrected Version of Record was published on May 18, 2023. For reference purposes, the VoR may still be accessed via the Supplemental Material section on this citation page.

Abstract

Over the past few years, researchers are showing huge interest in sentiment analysis and summarization of documents. The primary reason being that huge volumes of information are available in textual format, and this data has proven helpful for real-world applications and challenges. The sentiment analysis of a document will help the user comprehend the content’s emotional intent. Abstractive summarization algorithms generate a condensed version of the text, which can then be used to determine the emotion represented in the text using sentiment analysis. Recent research in abstractive summarization concentrates on neural network-based models, rather than conjunctions-based approaches, which might improve the overall efficiency. Neural network models like attention mechanism are tried out to handle complex works with promising results. The proposed work aims to present a novel framework that incorporates the part of speech tagging feature to the word embedding layer, which is then used as the input to the attention mechanism. With POS feature being part of the input layer, this framework is capable of dealing with words containing contextual and morphological information. The relevance of POS tagging here is due to its strong reliance on the language’s syntactic, contextual, and morphological information. The three main elements in the work are pre-processing, POS tagging feature in the embedding phase, and the incorporation of it into the attention mechanism. The word embedding provides the semantic concept about the word, while the POS tags give an idea about how significant the words are in the context of the content, which corresponds to the syntactic information. The proposed work was carried out in Malayalam, one of the prominent Indian languages. A widely used and accepted dataset from the English language was translated to Malayalam for conducting the experiments. The proposed framework gives a ROUGE score of 28, which outperformed the baseline models.

Supplementary Material

3561819-vor (3561819-vor.pdf)
Version of Record for "Abstractive Summarization of Text Document in Malayalam Language: Enhancing Attention Model Using POS Tagging Feature" by Nambiar et al., ACM Transactions on Asian and Low-Resource Language Information Processing, Volume 22, No. 2 (TALLIP 22:2).

References

[1]
A. P. Ajees and Sumam Mary Idicula. 2018. A POS tagger for Malayalam using conditional random fields. Int. J. Appl. Eng. Res. 13, 3 (2018).
[2]
Gabor Angeli, Melvin Jose Johnson Premkumar, and Christopher D. Manning. 2015. Leveraging linguistic structure for open domain information extraction. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing. 344–354. DOI:. arxiv:1603.07252.
[3]
Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2014. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 (2014).
[4]
Lidong Bing, Piji Li, Yi Liao, Wai Lam, Weiwei Guo, and Rebecca J. Passonneau. 2015. Abstractive multi-document summarization via phrase selection and merging. arXiv preprint arXiv:1506.01597 (2015).
[5]
Asli Celikyilmaz, Antoine Bosselut, Xiaodong He, and Yejin Choi. 2018. Deep communicating agents for abstractive summarization. arXiv preprint arXiv:1803.10357 (2018).
[6]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
[7]
Deepali K. Gaikwad and C. Namrata Mahender. 2016. A review paper on text summarization. International Journal of Advanced Research in Computer and Communication Engineering 5, 3 (2016), 154–160.
[8]
Sumam Mary Idicula and Peter S. David. 2007. A morphological processor for malayalam language. South Asia Res. 27, 2 (2007), 173–186.
[9]
Younes Jaafar and Karim Bouzoubaa. 2018. Towards a new hybrid approach for abstractive summarization. Procedia Comput. Sci. 142 (2018), 286–293.
[10]
Atif Khan and Naomie Salim. 2014. A review on abstractive summarization methods. J. Theoret. Appl. Inf. Technol. 59, 1 (2014), 64–72. DOI:
[11]
Atif Khan, Naomie Salim, and Yogan Jaya Kumar. 2015. A framework for multi-document abstractive summarization based on semantic role labelling. Appl. Soft Comput. J. 30 (2015), 737–747. DOI:
[12]
Renu Khandelwal. 2020. Attention: Sequence 2 Sequence model with Attention Mechanism. Retrieved from https://towardsdatascience.com/sequence-2-sequence-model-with-attention-mechanism-9e9ca2a613a.
[13]
Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov, and Luke Zettlemoyer. 2019. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461 (2019).
[14]
Chin-Yew Lin. 2004. Rouge: A package for automatic evaluation of summaries. In Text Summarization Branches Out. 74–81.
[15]
Yang Liu. 2019. Fine-tune BERT for extractive summarization. arXiv preprint arXiv:1903.10318 (2019).
[16]
Hans Peter Luhn. 1958. The automatic creation of literature abstracts. IBM J. Res. Devel. 2, 2 (1958), 159–165.
[17]
Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781 (2013).
[18]
Ibrahim F. Moawad and Mostafa Aref. 2012. Semantic graph reduction approach for abstractive text summarization. Proceedings of the International Conference on Computer Engineering and Systems. 132–138. DOI:
[19]
Karvanuur P. Mohanan. 1997. Grammatical relations and clause structure in Malayalam.
[20]
Sindhya K. Nambiar, Sumam Mary Idicula, et al. 2021. Attention based abstractive summarization of Malayalam document. Procedia Comput. Sci. 189 (2021), 250–257.
[21]
Sindhya K. Nambiar, S. David Peter, and Sumam Mary Idicula. 2021. Abstractive summarization of Malayalam document using sequence to sequence model. In Proceedings of the 7th International Conference on Advanced Computing and Communication Systems (ICACCS’21). IEEE, 347–352.
[22]
Romain Paulus, Caiming Xiong, and Richard Socher. 2018. A deep reinforced model for abstractive summarization. In Proceedings of the 6th International Conference on Learning Representations. 1–12. arxiv:1705.04304.
[23]
Mary Priya Sebastian, K. Sheena Kurian, and G. Santhosh Kumar. 2010. A classification of Sandhi rules for suffix separation in Malayalam. In Proceedings of the 38th All India Conference of Dravidian Linguists.
[24]
Abigail See, Peter J. Liu, and Christopher D. Manning. 2017. Get to the point: Summarization with pointer-generator networks. arXiv preprint arXiv:1704.04368 (2017).
[25]
Florian Strub, Harm De Vries, Jeremie Mary, Bilal Piot, Aaron Courvile, and Olivier Pietquin. 2017. End-to-end optimization of goal-driven and visually grounded dialogue systems. In Proceedings of the IJCAI International Joint Conference on Artificial Intelligence. 2765–2771. DOI:
[26]
George Tsatsaronis, Iraklis Varlamis, and Kjetil Nørvåg. 2010. SemanticRank: ranking keywords and sentences using semantic graphs. In Proceedings of the 23rd International Conference on Computational Linguistics (Coling’10). 1074–1082.
[27]
Jingqing Zhang, Yao Zhao, Mohammad Saleh, and Peter Liu. 2020. Pegasus: Pre-training with extracted gap-sentences for abstractive summarization. In Proceedings of the International Conference on Machine Learning. PMLR, 11328–11339.
[28]
Qingyu Zhou, Nan Yang, Furu Wei, Shaohan Huang, Ming Zhou, and Tiejun Zhao. 2018. Neural document summarization by jointly learning to score and select sentences. arXiv preprint arXiv:1807.02305 (2018).

Cited By

View all
  • (2024)Medical Question Summarization with Entity-driven Contrastive LearningACM Transactions on Asian and Low-Resource Language Information Processing10.1145/365216023:4(1-19)Online publication date: 15-Apr-2024
  • (2024)Multization: Multi-Modal Summarization Enhanced by Multi-Contextually Relevant and Irrelevant Attention AlignmentACM Transactions on Asian and Low-Resource Language Information Processing10.1145/365198323:5(1-29)Online publication date: 10-May-2024
  • (undefined)Social-sum-Mal: A Dataset for Abstractive Text Summarization in MalayalamACM Transactions on Asian and Low-Resource Language Information Processing10.1145/3696107

Index Terms

  1. Abstractive Summarization of Text Document in Malayalam Language: Enhancing Attention Model Using POS Tagging Feature

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Asian and Low-Resource Language Information Processing
    ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 22, Issue 2
    February 2023
    624 pages
    ISSN:2375-4699
    EISSN:2375-4702
    DOI:10.1145/3572719
    Issue’s Table of Contents

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 23 March 2023
    Online AM: 10 September 2022
    Accepted: 29 August 2022
    Revised: 21 July 2022
    Received: 27 April 2022
    Published in TALLIP Volume 22, Issue 2

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Abstractive summarization
    2. POS tagging
    3. attention mechanism
    4. encoder decoder
    5. Malayalam language

    Qualifiers

    • Research-article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)92
    • Downloads (Last 6 weeks)2
    Reflects downloads up to 03 Oct 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Medical Question Summarization with Entity-driven Contrastive LearningACM Transactions on Asian and Low-Resource Language Information Processing10.1145/365216023:4(1-19)Online publication date: 15-Apr-2024
    • (2024)Multization: Multi-Modal Summarization Enhanced by Multi-Contextually Relevant and Irrelevant Attention AlignmentACM Transactions on Asian and Low-Resource Language Information Processing10.1145/365198323:5(1-29)Online publication date: 10-May-2024
    • (undefined)Social-sum-Mal: A Dataset for Abstractive Text Summarization in MalayalamACM Transactions on Asian and Low-Resource Language Information Processing10.1145/3696107

    View Options

    Get Access

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    Full Text

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media