Text Summarization Using The T5 Transformer Model
Text Summarization Using The T5 Transformer Model
Text Summarization Using The T5 Transformer Model
1-5Department of Artificial Intelligence and Machine Learning, Dayananda Sagar University, Bangalore, India
---------------------------------------------------------------------***---------------------------------------------------------------------
Abstract - In our information-filled world, it is crucial to Finally, complete content and organizational editing
focus on the essential content amidst the overwhelming before formatting. Please take note of the following items
volume of information available. Unfortunately, people often when proofreading spelling and grammar:
spend a significant amount of time sifting through irrelevant
details, inadvertently overlooking crucial information. To 2. Literature Review
address this issue, we present a project that utilizes the T5
transformer model in natural language processing to develop Adhika Pramita Widyassari et al.[1] provides an overview of
an abstractive text summarization system. By leveraging various techniques and methods used in automatic text
advanced language modeling techniques, our project aims to summarization, with a particular focus on the Natural
enhance efficiency, comprehension, and decision-making Language Toolkit (NLTK). The author explores different
processes across various domains. approaches, including extractive and abstractive
summarization, and discusses how NLTK can be utilized in
Key Words: Abstractive summarization, T5 transformer these techniques.
model, Natural language processing.
Preprocessing: NLTK performs essential text
1.INTRODUCTION preprocessing tasks like tokenization, stemming, and
stop-word removal, aiding in information extraction
In our information-filled world, focusing on what truly by breaking text into words or sentences and
matters is essential for success. On average, a person spends reducing words to their root form.
a significant amount of their lifetime reading useless
information, often missing out on significant bits by Sentence Scoring: NLTK facilitates extractive
subconsciously dismissing them. To solve this problem, we summarization by offering tools to calculate
built a text summarizer that condenses lengthy text into sentence similarity (e.g., cosine similarity) and
shorter concise summaries, providing a quick overview of assign scores, enabling the selection of relevant
the main information. sentences based on their importance.
Text summarization is a vital tool in today's information- Feature Extraction: NLTK's part-of-speech tagging
driven world, allowing us to distil the essence of lengthy and named entity recognition assist in identifying
texts into concise summaries. By employing advanced entities and key terms, enhancing summary
natural language processing techniques, text summarizers accuracy and relevance.
extract key information, enabling readers to grasp the main
ideas quickly. In this report, we explore the effectiveness and Language Modeling: In abstractive summarization,
applications of text summarizers, shedding light on their NLTK helps build language models (e.g., n-gram
potential to enhance efficiency, comprehension, and models) for generating concise and coherent
decision-making processes across various domains. summaries by predicting probable next words or
phrases.
1.1 The T5 Transformer Model
Evaluation: NLTK includes evaluation metrics (e.g.,
To achieve this, we use the T5 transformer model which is ROUGE, BLEU) to assess summary quality by
a powerful language model that can understand and generate comparing them with reference summaries and
human-like text. Constructing a text summarizer based on T5 measuring similarity or effectiveness.
is beneficial because it allows for concise and accurate
summarization of lengthy documents. T5's ability to capture Khilji et al. [2] examines Abstractive Text Analysis, described
contextual relationships and generate coherent summaries as a natural language processing (NLP) technique that aims to
makes it an ideal choice for text summarization tasks, generate a concise and coherent summary of a given text by
enabling efficient information extraction and facilitating understanding its content and generating new sentences.
quick comprehension of complex texts. Abstractive summarization involves creating novel sentences
that capture the key information and main ideas of the source
text in a more human-like manner.
© 2023, IRJET | Impact Factor value: 8.226 | ISO 9001:2008 Certified Journal | Page 896
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 10 Issue: 08 | Aug 2023 www.irjet.net p-ISSN: 2395-0072
4. EXPERIMENTATION
4.1 Dataset
This dataset, multi_news found on HuggingFace, consists of Fig -2: ROUGE Scores for 10 epochs
two columns: a feature column containing news text
separated by "|||||," and a target column with human-written
summaries. The target column serves as the reference for
© 2023, IRJET | Impact Factor value: 8.226 | ISO 9001:2008 Certified Journal | Page 897
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 10 Issue: 08 | Aug 2023 www.irjet.net p-ISSN: 2395-0072
ACKNOWLEDGEMENTS
We are deeply grateful to our guide, Prof. Sasikala Nagarajan
for their support and mentorship throughout the course of
this project.
Fig 1 shows the ROUGE scores, it shows that over the span of [1] Adhika Pramita Widyassari, Supriadi Rustad, Guruh Fajar
10 epochs, the model's scores get better. Notably, the highest Shidik, Edi Noersasongko, Abdul Syukur, Affandy Affandy, De
scores are achieved in the order of ROUGE-L, followed by Rosal Ignatius Moses Setiadi, “Review of automatic text
ROUGE-2, ROUGE-1, and ROUGE. This pattern indicates the summarization techniques & methods", Journal of King Saud
model's ability to create coherent and fluent summaries while University 2022
preserving essential information. Despite this progress, the
[2] Khilji, Abdullah & Sinha, Utkarsh & Singh, Pintu & Ali,
ROUGE scores remain relatively low, which means there is
Adnan & Pakray, Dr. Partha "Abstractive Text Summarization
room to improve.
Approaches with Analysis of Evaluation Techniques",
Fig 2 shows the model’s progress when it is trained with 25 Computational Intelligence in Communications and Business
epochs. Throughout 25 epochs, the model's ROUGE scores Analytics 2021
demonstrate progressive enhancement. The highest scores
[3] Ilya Sutskever, Oriol Vinyals, Quoc V. Le, “Sequence to
are consistently observed in the order of ROUGE-L, followed
Sequence Learning with Neural Networks”, arXiv Cornell
by ROUGE-2, ROUGE-1, and ROUGE. This pattern highlights
University 2014.
the model's capability to generate summaries that are not
only coherent but also more fluent compared to the original [4 ]Jakob Uszkoreit, “Transformer: A Novel Neural Network
text, while preserving crucial information. Architecture for Language Understanding”, Google Research
2017
The model's improvement in ROUGE scores can be attributed
to a few key factors. Firstly, longer training exposes the [5] Abigail Rai, Study of Various Methods for Tokenization,
model to a wider range of information, leading to better Applications of Internet things pp 193-200 2020.
performance. Additionally, extended training duration
enhances the model's grasp of human language, resulting in
improved summaries. Furthermore, as the model learns
more, its accuracy in producing summaries that align with
human-generated content also increases, ensuring factual
correctness.
© 2023, IRJET | Impact Factor value: 8.226 | ISO 9001:2008 Certified Journal | Page 898