A Comparative Study of Transformer-based Models for Text Summarization of News Articles
Journal: International Journal of Advanced Trends in Computer Science and Engineering (IJATCSE) (Vol.13, No. 2)Publication Date: 2024-04-07
Authors : Charles Munyao Muia Aaron Mogeni Oirere Rachael Njeri Ndung'u;
Page : 37-43
Keywords : Natural Language Processing; ROUGE Metrics; Text Summarization; Transformers.;
Abstract
Transformer-based models such as GPT, T5, BART, and PEGASUS have made substantial progress in text summarization, a sub-domain of natural language processing that entails extracting important information from lengthy texts. The main objective of this research was to conduct a comparative analysis of these four transformer-based models based on their performance in text summarization of news articles. In achieving this objective, the transformer models pre-trained on extensive datasets were fine-tuned on the CNN/DailyMail dataset using a low learning rate to preserve the learned representations. The T5 transformer records the highest scores of 35.12, 22.75, 32.82, and 28.59 in ROUGE-1, ROUGE-2, ROUGE-L, and ROUGE-Lsum respectively, surpassing GPT, BART, and PEGASUS across all ROUGE metrics. The findings deduced from this study establish the proficiency of encoder-decoder models such as T5 in summary generation. Furthermore, the findings also demonstrated that the fine-tuning process's effectiveness in pre-trained models is improved when the pre-training objective closely aligns with the downstream task.
Other Latest Articles
- THE IMAGE OF RUSSIA IN THE LANGUAGE CONSCIOUSNESS OF RUSSIAN AND CHINESE STUDENTS ACCORDING TO THE DATA OF ASSOCIATIVE EXPERIMENTS (QUANTITATIVE ANALYSIS)
- INNOVATIVE DIGITAL TECHNOLOGIES IN MULTILINGUAL TEACHING OF FOREIGN STUDENTS
- A CASE STUDY OF A STUDENT PROJECT: USING ARTIFICIAL INTELLIGENCE TO DEVELOP AN INDIVIDUAL PROGRAM TO STUDY ENGLISH
- Grass Waste: A Highly Biosorbent for the Removal of Malachite Green Dye From Aqueous Solution
Last modified: 2024-04-19 21:42:18