A Survey of Abstractive Text Summarization Utilising Pretrained Language Models

被引:5
|
作者
Syed, Ayesha Ayub [1 ]
Gaol, Ford Lumban [1 ]
Boediman, Alfred [2 ]
Matsuo, Tokuro [3 ,4 ]
Budiharto, Widodo [5 ]
机构
[1] Bina Nusantara Univ, Dept Doctor Comp Sci, BINUS Grad Program, Jakarta, Indonesia
[2] Univ Chicago, Booth Sch Business, Dept Econometr & Stat, Chicago, IL 60637 USA
[3] Adv Inst Ind Technol, Grad Sch Ind Technol, Tokyo 1400011, Japan
[4] Asia Univ, Dept M Commerce & Multimedia Applicat, Taichung 41354, Taiwan
[5] Bina Nusantara Univ, Sch Comp Sci, Comp Sci Dept, Jakarta 11480, Indonesia
关键词
Abstractive text summarization; Performance improvement; Pretrained language models;
D O I
10.1007/978-3-031-21743-2_42
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We live in a digital era - an era of technology, artificial intelligence, big data, and information. The data and information on which we depend to fulfil several daily tasks and decision-making can become overwhelming to deal with and requires effective processing. This can be achieved by designing improved and robust automatic text summarization systems. These systems reduce the size of text document while retaining the salient information. The resurgence of deep learning and its progress from the Recurrent Neural Networks to deep transformer based PretrainedLanguage Models (PLM) with huge parameters and ample world and common-sense knowledge have opened the doors for huge success and improvement of theNatural Language Processing tasks including Abstractive Text Summarization (ATS). This work surveys the scientific literature to explore and analyze recent research on pre-trained language models and abstractive text summarization utilizing these models. The pretrained language models on abstractive summarization tasks have been analyzed quantitatively based on ROUGE scores on four standard datasets while the analysis of state-of-the-art ATS models has been conducted qualitatively to identify some issues and challenges encountered on finetuning large PLMs on downstream datasets for abstractive summarization. The survey further highlights some techniques that can help boost the performance of these systems. The findings in terms of performance improvement reveal that the models with better performance use either one or a combination of these strategies: (1) Domain Adaptation, (2) Model Augmentation, (3) Stable finetuning, and (4) Data Augmentation.
引用
收藏
页码:532 / 544
页数:13
相关论文
共 50 条
  • [1] Turkish abstractive text summarization using pretrained sequence-to-sequence models
    Baykara, Batuhan
    Gungor, Tunga
    [J]. NATURAL LANGUAGE ENGINEERING, 2023, 29 (05) : 1275 - 1304
  • [2] A Survey on Abstractive Text Summarization
    Moratanch, N.
    Chitrakala, S.
    [J]. PROCEEDINGS OF IEEE INTERNATIONAL CONFERENCE ON CIRCUIT, POWER AND COMPUTING TECHNOLOGIES (ICCPCT 2016), 2016,
  • [3] Survey on Abstractive Text Summarization
    Raphal, Nithin
    Duwarah, Hemanta
    Daniel, Philemon
    [J]. PROCEEDINGS OF THE 2018 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATION AND SIGNAL PROCESSING (ICCSP), 2018, : 513 - 517
  • [4] Pretrained Language Models for Text Generation: A Survey
    Li, Junyi
    Tang, Tianyi
    Zhao, Wayne Xin
    Wen, Ji-Rong
    [J]. PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 4492 - 4499
  • [5] A Survey of the State-of-the-Art Models in Neural Abstractive Text Summarization
    Syed, Ayesha Ayub
    Gaol, Ford Lumban
    Matsuo, Tokuro
    [J]. IEEE ACCESS, 2021, 9 : 13248 - 13265
  • [6] Reinforcement Learning Models for Abstractive Text Summarization
    Buciumas, Sergiu
    [J]. PROCEEDINGS OF THE 2019 ANNUAL ACM SOUTHEAST CONFERENCE (ACMSE 2019), 2019, : 270 - 271
  • [7] Leveraging ParsBERT and Pretrained mT5 for Persian Abstractive Text Summarization
    Farahani, Mehrdad
    Gharachorloo, Mohammad
    Manthouri, Mohammad
    [J]. 2021 26TH INTERNATIONAL COMPUTER CONFERENCE, COMPUTER SOCIETY OF IRAN (CSICC), 2021,
  • [8] Abstractive Text Summarization for the Urdu Language: Data and Methods
    Awais, Muhammad
    Muhammad Adeel Nawab, Rao
    [J]. IEEE ACCESS, 2024, 12 : 61198 - 61210
  • [9] Dense Text Retrieval Based on Pretrained Language Models: A Survey
    Zhao, Wayne Xin
    Liu, Jing
    Ren, Ruiyang
    Wen, Ji-Rong
    [J]. ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2024, 42 (04)
  • [10] A Survey of Extractive and Abstractive Automatic Text Summarization Techniques
    Dalal, Vipul
    Malik, Latesh
    [J]. 2013 Sixth International Conference on Emerging Trends in Engineering and Technology (ICETET 2013), 2013, : 109 - 110