Evaluating the Summarization Comprehension of Pre-Trained Language Models

被引:0
|
作者
Chernyshev, D. I. [1 ]
Dobrov, B. V. [1 ]
机构
[1] Moscow MV Lomonosov State Univ, Res Comp Ctr, Moscow 119991, Russia
基金
俄罗斯科学基金会;
关键词
abstractive summarization; neural networks; machine learning;
D O I
10.1134/S1995080223080115
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Recent advances in abstractive summarization demonstrate the importance of pre-training tasks, however, general purpose language models manage to outperform summarization-specialized pre-training approaches. While several works addressed the question of pseudosummarization pre-training efficiency in abstractive summarization fine-tuning, none has explored the properties of pre-trained models in a low-resource setting. This work attempts to fill this gap. We benchmark 5 state-of-the-art pre-trained language models on 5 single-document abstractive summarization datasets of different domains. To probe the models, we propose 4 novel task comprehension tests that evaluate themain components of summarizationmodels. Our experiments reveal that pseudo-summarization pre-training biases the models towards more extractive behavior and inhibits their ability to properly filter the salient content, leading to worse generalization.
引用
收藏
页码:3028 / 3039
页数:12
相关论文
共 50 条
  • [1] Evaluating the Summarization Comprehension of Pre-Trained Language Models
    D. I. Chernyshev
    B. V. Dobrov
    [J]. Lobachevskii Journal of Mathematics, 2023, 44 : 3028 - 3039
  • [2] Evaluating Commonsense in Pre-Trained Language Models
    Zhou, Xuhui
    Zhang, Yue
    Cui, Leyang
    Huang, Dandan
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 9733 - 9740
  • [3] Modeling Content Importance for Summarization with Pre-trained Language Models
    Xiao, Liqiang
    Lu Wang
    Hao He
    Jin, Yaohui
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3606 - 3611
  • [4] From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Models to Pre-trained Machine Reader
    Xu, Weiwen
    Li, Xin
    Zhang, Wenxuan
    Zhou, Meng
    Lam, Wai
    Si, Luo
    Bing, Lidong
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Evaluating and Inducing Personality in Pre-trained Language Models
    Jiang, Guangyuan
    Xu, Manjie
    Zhu, Song-Chun
    Han, Wenjuan
    Zhang, Chi
    Zhu, Yixin
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [6] Pre-trained language models with domain knowledge for biomedical extractive summarization
    Xie, Qianqian
    Bishop, Jennifer Amy
    Tiwari, Prayag
    Ananiadou, Sophia
    [J]. Knowledge-Based Systems, 2022, 252
  • [7] Pre-trained language models evaluating themselves - A comparative study
    Koch, Philipp
    Assenmacher, Matthias
    Heumann, Christian
    [J]. PROCEEDINGS OF THE THIRD WORKSHOP ON INSIGHTS FROM NEGATIVE RESULTS IN NLP (INSIGHTS 2022), 2022, : 180 - 187
  • [8] Hybrid multi-document summarization using pre-trained language models
    Ghadimi, Alireza
    Beigy, Hamid
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2022, 192
  • [9] Using Pre-Trained Language Models for Abstractive DBPEDIA Summarization: A Comparative Study
    Zahera, Hamada M.
    Vitiugin, Fedor
    Sherif, Mohamed Ahmed
    Castillo, Carlos
    Ngomo, Axel-Cyrille Ngonga
    [J]. KNOWLEDGE GRAPHS: SEMANTICS, MACHINE LEARNING, AND LANGUAGES, 2023, 56 : 19 - 37
  • [10] Somun: entity-centric summarization incorporating pre-trained language models
    Inan, Emrah
    [J]. NEURAL COMPUTING & APPLICATIONS, 2021, 33 (10): : 5301 - 5311