Data Selection Curriculum for Abstractive Text Summarization

被引:0
|
作者
Sun, Shichao [1 ]
Yuan, Ruifeng [1 ]
He, Jianfei [2 ]
Cao, Ziqiang [3 ]
Li, Wenjie [1 ]
Jia, Xiaohua [2 ]
机构
[1] Hong Kong Polytech Univ, Hong Kong, Peoples R China
[2] City Univ Hong Kong, Hong Kong, Peoples R China
[3] Soochow Univ, Suzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Abstractive Text Summarization (ATS) models are commonly trained using large-scale data that is randomly shuffled. However, the impact of data selection and data ordering on ATS models remains a relatively unexplored research area, where a significant challenge lies in accurately assessing the learning difficulty of each training instance. This study introduces a Data Selection Curriculum (DSC) scoring system that incorporates both the difficulty of improving ATS model via an instance and the expected performance on this instance. By selectively excluding excessively simple and overly complex instances, the training efficiency can be optimized. Furthermore, curriculum learning is integrated to accelerate convergence and improve performance by gradually increasing the learning difficulty, inspired by human learners. Experimental results on the CNN/DailyMail dataset demonstrate that our approach surpasses potent baselines, utilizing a mere 20% of the available instances.
引用
收藏
页码:7990 / 7995
页数:6
相关论文
共 50 条
  • [31] Sentence salience contrastive learning for abstractive text summarization
    Huang, Ying
    Li, Zhixin
    Chen, Zhenbin
    Zhang, Canlong
    Ma, Huifang
    NEUROCOMPUTING, 2024, 593
  • [32] Abstractive Arabic Text Summarization Based on Deep Learning
    Wazery, Y. M.
    Saleh, Marwa E.
    Alharbi, Abdullah
    Ali, Abdelmgeid A.
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [33] IWM-LSTM encoder for abstractive text summarization
    Gangundi R.
    Sridhar R.
    Multimedia Tools and Applications, 2025, 84 (09) : 5883 - 5904
  • [34] Turkish abstractive text document summarization using text to text transfer transformer
    Ay, Betul
    Ertam, Fatih
    Fidan, Guven
    Aydin, Galip
    ALEXANDRIA ENGINEERING JOURNAL, 2023, 68 : 1 - 13
  • [35] Multi-Fact Correction in Abstractive Text Summarization
    Dong, Yue
    Wang, Shuohang
    Gan, Zhe
    Cheng, Yu
    Cheung, Jackie Chi Kit
    Liu, Jingjing
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 9320 - 9331
  • [36] A Novel Framework for Semantic Oriented Abstractive Text Summarization
    Moratanch, N.
    Chitrakala, S.
    JOURNAL OF WEB ENGINEERING, 2018, 17 (08): : 675 - 716
  • [37] Neural Abstractive Summarization for Long Text and Multiple Tables
    Liu, Shuaiqi
    Cao, Jiannong
    Deng, Zhongfen
    Zhao, Wenting
    Yang, Ruosong
    Wen, Zhiyuan
    Yu, Philip S.
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (06) : 2572 - 2586
  • [38] Abstractive Text Summarization Using the BRIO Training Paradigm
    Khang Nhut Lam
    Thieu Gia Doan
    Khang Thua Pham
    Kalita, Jugal
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 92 - 99
  • [39] Keyword-Aware Encoder for Abstractive Text Summarization
    Hu, Tianxiang
    Liang, Jingxi
    Ye, Wei
    Zhang, Shikun
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2021), PT II, 2021, 12682 : 37 - 52
  • [40] Abstractive Text Summarization with Application to Bulgarian News Articles
    Taushanov, Nikola
    Koychev, Ivan
    Nakov, Preslav
    PROCEEDINGS OF THE THIRD INTERNATIONAL CONFERENCE COMPUTATIONAL LINGUISTICS IN BULGARIA (CLIB '18), 2018, : 15 - 22