Unsupervised extractive multi-document summarization method based on transfer learning from BERT multi-task fine-tuning

被引:21
|
作者
Lamsiyah, Salima [1 ]
El Mahdaouy, Abdelkader [3 ]
Ouatik, Said El Alaoui [1 ,2 ]
Espinasse, Bernard [4 ]
机构
[1] Sidi Mohamed Ben Abdellah Univ, FSDM, Lab Informat Signals Automat & Cognitivism, BP 1796, Fez Atlas 30003, Morocco
[2] Ibn Tofail Univ, Natl Sch Appl Sci, Lab Engn Sci, Kenitra, Morocco
[3] Mohammed VI Polytech Univ UM6P, Sch Comp Sci UM6P CS, Ben Guerir, Morocco
[4] Univ Toulon & Var, Aix Marseille Univ, CNRS, LIS,UMR 7020, Toulon, France
关键词
BERT fine-tuning; multi-document summarization; multi-task learning; sentence representation learning; transfer learning; SENTENCE SCORING TECHNIQUES;
D O I
10.1177/0165551521990616
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Text representation is a fundamental cornerstone that impacts the effectiveness of several text summarization methods. Transfer learning using pre-trained word embedding models has shown promising results. However, most of these representations do not consider the order and the semantic relationships between words in a sentence, and thus they do not carry the meaning of a full sentence. To overcome this issue, the current study proposes an unsupervised method for extractive multi-document summarization based on transfer learning from BERT sentence embedding model. Moreover, to improve sentence representation learning, we fine-tune BERT model on supervised intermediate tasks from GLUE benchmark datasets using single-task and multi-task fine-tuning methods. Experiments are performed on the standard DUC'2002-2004 datasets. The obtained results show that our method has significantly outperformed several baseline methods and achieves a comparable and sometimes better performance than the recent state-of-the-art deep learning-based methods. Furthermore, the results show that fine-tuning BERT using multi-task learning has considerably improved the performance.
引用
收藏
页码:164 / 182
页数:19
相关论文
共 50 条
  • [21] An Optimization Algorithm for Extractive Multi-document Summarization Based on Association of Sentences
    Chen, Chun-Hao
    Yang, Yi-Chen
    Lin, Jerry Chun-Wei
    ADVANCES AND TRENDS IN ARTIFICIAL INTELLIGENCE: THEORY AND PRACTICES IN ARTIFICIAL INTELLIGENCE, 2022, 13343 : 460 - 469
  • [22] An Extractive Multi-Document Summarization Technique Based on Fuzzy Logic approach
    Tsoumou, Evrard Stency Larys
    Yang, Shichong
    Lai, Linjing
    Varus, Mbembo Loundou
    2016 INTERNATIONAL CONFERENCE ON NETWORK AND INFORMATION SYSTEMS FOR COMPUTERS (ICNISC), 2016, : 346 - 351
  • [23] A Hybrid Solution To Abstractive Multi-Document Summarization Using Supervised and Unsupervised Learning
    Bhagchandani, Gaurav
    Bodra, Deep
    Gangan, Abhishek
    Mulla, Nikahat
    PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING AND CONTROL SYSTEMS (ICCS), 2019, : 566 - 570
  • [24] Multi-document Extractive Summarization Using Window-based Sentence Representation
    Zhang, Yong
    Er, Meng Joo
    Zhao, Rui
    2015 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2015, : 404 - 410
  • [25] MSCSO: Extractive Multi-document Summarization Based on a New Criterion of Sentences Overlapping
    Khaleghi, Zeynab
    Fakhredanesh, Mohammad
    Hourali, Maryam
    IRANIAN JOURNAL OF SCIENCE AND TECHNOLOGY-TRANSACTIONS OF ELECTRICAL ENGINEERING, 2021, 45 (01) : 195 - 205
  • [26] A semantic approach to extractive multi-document summarization: Applying sentence expansion for tuning of conceptual densities
    Bidoki, Mohammad
    Moosavi, Mohammad R.
    Fakhrahmad, Mostafa
    INFORMATION PROCESSING & MANAGEMENT, 2020, 57 (06)
  • [27] Extractive multi-document summarization using population-based multicriteria optimization
    John, Ansamma
    Premjith, P. S.
    Wilscy, M.
    EXPERT SYSTEMS WITH APPLICATIONS, 2017, 86 : 385 - 397
  • [28] MSCSO: Extractive Multi-document Summarization Based on a New Criterion of Sentences Overlapping
    Zeynab Khaleghi
    Mohammad Fakhredanesh
    Maryam Hourali
    Iranian Journal of Science and Technology, Transactions of Electrical Engineering, 2021, 45 : 195 - 205
  • [29] When to Use Multi-Task Learning vs Intermediate Fine-Tuning for Pre-Trained Encoder Transfer Learning
    Weller, Orion
    Seppi, Kevin
    Gardner, Matt
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): (SHORT PAPERS), VOL 2, 2022, : 272 - 282
  • [30] A decomposition-based multi-objective optimization approach for extractive multi-document text summarization
    Sanchez-Gomez, Jesus M.
    Vega-Rodriguez, Miguel A.
    Perez, Carlos J.
    APPLIED SOFT COMPUTING, 2020, 91