GreekT5: Sequence-to-Sequence Models for Greek News Summarization

被引:0
|
作者
Giarelis, Nikolaos [1 ]
Mastrokostas, Charalampos [1 ]
Karacapilidis, Nikos [1 ]
机构
[1] Univ Patras, Ind Management & Informat Syst Lab, MEAD, Rion, Greece
关键词
Deep Learning; Natural Language Processing; Greek NLP; Text Summarization; Pretrained Language Models; Greek Language;
D O I
10.1007/978-3-031-63215-0_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text summarization is a natural language processing subtask pertaining to the automatic formulation of a concise and coherent summary that covers the major concepts and topics from one or multiple documents. Recent advancements in deep learning have led to the development of abstractive summarization Transformer-based models, which outperform classical approaches. In any case, research in this field focuses on high resource languages such as English, while the corresponding work for low resource languages is limited. Dealing with modern Greek, this paper proposes a series of new abstractive models for news article summarization. The proposed models were thoroughly evaluated on the same dataset against GreekBART, the only existing model for Greek abstractive news summarization. Our evaluation results reveal that most of the proposed models perform better than GreekBART on various evaluation metrics. Our experiments indicate that multilingual Seq2Seq models, fine-tuned for a specific language and task, can achieve similar or even better performance compared to monolingual models pre-trained and fine-tuned for the same language and task, while requiring significantly less computational resources. We make our evaluation code public, aiming to increase the reproducibility of this work and facilitate future research in the field.
引用
收藏
页码:60 / 73
页数:14
相关论文
共 50 条
  • [31] Generating Bengali News Headlines: An Attentive Approach with Sequence-to-Sequence Networks
    Salehin, Mushfiqus
    Rafat, Ashik Ahamed Aman
    Khan, Fazle Rabby
    Abujar, Sheikh
    PROCEEDINGS OF THE 2019 8TH INTERNATIONAL CONFERENCE ON SYSTEM MODELING & ADVANCEMENT IN RESEARCH TRENDS (SMART-2019), 2019, : 256 - 261
  • [32] Improving Semantic Relevance for Sequence-to-Sequence Learning of Chinese Social Media Text Summarization
    Ma, Shuming
    Sun, Xu
    Xu, Jingjing
    Wang, Houfeng
    Li, Wenjie
    Su, Qi
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 2, 2017, : 635 - 640
  • [33] A Reinforced Topic-Aware Convolutional Sequence-to-Sequence Model for Abstractive Text Summarization
    Wang, Li
    Yao, Junlin
    Tao, Yunzhe
    Zhong, Li
    Liu, Wei
    Du, Qiang
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 4453 - 4460
  • [34] Sequence-to-Sequence Models and Their Evaluation for Spoken Language Normalization of Slovenian
    Maucec, Mirjam Sepesy
    Verdonik, Darinka
    Donaj, Gregor
    APPLIED SCIENCES-BASEL, 2024, 14 (20):
  • [35] Sequence-to-Sequence Models Can Directly Translate Foreign Speech
    Weiss, Ron J.
    Chorowski, Jan
    Jaitly, Navdeep
    Wu, Yonghui
    Chen, Zhifeng
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 2625 - 2629
  • [36] Unleashing the True Potential of Sequence-to-Sequence Models for Sequence Tagging and Structure Parsing
    He, Han
    Choi, Jinho D.
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2023, 11 : 582 - 599
  • [37] Retrosynthetic Reaction Prediction Using Neural Sequence-to-Sequence Models
    Liu, Bowen
    Ramsundar, Bharath
    Kawthekar, Prasad
    Shi, Jade
    Gomes, Joseph
    Quang Luu Nguyen
    Ho, Stephen
    Sloane, Jack
    Wender, Paul
    Pande, Vijay
    ACS CENTRAL SCIENCE, 2017, 3 (10) : 1103 - 1113
  • [38] Reformulating natural language queries using sequence-to-sequence models
    Xiaoyu Liu
    Shunda Pan
    Qi Zhang
    Yu-Gang Jiang
    Xuanjing Huang
    Science China Information Sciences, 2019, 62
  • [39] Runoff predictions in ungauged basins using sequence-to-sequence models
    Yin, Hanlin
    Guo, Zilong
    Zhang, Xiuwei
    Chen, Jiaojiao
    Zhang, Yanning
    JOURNAL OF HYDROLOGY, 2021, 603
  • [40] Reformulating natural language queries using sequence-to-sequence models
    Liu, Xiaoyu
    Pan, Shunda
    Zhang, Qi
    Jiang, Yu-Gang
    Huang, Xuanjing
    SCIENCE CHINA-INFORMATION SCIENCES, 2019, 62 (12)