GreekT5: Sequence-to-Sequence Models for Greek News Summarization

被引:0
|
作者
Giarelis, Nikolaos [1 ]
Mastrokostas, Charalampos [1 ]
Karacapilidis, Nikos [1 ]
机构
[1] Univ Patras, Ind Management & Informat Syst Lab, MEAD, Rion, Greece
关键词
Deep Learning; Natural Language Processing; Greek NLP; Text Summarization; Pretrained Language Models; Greek Language;
D O I
10.1007/978-3-031-63215-0_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text summarization is a natural language processing subtask pertaining to the automatic formulation of a concise and coherent summary that covers the major concepts and topics from one or multiple documents. Recent advancements in deep learning have led to the development of abstractive summarization Transformer-based models, which outperform classical approaches. In any case, research in this field focuses on high resource languages such as English, while the corresponding work for low resource languages is limited. Dealing with modern Greek, this paper proposes a series of new abstractive models for news article summarization. The proposed models were thoroughly evaluated on the same dataset against GreekBART, the only existing model for Greek abstractive news summarization. Our evaluation results reveal that most of the proposed models perform better than GreekBART on various evaluation metrics. Our experiments indicate that multilingual Seq2Seq models, fine-tuned for a specific language and task, can achieve similar or even better performance compared to monolingual models pre-trained and fine-tuned for the same language and task, while requiring significantly less computational resources. We make our evaluation code public, aiming to increase the reproducibility of this work and facilitate future research in the field.
引用
收藏
页码:60 / 73
页数:14
相关论文
共 50 条
  • [41] Guiding Attention in Sequence-to-Sequence Models for Dialogue Act prediction
    Colombo, Pierre
    Chapuis, Emile
    Manica, Matteo
    Vignon, Emmanuel
    Varni, Giovanna
    Clavel, Chloe
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7594 - 7601
  • [42] STATE-OF-THE-ART SPEECH RECOGNITION WITH SEQUENCE-TO-SEQUENCE MODELS
    Chiu, Chung-Cheng
    Sainath, Tara N.
    Wu, Yonghui
    Prabhavalkar, Rohit
    Nguyen, Patrick
    Chen, Zhifeng
    Kannan, Anjuli
    Weiss, Ron J.
    Rao, Kanishka
    Gonina, Ekaterina
    Jaitly, Navdeep
    Li, Bo
    Chorowski, Jan
    Bacchiani, Michiel
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 4774 - 4778
  • [43] COUPLED TRAINING OF SEQUENCE-TO-SEQUENCE MODELS FOR ACCENTED SPEECH RECOGNITION
    Unni, Vinit
    Joshi, Nitish
    Jyothi, Preethi
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 8254 - 8258
  • [44] BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese
    Nguyen Luong Tran
    Duong Minh Le
    Dat Quoc Nguyen
    INTERSPEECH 2022, 2022, : 1751 - 1755
  • [45] Multitask Sequence-to-Sequence Models for Grapheme-to-Phoneme Conversion
    Milde, Benjamin
    Schmidt, Christoph
    Koehler, Joachim
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 2536 - 2540
  • [46] Reformulating natural language queries using sequence-to-sequence models
    Xiaoyu LIU
    Shunda PAN
    Qi ZHANG
    Yu-Gang JIANG
    Xuanjing HUANG
    ScienceChina(InformationSciences), 2019, 62 (12) : 254 - 256
  • [47] EvoLSTM: context-dependent models of sequence evolution using a sequence-to-sequence LSTM
    Lim, Dongjoon
    Blanchette, Mathieu
    BIOINFORMATICS, 2020, 36 : 353 - 361
  • [48] Sequence-to-Sequence Language Models for Character and Emotion Detection in Dream Narratives
    Cortal, Gustave
    TRAITEMENT AUTOMATIQUE DES LANGUES, 2024, 65 (01): : 11 - 35
  • [49] Evaluating sequence-to-sequence models for simulating medical staff mobility on time
    Khovrichev, Mikhail A.
    Balakhontceva, Marina A.
    Ionov, Mikhail, V
    7TH INTERNATIONAL YOUNG SCIENTISTS CONFERENCE ON COMPUTATIONAL SCIENCE, YSC2018, 2018, 136 : 425 - 432
  • [50] Sequence-to-sequence models for workload interference prediction on batch processing datacenters
    Buchaca, David
    Marcual, Joan
    LLuis Berral, Josep
    Carrera, David
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2020, 110 (110): : 155 - 166