Abstractive text summarization for Hungarian

被引:7
|
作者
Yang, Zijian Gyozo [1 ,2 ,3 ]
Agocs, Adam [1 ]
Kusper, Gabor [1 ]
Varadi, Tamas [3 ]
机构
[1] Eszterhazy Karoly Univ, Fac Informat, Eger, Hungary
[2] MTA PPKE Hungarian Language Technol Res Grp, Budapest, Hungary
[3] Hungarian Res Ctr Linguist, Budapest, Hungary
来源
关键词
BERT; huBERT; ELECTRA; HILBERT; abstractive summarization; extractive summarization;
D O I
10.33039/ami.2021.04.002
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
In our research we have created a text summarization software tool for Hungarian using multilingual and Hungarian BERT-based models. Two types of text summarization method exist: abstractive and extractive. The abstractive summarization is more similar to human generated summarization. Target summaries may include phrases that the original text does not necessarily contain. This method generates the summarized text by applying keywords that were extracted from the original text. The extractive method summarizes the text by using the most important extracted phrases or sentences from the original text. In our research we have built both abstractive and extractive models for Hungarian. For abstractive models, we have used a multilingual BERT model and Hungarian monolingual BERT models. For extractive summarization, in addition to the BERT models, we have also made experiments with ELECTRA models. We find that the Hungarian monolingual models outperformed the multilingual BERT model in all cases. Furthermore, the ELECTRA small models achieved higher results than some of the BERT models. This result is important because the ELECTRA small models have much fewer parameters and were trained on only 1 GPU within a couple of days. Another important consideration is that the ELECTRA models are much smaller than the BERT models, which is important for the end users. To our best knowledge the first extractive and abstractive summarization systems reported in the present paper are the first such systems for Hungarian.
引用
收藏
页码:299 / 316
页数:18
相关论文
共 50 条
  • [1] An approach to Abstractive Text Summarization
    Huong Thanh Le
    Tien Manh Le
    [J]. 2013 INTERNATIONAL CONFERENCE OF SOFT COMPUTING AND PATTERN RECOGNITION (SOCPAR), 2013, : 371 - 376
  • [2] A Survey on Abstractive Text Summarization
    Moratanch, N.
    Chitrakala, S.
    [J]. PROCEEDINGS OF IEEE INTERNATIONAL CONFERENCE ON CIRCUIT, POWER AND COMPUTING TECHNOLOGIES (ICCPCT 2016), 2016,
  • [3] Survey on Abstractive Text Summarization
    Raphal, Nithin
    Duwarah, Hemanta
    Daniel, Philemon
    [J]. PROCEEDINGS OF THE 2018 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATION AND SIGNAL PROCESSING (ICCSP), 2018, : 513 - 517
  • [4] Abstractive Text Summarization Using Hybrid Technique of Summarization
    Liaqat, Muhammad Irfan
    Hamid, Isma
    Nawaz, Qamar
    Shafique, Nida
    [J]. 2022 14TH INTERNATIONAL CONFERENCE ON COMMUNICATION SOFTWARE AND NETWORKS (ICCSN 2022), 2022, : 141 - 144
  • [5] Dual Encoding for Abstractive Text Summarization
    Yao, Kaichun
    Zhang, Libo
    Du, Dawei
    Luo, Tiejian
    Tao, Lili
    Wu, Yanjun
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (03) : 985 - 996
  • [6] Abstractive text summarization and new large-scale datasets for agglutinative languages Turkish and Hungarian
    Batuhan Baykara
    Tunga Güngör
    [J]. Language Resources and Evaluation, 2022, 56 : 973 - 1007
  • [7] Abstractive text summarization and new large-scale datasets for agglutinative languages Turkish and Hungarian
    Baykara, Batuhan
    Gungor, Tunga
    [J]. LANGUAGE RESOURCES AND EVALUATION, 2022, 56 (03) : 973 - 1007
  • [8] Variational Neural Decoder for Abstractive Text Summarization
    Zhao, Huan
    Cao, Jie
    Xu, Mingquan
    Lu, Jian
    [J]. COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2020, 17 (02) : 537 - 552
  • [9] Reinforcement Learning Models for Abstractive Text Summarization
    Buciumas, Sergiu
    [J]. PROCEEDINGS OF THE 2019 ANNUAL ACM SOUTHEAST CONFERENCE (ACMSE 2019), 2019, : 270 - 271
  • [10] Inducing Causal Structure for Abstractive Text Summarization
    Chen, Lu
    Zhang, Ruqing
    Huang, Wei
    Chen, Wei
    Guo, Jiafeng
    Cheng, Xueqi
    [J]. PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 213 - 223