Abstractive Summarization of Broadcast News Stories for Estonian

被引:1
|
作者
Harm, Henry [1 ]
Alumae, Tanel [1 ]
机构
[1] Tallinn Univ Technol, Inst Software Sci, Tallinn, Estonia
来源
BALTIC JOURNAL OF MODERN COMPUTING | 2022年 / 10卷 / 03期
关键词
Abstractive summarization; low-resource languages; pre-trained models; multilingual models; machine-translation;
D O I
10.22364/bjmc.2022.10.3.23
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
We present an approach for generating abstractive summaries for Estonian spoken news stories in a low-resource setting. Given a recording of a radio news story, the goal is to create a summary that captures the essential information in a short format. The approach consists of two steps: automatically generating the transcript and applying a state-of-the-art text summarization system to generate the result. We evaluated a number of models, with the best-performing model leveraging the large English BART model pre-trained on CNN/DailyMail dataset and fine-tuned on machine-translated in-domain data, and with the test data translated to English and back. The method achieved a ROUGE-1 score of 17.22, improving on the alternatives and achieving the best result in human evaluation. The applicability of the proposed solution might be limited in languages where machine translation systems are not mature. In such cases multilingual BART should be considered, which achieved a ROUGE-1 score of 17.00 overall and a score of 16.22 without machine translation based data augmentation.
引用
收藏
页码:511 / 524
页数:14
相关论文
共 50 条
  • [41] Abstractive Summarization Model with Adaptive Sparsemax
    Guo, Shiqi
    Si, Yumeng
    Zhao, Jing
    [J]. NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 810 - 821
  • [42] Abstractive Summarization: A Survey of the State of the Art
    Lin, Hui
    Ng, Vincent
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9815 - 9822
  • [43] Reducing repetition in convolutional abstractive summarization
    Liu, Yizhu
    Chen, Xinyue
    Luo, Xusheng
    Zhu, Kenny Q.
    [J]. NATURAL LANGUAGE ENGINEERING, 2023, 29 (01) : 81 - 109
  • [44] Abstractive summarization: An overview of the state of the art
    Gupta, Som
    Gupta, S. K.
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2019, 121 : 49 - 65
  • [45] ABSUM: ABstractive SUMmarization of Lecture Videos
    Devi, M. S. Karthika
    Bhuvaneshwari, R.
    Baskaran, R.
    [J]. SMART TRENDS IN COMPUTING AND COMMUNICATIONS, VOL 3, SMARTCOM 2024, 2024, 947 : 237 - 248
  • [46] Neural Abstractive Summarization with Structural Attention
    Chowdhury, Tanya
    Kumar, Sachin
    Chakraborty, Tanmoy
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3716 - 3722
  • [47] Learning Cluster Patterns for Abstractive Summarization
    Jo, Sung-Guk
    Park, Seung-Hyeok
    Kim, Jeong-Jae
    On, Byung-Won
    [J]. IEEE ACCESS, 2023, 11 : 146065 - 146075
  • [48] A Combined Extractive With Abstractive Model for Summarization
    Liu, Wenfeng
    Gao, Yaling
    Li, Jinming
    Yang, Yuzhen
    [J]. IEEE ACCESS, 2021, 9 : 43970 - 43980
  • [49] BRIO: Bringing Order to Abstractive Summarization
    Liu, Yixin
    Liu, Pengfei
    Radev, Dragomir
    Neubig, Graham
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 2890 - 2903
  • [50] Attention Optimization for Abstractive Document Summarization
    Gui, Min
    Tian, Junfeng
    Wang, Rui
    Yang, Zhenglu
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1222 - 1228