Extractive Elementary Discourse Units for Improving Abstractive Summarization

被引:0
|
作者
Xiong, Ye [1 ]
Racharak, Teeradaj [1 ]
Minh Le Nguyen [1 ]
机构
[1] Japan Adv Inst Sci & Technol, Nomi, Ishikawa, Japan
关键词
Abstractive summarization; text generation; two-stage summarization;
D O I
10.1145/3477495.3531916
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The Abstractive summarization focuses on generating concise and fluent text from an original document while maintaining the original intent and containing the new words that do not appear in the original document. Recent studies point out that rewriting extractive summaries help improve the performance with a more concise and comprehensible output summary, which uses a sentence as a textual unit. However, a single document sentence normally cannot supply sufficient information. In this paper, we apply elementary discourse unit (EDU) as textual unit of content selection. In order to utilize EDU for generating a high quality summary, we propose a novel summarization model that first designs an EDU selector to choose salient content. Then, the generator model rewrites the selected EDUs as the final summary. To determine the relevancy of each EDU on the entire document, we choose to apply group tag embedding, which can establish the connection between summary sentences and relevant EDUs, so that our generator does not only focus on selected EDUs, but also ingest the entire original document. Extensive experiments on the CNN/Daily Mail dataset have demonstrated the effectiveness of our model.
引用
收藏
页码:2675 / 2679
页数:5
相关论文
共 50 条
  • [31] Question-driven text summarization using an extractive-abstractive framework
    Kia, Mahsa Abazari
    Garifullina, Aygul
    Kern, Mathias
    Chamberlain, Jon
    Jameel, Shoaib
    [J]. COMPUTATIONAL INTELLIGENCE, 2024, 40 (03)
  • [32] Extractive-Abstractive: A Two-Stage Model for Long Text Summarization
    Liang, Rui
    Li, Jianguo
    Huang, Li
    Lin, Ronghua
    Lai, Yu
    Xiong, Dan
    [J]. COMPUTER SUPPORTED COOPERATIVE WORK AND SOCIAL COMPUTING, CHINESECSCW 2021, PT II, 2022, 1492 : 173 - 184
  • [33] Extractive-Abstractive Summarization of Judgment Documents Using Multiple Attention Networks
    Gao, Yan
    Liu, Zhengtao
    Li, Juan
    Guo, Fan
    Xiao, Fei
    [J]. LOGIC AND ARGUMENTATION, CLAR 2021, 2021, 13040 : 486 - 494
  • [34] A comparative study of abstractive and extractive summarization techniques to label subgroups on patent dataset
    Souza, Cinthia M.
    Meireles, Magali R. G.
    Almeida, Paulo E. M.
    [J]. SCIENTOMETRICS, 2021, 126 (01) : 135 - 156
  • [35] Leveraging peer-review aspects for extractive and abstractive summarization of scientific articles
    Majadly, Muhammad
    Last, Mark
    [J]. INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2024,
  • [36] Automatic Multi-Document Summarization for Indonesian Documents Using Hybrid Abstractive-Extractive Summarization Technique
    Yapinus, Glorian
    Erwin, Alva
    Galinium, Maulahikmah
    Muliady, Wahyu
    [J]. 2014 6TH INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY AND ELECTRICAL ENGINEERING (ICITEE), 2014, : 39 - 43
  • [37] Improving abstractive summarization of legal rulings through textual entailment
    Diego de Vargas Feijo
    Viviane P. Moreira
    [J]. Artificial Intelligence and Law, 2023, 31 : 91 - 113
  • [38] Improving extractive document summarization with sentence centrality
    Gong, Shuai
    Zhu, Zhenfang
    Qi, Jiangtao
    Tong, Chunling
    Lu, Qiang
    Wu, Wenqing
    [J]. PLOS ONE, 2022, 17 (07):
  • [39] A Study of Abstractive Summarization Using Semantic Representations and Discourse Level Information
    Valderrama Vilca, Gregory Cesar
    Sobrevilla Cabezudo, Marco Antonio
    [J]. TEXT, SPEECH, AND DIALOGUE, TSD 2017, 2017, 10415 : 482 - 490
  • [40] Improving Faithfulness in Abstractive Summarization with Contrast Candidate Generation and Selection
    Chen, Sihao
    Zhang, Fan
    Sone, Kazoo
    Roth, Dan
    [J]. 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 5935 - 5941