Abstractive Document Summarization via Neural Model with Joint Attention

被引:19
|
作者
Hou, Liwei [1 ]
Hu, Po [1 ]
Bei, Chao [2 ]
机构
[1] Cent China Normal Univ, Sch Comp Sci, Wuhan 430079, Hubei, Peoples R China
[2] Global Tone Commun Technol Co Ltd, Beijing 100043, Peoples R China
基金
中国国家自然科学基金;
关键词
Abstractive summarization; Attentional mechanism; Encoder-decoder framework; Neural network;
D O I
10.1007/978-3-319-73618-1_28
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the difficulty of abstractive summarization, the great majority of past work on document summarization has been extractive, while the recent success of sequence-to-sequence framework has made abstractive summarization viable, in which a set of recurrent neural networks models based on attention encoder-decoder have achieved promising performance on short-text summarization tasks. Unfortunately, these attention encoder-decoder models often suffer from the undesirable shortcomings of generating repeated words or phrases and inability to deal with out-of-vocabulary words appropriately. To address these issues, in this work we propose to add an attention mechanism on output sequence to avoid repetitive contents and use the subword method to deal with the rare and unknown words. We applied our model to the public dataset provided by NLPCC 2017 shared task3. The evaluation results show that our system achieved the best ROUGE performance among all the participating teams and is also competitive with some state-of-the-art methods.
引用
收藏
页码:329 / 338
页数:10
相关论文
共 50 条
  • [1] Neural attention model with keyword memory for abstractive document summarization
    Choi, YunSeok
    Kim, Dahae
    Lee, Jee-Hyong
    [J]. CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2020, 32 (18):
  • [2] Abstractive Summarization by Neural Attention Model with Document Content Memory
    Choi, Yunseok
    Kim, Dahae
    Lee, Jee-Hyong
    [J]. PROCEEDINGS OF THE 2018 CONFERENCE ON RESEARCH IN ADAPTIVE AND CONVERGENT SYSTEMS (RACS 2018), 2018, : 11 - 16
  • [3] Attention Optimization for Abstractive Document Summarization
    Gui, Min
    Tian, Junfeng
    Wang, Rui
    Yang, Zhenglu
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1222 - 1228
  • [4] Abstractive Multi-Document Summarization via Joint Learning with Single-Document Summarization
    Jin, Hanqi
    Wan, Xiaojun
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 2545 - 2554
  • [5] Neural Abstractive Summarization with Structural Attention
    Chowdhury, Tanya
    Kumar, Sachin
    Chakraborty, Tanmoy
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3716 - 3722
  • [6] Attention based Abstractive Summarization of Malayalam Document
    Nambiar, Sindhya K.
    Peter, David S.
    Idicula, Sumam Mary
    [J]. AI IN COMPUTATIONAL LINGUISTICS, 2021, 189 : 250 - 257
  • [7] Abstractive Document Summarization via Bidirectional Decoder
    Wan, Xin
    Li, Chen
    Wang, Ruijia
    Xiao, Ding
    Shi, Chuan
    [J]. ADVANCED DATA MINING AND APPLICATIONS, ADMA 2018, 2018, 11323 : 364 - 377
  • [8] Abstractive Document Summarization with a Graph-Based Attentional Neural Model
    Tan, Jiwei
    Wan, Xiaojun
    Xiao, Jianguo
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 1171 - 1181
  • [9] MeanSum : A Neural Model for Unsupervised Multi-Document Abstractive Summarization
    Chu, Eric
    Liu, Peter J.
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [10] Topic Attentional Neural Network for Abstractive Document Summarization
    Liu, Hao
    Zheng, Hai-Tao
    Wang, Wei
    [J]. ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2019, PT II, 2019, 11440 : 70 - 81