Attention History-based Attention for Abstractive Text Summarization

被引:2
|
作者
Lee, Hyunsoo [1 ]
Choi, YunSeok [1 ]
Lee, Jee-Hyong [1 ]
机构
[1] Sungkyunkwan Univ, Dept Software, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
Abstractive Text Summarization; Attention Mechanism; Pointer Mechanism; Accumulation-based Attention;
D O I
10.1145/3341105.3373892
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, encoder-decoder model using attention has shown meaningful results in the abstractive summarization tasks. In the attention mechanism, the attention distribution is generated based only on the current decoder state. However, since there are patterns in the process of writing summaries, patterns will exist even in the process of paying attention. In this work, we propose the attention history-based attention model that considers such patterns of the attention history. We build an additional recurrent network, the attention reader network to model the attention patterns. Also, we employ an accumulation vector that keeps the total amount of effective attention to each part of the input text, which is guided by an additional network named the accumulation network. Both the attention reader network and the accumulation vector are used as the additional inputs to the attention mechanism. The evaluation results on the CNN/Daily Mail dataset show that our method better captures the attention pattern and achieves higher ROUGE scores than strong baselines.
引用
收藏
页码:1075 / 1081
页数:7
相关论文
共 50 条
  • [1] Abstractive Text Summarization Using Enhanced Attention Model
    Roul, Rajendra Kumar
    Joshi, Pratik Madhav
    Sahoo, Jajati Keshari
    [J]. INTELLIGENT HUMAN COMPUTER INTERACTION (IHCI 2019), 2020, 11886 : 63 - 76
  • [2] Abstractive Text Summarization with Multi-Head Attention
    Li, Jinpeng
    Zhang, Chuang
    Chen, Xiaojun
    Cao, Yanan
    Liao, Pengcheng
    Zhang, Peng
    [J]. 2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [3] Attention based Abstractive Summarization of Malayalam Document
    Nambiar, Sindhya K.
    Peter, David S.
    Idicula, Sumam Mary
    [J]. AI IN COMPUTATIONAL LINGUISTICS, 2021, 189 : 250 - 257
  • [4] A Novel Attention Mechanism considering Decoder Input for Abstractive Text Summarization
    Niu, Jianwei
    Sun, Mingsheng
    Rodrigues, Joel J. P. C.
    Liu, Xuefeng
    [J]. ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [5] Attention Optimization for Abstractive Document Summarization
    Gui, Min
    Tian, Junfeng
    Wang, Rui
    Yang, Zhenglu
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1222 - 1228
  • [6] Neural Abstractive Summarization with Structural Attention
    Chowdhury, Tanya
    Kumar, Sachin
    Chakraborty, Tanmoy
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3716 - 3722
  • [7] Neural Attention Model for Abstractive Text Summarization Using Linguistic Feature Space
    Dilawari, Aniqa
    Khan, Muhammad Usman Ghani
    Saleem, Summra
    Zahoor-Ur-Rehman
    Shaikh, Fatema Sabeen
    [J]. IEEE ACCESS, 2023, 11 : 23557 - 23564
  • [8] A Novel Deep Learning Attention Based Sequence to Sequence Model for Automatic Abstractive Text Summarization
    Yousef Methkal Abd Algani
    [J]. International Journal of Information Technology, 2024, 16 (6) : 3597 - 3603
  • [9] See, hear, read: Leveraging multimodality with guided attention for abstractive text summarization
    Atri, Yash Kumar
    Pramanick, Shraman
    Goyal, Vikram
    Chakraborty, Tanmoy
    [J]. KNOWLEDGE-BASED SYSTEMS, 2021, 227
  • [10] Summary-aware attention for social media short text abstractive summarization
    Wang, Qianlong
    Ren, Jiangtao
    [J]. NEUROCOMPUTING, 2021, 425 : 290 - 299