Abstractive Text Summarization with Multi-Head Attention

被引:1
|
作者
Li, Jinpeng [1 ,2 ]
Zhang, Chuang [1 ]
Chen, Xiaojun [1 ]
Cao, Yanan [1 ]
Liao, Pengcheng [1 ,2 ]
Zhang, Peng [1 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ijcnn.2019.8851885
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we present a novel sequence-to-sequence architecture with multi-head attention for automatic summarization of long text. Summaries generated by previous abstractive methods have the problems of duplicate and missing original information commonly. To address these problems, we propose a multi-head attention summarization (MHAS) model, which uses multi-head attention mechanism to learn relevant information in different representation subspaces. The MHAS model can consider the previously predicted words when generating new words to avoid generating a summary of redundant repetition words. And it can learn the internal structure of the article by adding self-attention layer to the traditional encoder and decoder and make the model better preserve the original information. We also integrate the multi-head attention distribution into pointer network creatively to improve the performance of the model. Experiments are conducted on CNN/Daily Mail dataset, which is a long text English corpora. Experimental results show that our proposed model outperforms the previous extractive and abstractive models.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] VATMAN: Integrating Video-Audio-Text for Multimodal Abstractive SummarizatioN via Crossmodal Multi-Head Attention Fusion
    Baek, Doosan
    Kim, Jiho
    Lee, Hongchul
    [J]. IEEE ACCESS, 2024, 12 : 119174 - 119184
  • [2] VIDEO SUMMARIZATION WITH ANCHORS AND MULTI-HEAD ATTENTION
    Sung, Yi-Lin
    Hong, Cheng-Yao
    Hsu, Yen-Chi
    Liu, Tyng-Luh
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 2396 - 2400
  • [3] Text summarization based on multi-head self-attention mechanism and pointer network
    Qiu, Dong
    Yang, Bing
    [J]. COMPLEX & INTELLIGENT SYSTEMS, 2022, 8 (01) : 555 - 567
  • [4] Text summarization based on multi-head self-attention mechanism and pointer network
    Dong Qiu
    Bing Yang
    [J]. Complex & Intelligent Systems, 2022, 8 : 555 - 567
  • [5] MS-Pointer Network: Abstractive Text Summary Based on Multi-Head Self-Attention
    Guo, Qian
    Huang, Jifeng
    Xiong, Naixue
    Wang, Pan
    [J]. IEEE ACCESS, 2019, 7 : 138603 - 138613
  • [6] Attention History-based Attention for Abstractive Text Summarization
    Lee, Hyunsoo
    Choi, YunSeok
    Lee, Jee-Hyong
    [J]. PROCEEDINGS OF THE 35TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING (SAC'20), 2020, : 1075 - 1081
  • [7] Abstractive Text Summarization Using Enhanced Attention Model
    Roul, Rajendra Kumar
    Joshi, Pratik Madhav
    Sahoo, Jajati Keshari
    [J]. INTELLIGENT HUMAN COMPUTER INTERACTION (IHCI 2019), 2020, 11886 : 63 - 76
  • [8] Multi-Fact Correction in Abstractive Text Summarization
    Dong, Yue
    Wang, Shuohang
    Gan, Zhe
    Cheng, Yu
    Cheung, Jackie Chi Kit
    Liu, Jingjing
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 9320 - 9331
  • [9] On the diversity of multi-head attention
    Li, Jian
    Wang, Xing
    Tu, Zhaopeng
    Lyu, Michael R.
    [J]. NEUROCOMPUTING, 2021, 454 : 14 - 24
  • [10] Hierarchical Gated Convolutional Networks with Multi-Head Attention for Text Classification
    Du, Haizhou
    Qian, Jingu
    [J]. 2018 5TH INTERNATIONAL CONFERENCE ON SYSTEMS AND INFORMATICS (ICSAI), 2018, : 1170 - 1175