Improving Transformer with Sequential Context Representations for Abstractive Text Summarization

被引:20
|
作者
Cai, Tian [1 ,2 ]
Shen, Mengjun [1 ,2 ]
Peng, Huailiang [1 ,2 ]
Jiang, Lei [1 ]
Dai, Qiong [1 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
基金
美国国家科学基金会;
关键词
Transformer; Abstractive summarization;
D O I
10.1007/978-3-030-32233-5_40
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent dominant approaches for abstractive text summarization are mainly RNN-based encoder-decoder framework, these methods usually suffer from the poor semantic representations for long sequences. In this paper, we propose a new abstractive summarization model, called RC-Transformer (RCT). The model is not only capable of learning longterm dependencies, but also addresses the inherent shortcoming of Transformer on insensitivity to word order information. We extend the Transformer with an additional RNN-based encoder to capture the sequential context representations. In order to extract salient information effectively, we further construct a convolution module to filter the sequential context with local importance. The experimental results on Gigaword and DUC-2004 datasets show that our proposed model achieves the state-of-the-art performance, even without introducing external information. In addition, our model also owns an advantage in speed over the RNN-based models.
引用
收藏
页码:512 / 524
页数:13
相关论文
共 50 条
  • [31] Arabic abstractive text summarization using RNN-based and transformer-based architectures
    Bani-Almarjeh, Mohammad
    Kurdy, Mohamad-Bassam
    INFORMATION PROCESSING & MANAGEMENT, 2023, 60 (02)
  • [32] Hie-Transformer: A Hierarchical Hybrid Transformer for Abstractive Article Summarization
    Zhang, Xuewen
    Meng, Kui
    Liu, Gongshen
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT III, 2019, 11955 : 248 - 258
  • [33] Friendly Topic Assistant for Transformer Based Abstractive Summarization
    Wang, Zhengjue
    Duan, Zhibin
    Zhang, Hao
    Wang, Chaojie
    Tian, Long
    Chen, Bo
    Zhou, Mingyuan
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 485 - 497
  • [34] ATSSI: Abstractive Text Summarization using Sentiment Infusion
    Bhargava, Rupal
    Sharma, Yashvardhan
    Sharma, Gargi
    TWELFTH INTERNATIONAL CONFERENCE ON COMMUNICATION NETWORKS, ICCN 2016 / TWELFTH INTERNATIONAL CONFERENCE ON DATA MINING AND WAREHOUSING, ICDMW 2016 / TWELFTH INTERNATIONAL CONFERENCE ON IMAGE AND SIGNAL PROCESSING, ICISP 2016, 2016, 89 : 404 - 411
  • [35] Unsupervised Abstractive Text Summarization with Length Controlled Autoencoder
    Dugar, Abhinav
    Singh, Gaurav
    Navyasree, B.
    Kumar, Anand M.
    2022 IEEE 19TH INDIA COUNCIL INTERNATIONAL CONFERENCE, INDICON, 2022,
  • [36] Abstractive Text Summarization Using Enhanced Attention Model
    Roul, Rajendra Kumar
    Joshi, Pratik Madhav
    Sahoo, Jajati Keshari
    INTELLIGENT HUMAN COMPUTER INTERACTION (IHCI 2019), 2020, 11886 : 63 - 76
  • [37] Abstractive Text Summarization Based on Semantic Alignment Network
    Wu S.
    Huang D.
    Li J.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2021, 57 (01): : 1 - 6
  • [38] Sentence salience contrastive learning for abstractive text summarization
    Huang, Ying
    Li, Zhixin
    Chen, Zhenbin
    Zhang, Canlong
    Ma, Huifang
    NEUROCOMPUTING, 2024, 593
  • [39] Abstractive Arabic Text Summarization Based on Deep Learning
    Wazery, Y. M.
    Saleh, Marwa E.
    Alharbi, Abdullah
    Ali, Abdelmgeid A.
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [40] IWM-LSTM encoder for abstractive text summarization
    Ravindra Gangundi
    Rajeswari Sridhar
    Multimedia Tools and Applications, 2025, 84 (9) : 5883 - 5904