Incorporating word attention with convolutional neural networks for abstractive summarization

被引:0
|
作者
Chengzhe Yuan
Zhifeng Bao
Mark Sanderson
Yong Tang
机构
[1] South China Normal University,School of Computer Science
[2] RMIT University,School of Science, Computer Science and Information Technology
来源
World Wide Web | 2020年 / 23卷
关键词
Abstractive summarization; Word attention; Convolutional neural networks; Sequence-to-sequence model;
D O I
暂无
中图分类号
学科分类号
摘要
Neural sequence-to-sequence (seq2seq) models have been widely used in abstractive summarization tasks. One of the challenges of this task is redundant contents in the input document often confuses the models and leads to poor performance. An efficient way to solve this problem is to select salient information from the input document. In this paper, we propose an approach that incorporates word attention with multilayer convolutional neural networks (CNNs) to extend a standard seq2seq model for abstractive summarization. First, by concentrating on a subset of source words during encoding an input sentence, word attention is able to extract informative keywords in the input, which gives us the ability to interpret generated summaries. Second, these keywords are further distilled by multilayer CNNs to capture the coarse-grained contextual features of the input sentence. Thus, the combined word attention and multilayer CNNs modules provide a better-learned representation of the input document, which helps the model generate interpretable, coherent and informative summaries in an abstractive summarization task. We evaluate the effectiveness of our model on the English Gigaword, DUC2004 and Chinese summarization dataset LCSTS. Experimental results show the effectiveness of our approach.
引用
收藏
页码:267 / 287
页数:20
相关论文
共 50 条
  • [1] Incorporating word attention with convolutional neural networks for abstractive summarization
    Yuan, Chengzhe
    Bao, Zhifeng
    Sanderson, Mark
    Tang, Yong
    [J]. WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2020, 23 (01): : 267 - 287
  • [2] Gated Graph Neural Attention Networks for abstractive summarization
    Liang, Zeyu
    Du, Junping
    Shao, Yingxia
    Ji, Houye
    [J]. NEUROCOMPUTING, 2021, 431 : 128 - 136
  • [3] Abstractive Sentence Summarization with Encoder-Convolutional Neural Networks
    Toi Nguyen
    Toai Le
    Nhi-Thao Tran
    [J]. 2020 12TH INTERNATIONAL CONFERENCE ON KNOWLEDGE AND SYSTEMS ENGINEERING (IEEE KSE 2020), 2020, : 13 - 18
  • [4] Neural Abstractive Summarization with Structural Attention
    Chowdhury, Tanya
    Kumar, Sachin
    Chakraborty, Tanmoy
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3716 - 3722
  • [5] Abstractive Summarization with Keyword and Generated Word Attention
    Wang, Qianlong
    Ren, Jiangtao
    [J]. 2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [6] Controlling Length in Abstractive Summarization Using a Convolutional Neural Network
    Liu, Yizhu
    Luo, Zhiyi
    Zhu, Kenny Q.
    [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4110 - 4119
  • [7] Neural attention model with keyword memory for abstractive document summarization
    Choi, YunSeok
    Kim, Dahae
    Lee, Jee-Hyong
    [J]. CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2020, 32 (18):
  • [8] Abstractive Summarization by Neural Attention Model with Document Content Memory
    Choi, Yunseok
    Kim, Dahae
    Lee, Jee-Hyong
    [J]. PROCEEDINGS OF THE 2018 CONFERENCE ON RESEARCH IN ADAPTIVE AND CONVERGENT SYSTEMS (RACS 2018), 2018, : 11 - 16
  • [9] Abstractive Document Summarization via Neural Model with Joint Attention
    Hou, Liwei
    Hu, Po
    Bei, Chao
    [J]. NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2017, 2018, 10619 : 329 - 338
  • [10] Reducing repetition in convolutional abstractive summarization
    Liu, Yizhu
    Chen, Xinyue
    Luo, Xusheng
    Zhu, Kenny Q.
    [J]. NATURAL LANGUAGE ENGINEERING, 2023, 29 (01) : 81 - 109