Variational Neural Decoder for Abstractive Text Summarization

被引:8
|
作者
Zhao, Huan [1 ]
Cao, Jie [1 ]
Xu, Mingquan [1 ]
Lu, Jian [1 ]
机构
[1] Hunan Univ, Coll Comp Sci & Elect Engn, Changsha 410000, Hunan, Peoples R China
基金
美国国家科学基金会; 国家重点研发计划;
关键词
abstractive summarization; sequence-to-sequence; variational auto-encoder; variation neural inferer;
D O I
10.2298/CSIS200131012Z
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the conventional sequence-to-sequence (seq2seq) model for abstrac-tive summarization, the internal transformation structure of recurrent neural net-works (RNNs) is completely determined. Therefore, the learned semantic informa-tion is far from enough to represent all semantic details and context dependencies, resulting in a redundant summary and poor consistency. In this paper, we propose a variational neural decoder text summarization model (VND). The model introduces a series of implicit variables by combining variational RNN and variational auto -encoder, which is used to capture complex semantic representation at each step of decoding. It includes a standard RNN layer and a variational RNN layer [5]. These two network layers respectively generate a deterministic hidden state and a random hidden state. We use these two RNN layers to establish the dependence between implicit variables between adjacent time steps. In this way, the model structure can better capture the complex semantics and the strong dependence between the adja-cent time steps when outputting the summary, thereby improving the performance of generating the summary. The experimental results show that, on the text summary LCSTS and English Gigaword dataset, our model has a significant improvement over the baseline model.
引用
收藏
页码:537 / 552
页数:16
相关论文
共 50 条
  • [1] A Novel Attention Mechanism considering Decoder Input for Abstractive Text Summarization
    Niu, Jianwei
    Sun, Mingsheng
    Rodrigues, Joel J. P. C.
    Liu, Xuefeng
    [J]. ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [2] Neural Abstractive Summarization for Long Text and Multiple Tables
    Liu, Shuaiqi
    Cao, Jiannong
    Deng, Zhongfen
    Zhao, Wenting
    Yang, Ruosong
    Wen, Zhiyuan
    Yu, Philip S.
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (06) : 2572 - 2586
  • [3] Multilayer encoder and single-layer decoder for abstractive Arabic text summarization
    Suleiman, Dima
    Awajan, Arafat
    [J]. KNOWLEDGE-BASED SYSTEMS, 2022, 237
  • [4] Keyphrase Guided Beam Search for Neural Abstractive Text Summarization
    Chen, Xuewen
    Li, Jinlong
    Wang, Haihan
    [J]. 2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [5] Neural Abstractive Text Summarization with Sequence-to-Sequence Models
    Shi, Tian
    Keneshloo, Yaser
    Ramakrishnan, Naren
    Reddy, Chandan K.
    [J]. ACM/IMS Transactions on Data Science, 2021, 2 (01):
  • [6] Abstractive text summarization for Hungarian
    Yang, Zijian Gyozo
    Agocs, Adam
    Kusper, Gabor
    Varadi, Tamas
    [J]. ANNALES MATHEMATICAE ET INFORMATICAE, 2021, 53 : 299 - 316
  • [7] A Survey on Abstractive Text Summarization
    Moratanch, N.
    Chitrakala, S.
    [J]. PROCEEDINGS OF IEEE INTERNATIONAL CONFERENCE ON CIRCUIT, POWER AND COMPUTING TECHNOLOGIES (ICCPCT 2016), 2016,
  • [8] An approach to Abstractive Text Summarization
    Huong Thanh Le
    Tien Manh Le
    [J]. 2013 INTERNATIONAL CONFERENCE OF SOFT COMPUTING AND PATTERN RECOGNITION (SOCPAR), 2013, : 371 - 376
  • [9] A Review on Neural network based Abstractive Text Summarization models
    Tandel, Jinal
    Mistree, Kinjal
    Shah, Parth
    [J]. 2019 IEEE 5TH INTERNATIONAL CONFERENCE FOR CONVERGENCE IN TECHNOLOGY (I2CT), 2019,
  • [10] Survey on Abstractive Text Summarization
    Raphal, Nithin
    Duwarah, Hemanta
    Daniel, Philemon
    [J]. PROCEEDINGS OF THE 2018 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATION AND SIGNAL PROCESSING (ICCSP), 2018, : 513 - 517