Text Summary Generation Model Based on Sentence Fusion and Self-Supervised Training

被引:0
|
作者
Zou A. [1 ]
Hao W. [1 ]
Jin D. [1 ]
Chen G. [1 ]
机构
[1] College of Command and Control Engineering, Army Engineering University of PLA, Nanjing
基金
中国国家自然科学基金;
关键词
Automatic Text Summarization; Deep Neural Network; Pre-trained Language Model; Self-Supervised Training; Sentence Fusion;
D O I
10.16451/j.cnki.issn1003-6059.202205002
中图分类号
学科分类号
摘要
To improve the capability of sentence fusion of deep neural network text generation technique, a text summary generation model based on sentence fusion and self-supervised training is proposed. Before the model training, the training data are firstly pre-processed according to the concept of points of correspondence in the theory of sentence fusion, and thus the data can meet the needs of model training. The training of the proposed model falls into two parts. In the first stage, according to the distribution of the sentence fusion phenomenon in the dataset, the training task of the permutation language model is designed with the points of correspondence as the minimum semantic unit to enhance the ability to capture the information of the fused sentence context. In the second stage, an attention masking strategy based on the fusion information is utilized to control the information intake of the model during the text generation process to enhance the fusion ability in the text generation stage. Experiments on the open dataset show that the proposed model is superior in several evaluation metrics, including those based on statistics, deep semantics and sentence fusion ratio. © 2022, Science Press. All right reserved.
引用
收藏
页码:401 / 411
页数:10
相关论文
共 48 条
  • [1] LUHN H P., The Automatic Creation of Literature Abstracts, IBM Journal of Research and Development, 2, 2, pp. 159-165, (1958)
  • [2] LI J P, ZHANG C, CHEN X J, Et al., Survey on Automatic Text Summarization, Journal of Computer Research and Development, 58, 1, pp. 1-21, (2021)
  • [3] CAO Z Q, WEI F R, DONG L, Et al., Ranking with Recursive Neural Networks and Its Application to Multi-document Summarization, Proc of the 29th AAAI Conference on Artificial Intelligence, pp. 2153-2159, (2015)
  • [4] CHENG J P, LAPATA M., Neural Summarization by Extracting Sentences and Words
  • [5] NALLAPATI R, ZHAI F F, ZHOU B W., SummaRuNNer: A Recurrent Neural Network Based Sequence Model for Extractive Summarization of Documents, Proc of the 31st AAAI Conference on Artificial Intelligence, pp. 3075-3081, (2017)
  • [6] SEE A, LIU P J, MANNING C D., Get to the Point: Summarization with Pointer-Generator Networks
  • [7] KRYSCINSKI W, PAULUS R, XIONG C M, Et al., Improving Abstraction in Text Summarization, Proc of the Conference on Empirical Methods in Natural Language Processing, pp. 1808-1817, (2018)
  • [8] CHEN Y C, BANSAL M., Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting, Proc of the 56th Annual Meeting of the Association for Computational Linguistics(Long Papers), pp. 675-686, (2018)
  • [9] GEHRMANN S, DENG Y T, RUSH A M., Bottom-Up Abstractive Summarization, Proc of the Conference on Empirical Methods in Natural Language Processing, pp. 4098-4109, (2018)
  • [10] CELIKYILMAZ A, BOSSELUT A, HE X D, Et al., Deep Communicating Agents for Abstractive Summarization, Proc of the NAACL-HLT 2018, pp. 1662-1675, (2018)