Enhanced Seq2Seq Autoencoder via Contrastive Learning for Abstractive Text Summarization

被引:4
|
作者
Zheng, Chujie [1 ]
Zhang, Kunpeng [2 ]
Wang, Harry Jiannan [1 ]
Fan, Ling [3 ,4 ]
Wang, Zhe [4 ]
机构
[1] Univ Delaware, Newark, DE 19716 USA
[2] Univ Maryland, College Pk, MD 20742 USA
[3] Tongji Univ, Shanghai, Peoples R China
[4] Tezign Com, Shanghai, Peoples R China
关键词
Abstractive Text Summarization; Contrastive Learning; Data Augmentation; Seq2seq;
D O I
10.1109/BigData52589.2021.9671819
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we present a denoising sequence-to-sequence (seq2seq) autoencoder via contrastive learning for abstractive text summarization. Our model adopts a standard Transformer-based architecture with a multi-layer bi-directional encoder and an auto-regressive decoder. To enhance its denoising ability, we incorporate self-supervised contrastive learning along with various sentence-level document augmentation. These two components, seq2seq autoencoder and contrastive learning, are jointly trained through fine-tuning, w hich i mproves t he performance of text summarization with regard to ROUGE scores and human evaluation. We conduct experiments on two datasets and demonstrate that our model outperforms many existing benchmarks and even achieves comparable performance to the state-of-the-art abstractive systems trained with more complex architecture and extensive computation resources.
引用
收藏
页码:1764 / 1771
页数:8
相关论文
共 50 条
  • [1] Abstractive Summarization Model with a Feature-Enhanced Seq2Seq Structure
    Hao, Zepeng
    Ji, Jingzhou
    Xie, Tao
    Xue, Bin
    2020 5TH ASIA-PACIFIC CONFERENCE ON INTELLIGENT ROBOT SYSTEMS (ACIRS 2020), 2020, : 163 - 167
  • [2] Abstractive social media text summarization using selective reinforced Seq2Seq attention model
    Liang, Zeyu
    Du, Junping
    Li, Chaoyang
    NEUROCOMPUTING, 2020, 410 : 432 - 440
  • [3] Abstract Text Summarization with a Convolutional Seq2seq Model
    Zhang, Yong
    Li, Dan
    Wang, Yuheng
    Fang, Yang
    Xiao, Weidong
    APPLIED SCIENCES-BASEL, 2019, 9 (08):
  • [4] CFCSS : Based on CF Network Convolutional Seq2Seq Model for Abstractive Summarization
    Liang, Qingmin
    Lu, Ling
    Chang, Tianji
    Yang, Wu
    PROCEEDINGS OF THE 15TH IEEE CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA 2020), 2020, : 1160 - 1164
  • [5] A Hierarchical Attention Seq2seq Model with CopyNet for Text Summarization
    Zhang, Yong
    Wang, Yuheng
    Liao, Jinzhi
    Xiao, Weidong
    2018 INTERNATIONAL CONFERENCE ON ROBOTS & INTELLIGENT SYSTEM (ICRIS 2018), 2018, : 316 - 320
  • [6] Enhanced matrix inference with Seq2seq models via diagonal sorting
    Wei Peng
    Yisong Wang
    Maonian Wu
    Scientific Reports, 14
  • [7] Sentence salience contrastive learning for abstractive text summarization
    Huang, Ying
    Li, Zhixin
    Chen, Zhenbin
    Zhang, Canlong
    Ma, Huifang
    NEUROCOMPUTING, 2024, 593
  • [8] Enhanced matrix inference with Seq2seq models via diagonal sorting
    Peng, Wei
    Wang, Yisong
    Wu, Maonian
    SCIENTIFIC REPORTS, 2024, 14 (01)
  • [9] A Chinese text corrector based on seq2seq model
    Gu, Sunyan
    Lang, Fei
    2017 INTERNATIONAL CONFERENCE ON CYBER-ENABLED DISTRIBUTED COMPUTING AND KNOWLEDGE DISCOVERY (CYBERC), 2017, : 322 - 325
  • [10] A Study on Hierarchical Text Classification as a Seq2seq Task
    Torba, Fatos
    Gravier, Christophe
    Laclau, Charlotte
    Kammoun, Abderrhammen
    Subercaze, Julien
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT III, 2024, 14610 : 287 - 296