Enhanced Seq2Seq Autoencoder via Contrastive Learning for Abstractive Text Summarization

被引:4
|
作者
Zheng, Chujie [1 ]
Zhang, Kunpeng [2 ]
Wang, Harry Jiannan [1 ]
Fan, Ling [3 ,4 ]
Wang, Zhe [4 ]
机构
[1] Univ Delaware, Newark, DE 19716 USA
[2] Univ Maryland, College Pk, MD 20742 USA
[3] Tongji Univ, Shanghai, Peoples R China
[4] Tezign Com, Shanghai, Peoples R China
关键词
Abstractive Text Summarization; Contrastive Learning; Data Augmentation; Seq2seq;
D O I
10.1109/BigData52589.2021.9671819
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we present a denoising sequence-to-sequence (seq2seq) autoencoder via contrastive learning for abstractive text summarization. Our model adopts a standard Transformer-based architecture with a multi-layer bi-directional encoder and an auto-regressive decoder. To enhance its denoising ability, we incorporate self-supervised contrastive learning along with various sentence-level document augmentation. These two components, seq2seq autoencoder and contrastive learning, are jointly trained through fine-tuning, w hich i mproves t he performance of text summarization with regard to ROUGE scores and human evaluation. We conduct experiments on two datasets and demonstrate that our model outperforms many existing benchmarks and even achieves comparable performance to the state-of-the-art abstractive systems trained with more complex architecture and extensive computation resources.
引用
收藏
页码:1764 / 1771
页数:8
相关论文
共 50 条
  • [31] MTrajRec: Map-Constrained Trajectory Recovery via Seq2Seq Multi-task Learning
    Ren, Huimin
    Ruan, Sijie
    Li, Yanhua
    Bao, Jie
    Meng, Chuishi
    Li, Ruiyuan
    Zheng, Yu
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 1410 - 1419
  • [32] Application of Seq2Seq Models on Code Correction
    Huang, Shan
    Zhou, Xiao
    Chin, Sang
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2021, 4
  • [33] SEQ2SEQ ATTENTIONAL SIAMESE NEURAL NETWORKS FOR TEXT-DEPENDENT SPEAKER VERIFICATION
    Zhang, Yichi
    Yu, Meng
    Li, Na
    Yu, Chengzhu
    Cui, Jia
    Yu, Dong
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 6131 - 6135
  • [34] A Seq2seq Learning Approach for Modeling Semantic Trajectories and Predicting the Next Location
    Karatzoglou, Antonios
    Jablonski, Adrian
    Beigl, Michael
    26TH ACM SIGSPATIAL INTERNATIONAL CONFERENCE ON ADVANCES IN GEOGRAPHIC INFORMATION SYSTEMS (ACM SIGSPATIAL GIS 2018), 2018, : 528 - 531
  • [35] Neural Question Generation based on Seq2Seq
    Liu, Bingran
    2020 5TH INTERNATIONAL CONFERENCE ON MATHEMATICS AND ARTIFICIAL INTELLIGENCE (ICMAI 2020), 2020, : 119 - 123
  • [36] A Primer on Seq2Seq Models for Generative Chatbots
    Scotti, Vincenzo
    Sbattella, Licia
    Tedesco, Roberto
    ACM COMPUTING SURVEYS, 2024, 56 (03)
  • [37] A seq2seq learning method for microscopic emission estimation of on-road vehicles
    Zhenyi Zhao
    Yang Cao
    Zhenyi Xu
    Yu Kang
    Neural Computing and Applications, 2024, 36 : 8565 - 8576
  • [38] Learning Seq2Seq Model with Dynamic Schema Linking for NL2SQL
    Ning, Xingxing
    Zhao, Yupeng
    Liu, Jie
    CCKS 2022 - EVALUATION TRACK, 2022, 1711 : 148 - 153
  • [39] A seq2seq learning method for microscopic emission estimation of on-road vehicles
    Zhao, Zhenyi
    Cao, Yang
    Xu, Zhenyi
    Kang, Yu
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (15): : 8565 - 8576
  • [40] History-based attention in Seq2Seq model for multi-label text classification
    Xiao, Yaoqiang
    Li, Yi
    Yuan, Jin
    Guo, Songrui
    Xiao, Yi
    Li, Zhiyong
    KNOWLEDGE-BASED SYSTEMS, 2021, 224