Enhanced Seq2Seq Autoencoder via Contrastive Learning for Abstractive Text Summarization

被引:4
|
作者
Zheng, Chujie [1 ]
Zhang, Kunpeng [2 ]
Wang, Harry Jiannan [1 ]
Fan, Ling [3 ,4 ]
Wang, Zhe [4 ]
机构
[1] Univ Delaware, Newark, DE 19716 USA
[2] Univ Maryland, College Pk, MD 20742 USA
[3] Tongji Univ, Shanghai, Peoples R China
[4] Tezign Com, Shanghai, Peoples R China
关键词
Abstractive Text Summarization; Contrastive Learning; Data Augmentation; Seq2seq;
D O I
10.1109/BigData52589.2021.9671819
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we present a denoising sequence-to-sequence (seq2seq) autoencoder via contrastive learning for abstractive text summarization. Our model adopts a standard Transformer-based architecture with a multi-layer bi-directional encoder and an auto-regressive decoder. To enhance its denoising ability, we incorporate self-supervised contrastive learning along with various sentence-level document augmentation. These two components, seq2seq autoencoder and contrastive learning, are jointly trained through fine-tuning, w hich i mproves t he performance of text summarization with regard to ROUGE scores and human evaluation. We conduct experiments on two datasets and demonstrate that our model outperforms many existing benchmarks and even achieves comparable performance to the state-of-the-art abstractive systems trained with more complex architecture and extensive computation resources.
引用
收藏
页码:1764 / 1771
页数:8
相关论文
共 50 条
  • [41] Residual Seq2Seq model for Building energy management
    Kim, Marie
    Kim, Nae-soo
    Song, YuJin
    Pyo, Cheol Sig
    2019 10TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGY CONVERGENCE (ICTC): ICT CONVERGENCE LEADING THE AUTONOMOUS FUTURE, 2019, : 1126 - 1128
  • [42] Automatic Generation of Pseudocode with Attention Seq2seq Model
    Xu, Shaofeng
    Xiong, Yun
    2018 25TH ASIA-PACIFIC SOFTWARE ENGINEERING CONFERENCE (APSEC 2018), 2018, : 711 - 712
  • [43] Map Matching Based on Seq2Seq with Topology Information
    Bai, Yulong
    Li, Guolian
    Lu, Tianxiu
    Wu, Yadong
    Zhang, Weihan
    Feng, Yidan
    APPLIED SCIENCES-BASEL, 2023, 13 (23):
  • [44] Seq2Seq模型的短期水位预测
    刘艳
    张婷
    康爱卿
    李建柱
    雷晓辉
    水利水电科技进展, 2022, (03) : 57 - 63
  • [45] Smoothing and Shrinking the Sparse Seq2Seq Search Space
    Peters, Ben
    Martins, Andre F. T.
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 2642 - 2654
  • [46] A Seq2Seq Model Improved by Transcendental Learning and Imaged Sequence Samples for Porosity Prediction
    Zhou, Lijian
    Wang, Lijun
    Zhao, Zhiang
    Liu, Yuwei
    Liu, Xiwu
    MATHEMATICS, 2023, 11 (01)
  • [47] Sliding Window Seq2seq Modeling for Engagement Estimation
    Yu, Jun
    Lu, Keda
    Jing, Mohan
    Liang, Ziqi
    Zhang, Bingyuan
    Sun, Jianqing
    Liang, Jiaen
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 9496 - 9500
  • [48] Untargeted Code Authorship Evasion with Seq2Seq Transformation
    Choi, Soohyeon
    Jang, Rhongho
    Nyang, DaeHun
    Mohaisen, David
    arXiv, 2023,
  • [49] SparQL Query Prediction Based on Seq2Seq Model
    Yang D.-H.
    Zou K.-F.
    Wang H.-Z.
    Wang J.-B.
    Ruan Jian Xue Bao/Journal of Software, 2021, 32 (03): : 805 - 817
  • [50] SimCLS: A Simple Framework for Contrastive Learning of Abstractive Summarization
    Liu, Yixin
    Liu, Pengfei
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 1065 - 1072