Enhanced Seq2Seq Autoencoder via Contrastive Learning for Abstractive Text Summarization

被引:4
|
作者
Zheng, Chujie [1 ]
Zhang, Kunpeng [2 ]
Wang, Harry Jiannan [1 ]
Fan, Ling [3 ,4 ]
Wang, Zhe [4 ]
机构
[1] Univ Delaware, Newark, DE 19716 USA
[2] Univ Maryland, College Pk, MD 20742 USA
[3] Tongji Univ, Shanghai, Peoples R China
[4] Tezign Com, Shanghai, Peoples R China
关键词
Abstractive Text Summarization; Contrastive Learning; Data Augmentation; Seq2seq;
D O I
10.1109/BigData52589.2021.9671819
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we present a denoising sequence-to-sequence (seq2seq) autoencoder via contrastive learning for abstractive text summarization. Our model adopts a standard Transformer-based architecture with a multi-layer bi-directional encoder and an auto-regressive decoder. To enhance its denoising ability, we incorporate self-supervised contrastive learning along with various sentence-level document augmentation. These two components, seq2seq autoencoder and contrastive learning, are jointly trained through fine-tuning, w hich i mproves t he performance of text summarization with regard to ROUGE scores and human evaluation. We conduct experiments on two datasets and demonstrate that our model outperforms many existing benchmarks and even achieves comparable performance to the state-of-the-art abstractive systems trained with more complex architecture and extensive computation resources.
引用
收藏
页码:1764 / 1771
页数:8
相关论文
共 50 条
  • [21] Knowledge-based Questions Generation with Seq2Seq Learning
    Tang, Xiangru
    Gao, Hanning
    Gao, Junjie
    PROCEEDINGS OF THE 2018 IEEE INTERNATIONAL CONFERENCE ON PROGRESS IN INFORMATICS AND COMPUTING (PIC), 2018, : 180 - 184
  • [22] Open-Domain Table-to-Text Generation based on Seq2seq
    Cao, Juan
    Gong, Junpeng
    Zhang, Pengzhou
    2018 INTERNATIONAL CONFERENCE ON ALGORITHMS, COMPUTING AND ARTIFICIAL INTELLIGENCE (ACAI 2018), 2018,
  • [23] Sparsing and Smoothing for the seq2seq Models
    Zhao S.
    Liang Z.
    Wen J.
    Chen J.
    IEEE Transactions on Artificial Intelligence, 2023, 4 (03): : 464 - 472
  • [24] Seq2Seq Imitation Learning for Tactile Feedback-based Manipulation
    Yang, Wenyan
    Angleraud, Alexandre
    Pieters, Roel S.
    Pajarinen, Joni
    Kamarainen, Joni-Kristian
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 5829 - 5836
  • [25] Linguistic Descriptions of Human Motion with Generative Adversarial Seq2Seq Learning
    Goutsu, Yusuke
    Inamura, Tetsunari
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 4281 - 4287
  • [26] Fine Grained Named Entity Recognition via Seq2seq Framework
    Zhu, Huiming
    He, Chunhui
    Fang, Yang
    Xiao, Weidong
    IEEE ACCESS, 2020, 8 : 53953 - 53961
  • [27] Energy Investment Risk Assessment for Nations Via Seq2seq Model
    Liang, Panning
    Yu, Mingyang
    Jiang, Lincheng
    FRONTIERS IN ENVIRONMENTAL SCIENCE, 2021, 9
  • [28] Unsupervised Abstractive Text Summarization with Length Controlled Autoencoder
    Dugar, Abhinav
    Singh, Gaurav
    Navyasree, B.
    Kumar, Anand M.
    2022 IEEE 19TH INDIA COUNCIL INTERNATIONAL CONFERENCE, INDICON, 2022,
  • [29] Powering Predictive Treatment Planning by a Seq2seq Deep Learning Predictor
    Lee, D.
    Hu, Y.
    Kuo, L.
    Alam, S.
    Yorke, E.
    Rimner, A.
    Zhang, P.
    MEDICAL PHYSICS, 2021, 48 (06)
  • [30] Learning the Dyck Language with Attention-based Seq2Seq Models
    Yu, Xiang
    Ngoc Thang Vu
    Kuhn, Jonas
    BLACKBOXNLP WORKSHOP ON ANALYZING AND INTERPRETING NEURAL NETWORKS FOR NLP AT ACL 2019, 2019, : 138 - 146