Self-supervised Pre-training and Semi-supervised Learning for Extractive Dialog Summarization

被引:0
|
作者
Zhuang, Yingying [1 ]
Song, Jiecheng [1 ]
Sadagopan, Narayanan [1 ]
Beniwal, Anurag [1 ]
机构
[1] Amazon, San Francisco, CA 94107 USA
关键词
summarization; twitter; dialog; self-supervised pre-training; semi-supervised learning;
D O I
10.1145/3543873.3587680
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Language model pre-training has led to state-of-the-art performance in text summarization. While a variety of pre-trained transformer models are available nowadays, they are mostly trained on documents. In this study we introduce self-supervised pre-training to enhance the BERT model's semantic and structural understanding of dialog texts from social media. We also propose a semi-supervised teacher-student learning framework to address the common issue of limited available labels in summarization datasets. We empirically evaluate our approach on extractive summarization task with the TWEETSUMM corpus, a recently introduced dialog summarization dataset from Twitter customer care conversations and demonstrate that our self-supervised pre-training and semi-supervised teacher-student learning are both beneficial in comparison to other pre-trained models. Additionally, we compare pre-training and teacher-student learning in various low data-resource settings, and find that pre-training outperforms teacher-student learning and the differences between the two are more significant when the available labels are scarce.
引用
收藏
页码:1069 / 1076
页数:8
相关论文
共 50 条
  • [41] Correlational Image Modeling for Self-Supervised Visual Pre-Training
    Li, Wei
    Xie, Jiahao
    Loy, Chen Change
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 15105 - 15115
  • [42] MEASURING THE IMPACT OF DOMAIN FACTORS IN SELF-SUPERVISED PRE-TRAINING
    Sanabria, Ramon
    Wei-Ning, Hsu
    Alexei, Baevski
    Auli, Michael
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW, 2023,
  • [43] Contrastive Self-Supervised Pre-Training for Video Quality Assessment
    Chen, Pengfei
    Li, Leida
    Wu, Jinjian
    Dong, Weisheng
    Shi, Guangming
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 458 - 471
  • [44] SSGait: enhancing gait recognition via semi-supervised self-supervised learning
    Xi, Hao
    Ren, Kai
    Lu, Peng
    Li, Yongqiang
    Hu, Chuanping
    [J]. APPLIED INTELLIGENCE, 2024, 54 (07) : 5639 - 5657
  • [45] AN ADAPTER BASED PRE-TRAINING FOR EFFICIENT AND SCALABLE SELF-SUPERVISED SPEECH REPRESENTATION LEARNING
    Kessler, Samuel
    Thomas, Bethan
    Karout, Salah
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3179 - 3183
  • [46] Self-supervised graph neural network with pre-training generative learning for recommendation systems
    Min, Xin
    Li, Wei
    Yang, Jinzhao
    Xie, Weidong
    Zhao, Dazhe
    [J]. SCIENTIFIC REPORTS, 2022, 12 (01)
  • [47] Deep learning based on self-supervised pre-training: Application on sandstone content prediction
    Wang, Chong Ming
    Wang, Xing Jian
    Chen, Yang
    Wen, Xue Mei
    Zhang, Yong Heng
    Li, Qing Wu
    [J]. FRONTIERS IN EARTH SCIENCE, 2023, 10
  • [48] Self-supervised graph neural network with pre-training generative learning for recommendation systems
    Xin Min
    Wei Li
    Jinzhao Yang
    Weidong Xie
    Dazhe Zhao
    [J]. Scientific Reports, 12
  • [49] LPCL: Localized prominence contrastive learning for self-supervised dense visual pre-training
    Chen, Zihan
    Zhu, Hongyuan
    Cheng, Hao
    Mi, Siya
    Zhang, Yu
    Geng, Xin
    [J]. PATTERN RECOGNITION, 2023, 135
  • [50] Self-supervised learning and semi-supervised learning for multi-sequence medical image classification
    Wang, Yueyue
    Song, Danjun
    Wang, Wentao
    Rao, Shengxiang
    Wang, Xiaoying
    Wang, Manning
    [J]. NEUROCOMPUTING, 2022, 513 : 383 - 394