Enhancing Pre-trained Language Models by Self-supervised Learning for Story Cloze Test

被引:1
|
作者
Xie, Yuqiang [1 ,2 ]
Hu, Yue [1 ,2 ]
Xing, Luxi [1 ,2 ]
Wang, Chunhui [3 ]
Hu, Yong [3 ]
Wei, Xiangpeng [1 ,2 ]
Sun, Yajing [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[3] Effy Intelligent Technol Beijing Co Ltd, Beijing, Peoples R China
关键词
Self-supervised learning; Story comprehension; Story; Cloze Test;
D O I
10.1007/978-3-030-55130-8_24
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Story Cloze Test (SCT) gains increasing attention in evaluating the ability of story comprehension, which requires a story comprehension model to select the correct ending to a story context from two candidate endings. Recent advances, such as GPT and BERT, have shown success in incorporating a pre-trained transformer language model and fine-tuning operation to improve SCT. However, this framework still has some fundamental problems in effectively incorporating story-level knowledge from related corpus. In this paper, we introduce three self-supervised learning tasks (Drop, Replace and TOV) to transfer the story-level knowledge of ROCStories into the backbone model including vanilla BERT and Multi-Choice Head. We evaluate our approach on both SCT-v1.0 and SCT-v1.5 benchmarks. The experimental results demonstrate that our approach achieves state-of-the-art results compared with baseline models.
引用
收藏
页码:271 / 279
页数:9
相关论文
共 50 条
  • [1] A Systematic Review of Transformer-Based Pre-Trained Language Models through Self-Supervised Learning
    Kotei, Evans
    Thirunavukarasu, Ramkumar
    [J]. INFORMATION, 2023, 14 (03)
  • [2] BadEncoder: Backdoor Attacks to Pre-trained Encoders in Self-Supervised Learning
    Jia, Jinyuan
    Liu, Yupei
    Gong, Neil Zhenqiang
    [J]. 43RD IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP 2022), 2022, : 2043 - 2059
  • [3] From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Models to Pre-trained Machine Reader
    Xu, Weiwen
    Li, Xin
    Zhang, Wenxuan
    Zhou, Meng
    Lam, Wai
    Si, Luo
    Bing, Lidong
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [4] Pre-trained Encoders in Self-Supervised Learning Improve Secure and Privacy-preserving Supervised Learning
    Liu, Hongbin
    Qu, Wenjie
    Jia, Jinyuan
    Gong, Neil Zhenqiang
    [J]. PROCEEDINGS 45TH IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS, SPW 2024, 2024, : 144 - 156
  • [6] Improving Speech Separation with Knowledge Distilled from Self-supervised Pre-trained Models
    Qu, Bowen
    Li, Chenda
    Bai, Jinfeng
    Qian, Yanmin
    [J]. 2022 13TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2022, : 329 - 333
  • [7] KNOWLEDGE DISTILLATION FOR NEURAL TRANSDUCERS FROM LARGE SELF-SUPERVISED PRE-TRAINED MODELS
    Yang, Xiaoyu
    Li, Qiujia
    Woodland, Philip C.
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8527 - 8531
  • [8] Self-supervised Learning Based on a Pre-trained Method for the Subtype Classification of Spinal Tumors
    Jiao, Menglei
    Liu, Hong
    Yang, Zekang
    Tian, Shuai
    Ouyang, Hanqiang
    Li, Yuan
    Yuan, Yuan
    Liu, Jianfang
    Wang, Chunjie
    Lang, Ning
    Jiang, Liang
    Yuan, Huishu
    Qian, Yueliang
    Wang, Xiangdong
    [J]. COMPUTATIONAL MATHEMATICS MODELING IN CANCER ANALYSIS, CMMCA 2022, 2022, 13574 : 58 - 67
  • [9] Self-supervised Bidirectional Prompt Tuning for Entity-enhanced Pre-trained Language Model
    Zou, Jiaxin
    Xu, Xianghong
    Hou, Jiawei
    Yang, Qiang
    Zheng, Hai-Tao
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [10] SPIQ: A Self-Supervised Pre-Trained Model for Image Quality Assessment
    Chen, Pengfei
    Li, Leida
    Wu, Qingbo
    Wu, Jinjian
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 513 - 517