eMLM: A New Pre-training Objective for Emotion Related Tasks

被引:0
|
作者
Sosea, Tiberiu [1 ]
Caragea, Cornelia [1 ]
机构
[1] Univ Illinois, Comp Sci, Chicago, IL 60680 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Bidirectional Encoder Representations from Transformers (BERT) have been shown to be extremely effective on a wide variety of natural language processing tasks, including sentiment analysis and emotion detection. However, the proposed pre-training objectives of BERT do not induce any sentiment or emotion-specific biases into the model. In this paper, we present Emotion Masked Language Modeling, a variation of Masked Language Modeling, aimed at improving the BERT language representation model for emotion detection and sentiment analysis tasks. Using the same pre-training corpora as the original BERT model, Wikipedia and BookCorpus, our BERT variation manages to improve the downstream performance on 4 tasks for emotion detection and sentiment analysis by an average of 1:2% F1. Moreover, Your approach shows an increased performance in our task-specific robustness tests. We make our code and pre-trained model available at https://github.com/tsosea2/eMLM.
引用
收藏
页码:286 / 293
页数:8
相关论文
共 50 条
  • [1] Table Pre-training: A Survey on Model Architectures, Pre-training Objectives, and Downstream Tasks
    Dong, Haoyu
    Cheng, Zhoujun
    He, Xinyi
    Zhou, Mengyu
    Zhou, Anda
    Zhou, Fan
    Liu, Ao
    Han, Shi
    Zhang, Dongmei
    PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, 2022, : 5426 - 5435
  • [2] Automating Code-Related Tasks Through Transformers: The Impact of Pre-training
    Tufano, Rosalia
    Pascarella, Luca
    Bavota, Gabriele
    2023 IEEE/ACM 45TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ICSE, 2023, : 2425 - 2437
  • [3] Evaluating synthetic pre-Training for handwriting processing tasks
    Pippi, Vittorio
    Cascianelli, Silvia
    Baraldi, Lorenzo
    Cucchiara, Rita
    PATTERN RECOGNITION LETTERS, 2023, 172 : 44 - 50
  • [4] Insights into Pre-training via Simpler Synthetic Tasks
    Wu, Yuhuai
    Li, Felix
    Liang, Percy
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [5] Synthetic Pre-Training Tasks for Neural Machine Translation
    He, Zexue
    Blackwood, Graeme
    Panda, Rameswar
    McAuley, Julian
    Feris, Rogerio
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8080 - 8098
  • [6] Unsupervised Pre-training for Temporal Action Localization Tasks
    Zhang, Can
    Yang, Tianyu
    Weng, Junwu
    Cao, Meng
    Wang, Jue
    Zou, Yuexian
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 14011 - 14021
  • [7] DOBF: A Deobfuscation Pre-Training Objective for Programming Languages
    Lachaux, Marie-Anne
    Roziere, Baptiste
    Szafraniec, Marc
    Lample, Guillaume
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [8] Unifying Structure Reasoning and Language Pre-Training for Complex Reasoning Tasks
    Wang, Siyuan
    Wei, Zhongyu
    Xu, Jiarong
    Li, Taishan
    Fan, Zhihao
    IEEE/ACM Transactions on Audio Speech and Language Processing, 2024, 32 : 1586 - 1595
  • [9] Dialogue Specific Pre-training Tasks for Improved Dialogue State Tracking
    An, Jinwon
    Kim, Misuk
    NEURAL PROCESSING LETTERS, 2023, 55 (06) : 7761 - 7776
  • [10] Improving Information Extraction on Business Documents with Specific Pre-training Tasks
    Douzon, Thibault
    Duffner, Stefan
    Garcia, Christophe
    Espinas, Jeremy
    DOCUMENT ANALYSIS SYSTEMS, DAS 2022, 2022, 13237 : 111 - 125