Affect Analysis in Arabic Text: Further Pre-Training Language Models for Sentiment and Emotion

被引:3
|
作者
Alshehri, Wafa [1 ,2 ,3 ]
Al-Twairesh, Nora [1 ,4 ]
Alothaim, Abdulrahman [1 ,2 ]
机构
[1] King Saud Univ, Coll Comp & Informat Sci, STCs Artificial Intelligence Chair, Riyadh 11451, Saudi Arabia
[2] King Saud Univ, Coll Comp & Informat Sci, Dept Informat Syst, Riyadh 11451, Saudi Arabia
[3] King Khalid Univ, Coll Sci & Arts, Dept Comp Sci, Almajarda 63931, Saudi Arabia
[4] King Saud Univ, Coll Comp & Informat Sci, Dept Informat Technol, Riyadh 11451, Saudi Arabia
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 09期
关键词
sentiment analysis; emotion detection; pretrained language models; model adaptation; task-adaptation approach;
D O I
10.3390/app13095609
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
One of the main tasks in the field of natural language processing (NLP) is the analysis of affective states (sentiment and emotional) based on written text, and attempts have improved dramatically in recent years. However, in studies on the Arabic language, machine learning or deep learning algorithms were utilised to analyse sentiment and emotion more often than current pre-trained language models. Additionally, further pre-training the language model on specific tasks (i.e., within-task and cross-task adaptation) has not yet been investigated for Arabic in general, and for the sentiment and emotion task in particular. In this paper, we adapt a BERT-based Arabic pretrained language model for the sentiment and emotion tasks by further pre-training it on a sentiment and emotion corpus. Hence, we developed five new Arabic models: QST, QSR, QSRT, QE3, and QE6. Five sentiment and two emotion datasets spanning both small- and large-resource settings were used to evaluate the developed models. The adaptation approaches significantly enhanced the performance of seven Arabic sentiment and emotion datasets. The developed models showed excellent improvements over the sentiment and emotion datasets, which ranged from 0.15-4.71%.
引用
收藏
页数:26
相关论文
共 50 条
  • [1] Pre-Training Language Models for Identifying Patronizing and Condescending Language: An Analysis
    Perez-Almendros, Carla
    Espinosa-Anke, Luis
    Schockaert, Steven
    [J]. LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 3902 - 3911
  • [2] SKEP: Sentiment Knowledge Enhanced Pre-training for Sentiment Analysis
    Tian, Hao
    Gao, Can
    Xiao, Xinyan
    Liu, Hao
    He, Bolei
    Wu, Hua
    Wang, Haifeng
    Wu, Feng
    [J]. 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 4067 - 4076
  • [3] A Multi-Channel Text Sentiment Analysis Model Integrating Pre-training Mechanism
    Liang, Shengbin
    Jin, Jiangyong
    Du, Wencai
    Qu, Shenming
    [J]. INFORMATION TECHNOLOGY AND CONTROL, 2023, 52 (02): : 263 - 275
  • [4] Sentiment-aware multimodal pre-training for multimodal sentiment analysis
    Ye, Junjie
    Zhou, Jie
    Tian, Junfeng
    Wang, Rui
    Zhou, Jingyi
    Gui, Tao
    Zhang, Qi
    Huang, Xuanjing
    [J]. KNOWLEDGE-BASED SYSTEMS, 2022, 258
  • [5] Vision-Language Pre-Training for Multimodal Aspect-Based Sentiment Analysis
    Ling, Yan
    Yu, Jianfei
    Xia, Rui
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 2149 - 2159
  • [6] Improving the Sample Efficiency of Pre-training Language Models
    Berend, Gabor
    [J]. ERCIM NEWS, 2024, (136): : 38 - 40
  • [7] Research on Pre-Training Models for Tibetan Text with Character Awareness
    Gadeng, Luosang
    Nyima, Tashi
    [J]. Computer Engineering and Applications, 2024, 60 (21) : 127 - 133
  • [8] Vision-Language Pre-Training for Boosting Scene Text Detectors
    Song, Sibo
    Wan, Jianqiang
    Yang, Zhibo
    Tang, Jun
    Cheng, Wenqing
    Bai, Xiang
    Yao, Cong
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 15660 - 15670
  • [9] CLAP: Contrastive Language-Audio Pre-training Model for Multi-modal Sentiment Analysis
    Zhao, Tianqi
    Kong, Ming
    Liang, Tian
    Zhu, Qiang
    Kuang, Kun
    Wu, Fei
    [J]. PROCEEDINGS OF THE 2023 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2023, 2023, : 622 - 626
  • [10] Improving Medical Speech-to-Text Accuracy using Vision-Language Pre-training Models
    Huh, Jaeyoung
    Park, Sangjoon
    Lee, Jeong Eun
    Ye, Jong Chul
    [J]. IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (03) : 1692 - 1703