Pre-Trained Language Model-Based Deep Learning for Sentiment Classification of Vietnamese Feedback

被引:0
|
作者
Loc, Cu Vinh [1 ]
Viet, Truong Xuan [1 ]
Viet, Tran Hoang [1 ]
Thao, Le Hoang [1 ]
Viet, Nguyen Hoang [1 ]
机构
[1] Can Tho Univ, Software Ctr, Can Tho city, Vietnam
关键词
Sentiment analysis; PhoBERT; deep learning; text classification; Vietnamese feedback;
D O I
10.1142/S1469026823500165
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, with the strong and outstanding development of the Internet, the need to refer to the feedback of previous customers when shopping online is increasing. Therefore, websites are developed to allow users to share experiences, reviews, comments and feedback about the services and products of businesses and organizations. The organizations also collect user feedback about their products and services to give better directions. However, with a large amount of user feedback about certain services and products, it is difficult for users, businesses, and organizations to pay attention to them all. Thus, an automatic system is necessary to analyze the sentiment of a customer feedback. Recently, the well-known pre-trained language models for Vietnamese (PhoBERT) achieved high performance in comparison with other approaches. However, this method may not focus on the local information in the text like phrases or fragments. In this paper, we propose a Convolutional Neural Network (CNN) model based on PhoBERT for sentiment classification. The output of contextualized embeddings of the PhoBERT's last four layers is fed into the CNN. This makes the network capable of obtaining more local information from the sentiment. Besides, the PhoBERT output is also given to the transformer encoder layers in order to employ the self-attention technique, and this also makes the model more focused on the important information of the sentiment segments. The experimental results demonstrate that the proposed approach gives competitive performance compared to the existing studies on three public datasets with the opinions of Vietnamese people.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] A Study of Vietnamese Sentiment Classification with Ensemble Pre-trained Language Models
    Thin, Dang Van
    Hao, Duong Ngoc
    Nguyen, Ngan Luu-Thuy
    VIETNAM JOURNAL OF COMPUTER SCIENCE, 2024, 11 (01) : 137 - 165
  • [2] Chinese-Korean Weibo Sentiment Classification Based on Pre-trained Language Model and Transfer Learning
    Wang, Hengxuan
    Zhang, Zhenguo
    Cui, Xu
    Cui, Rongyi
    2022 IEEE 2ND INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND ARTIFICIAL INTELLIGENCE (CCAI 2022), 2022, : 49 - 54
  • [3] ViDeBERTa: A powerful pre-trained language model for Vietnamese
    Tran, Cong Dao
    Pham, Nhut Huy
    Nguyen, Anh
    Hy, Truong Son
    Vu, Tu
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1071 - 1078
  • [4] Neural Transfer Learning For Vietnamese Sentiment Analysis Using Pre-trained Contextual Language Models
    An Pha Le
    Tran Vu Pham
    Thanh-Van Le
    Huynh, Duy, V
    2021 IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLIED NETWORK TECHNOLOGIES (ICMLANT II), 2021, : 84 - 88
  • [5] A survey of text classification based on pre-trained language model
    Wu, Yujia
    Wan, Jun
    NEUROCOMPUTING, 2025, 616
  • [6] Chinese Fine-Grained Sentiment Classification Based on Pre-trained Language Model and Attention Mechanism
    Zhou, Faguo
    Zhang, Jing
    Song, Yanan
    SMART COMPUTING AND COMMUNICATION, 2022, 13202 : 37 - 47
  • [7] PILOT: a pre-trained model-based continual learning toolbox
    Sun, Hai-Long
    Zhou, Da-Wei
    Zhan, De-Chuan
    Ye, Han-Jia
    SCIENCE CHINA-INFORMATION SCIENCES, 2025, 68 (04)
  • [8] PILOT: a pre-trained model-based continual learning toolbox
    HaiLong SUN
    DaWei ZHOU
    DeChuan ZHAN
    HanJia YE
    Science China(Information Sciences), 2025, 68 (04) : 383 - 384
  • [9] Leveraging Pre-trained Language Model for Speech Sentiment Analysis
    Shon, Suwon
    Brusco, Pablo
    Pan, Jing
    Han, Kyu J.
    Watanabe, Shinji
    INTERSPEECH 2021, 2021, : 3420 - 3424
  • [10] AraXLNet: pre-trained language model for sentiment analysis of Arabic
    Alduailej, Alhanouf
    Alothaim, Abdulrahman
    JOURNAL OF BIG DATA, 2022, 9 (01)