TopicBERT: A Topic-Enhanced Neural Language Model Fine-Tuned for Sentiment Classification

被引:18
|
作者
Zhou, Yuxiang [1 ]
Liao, Lejian [1 ]
Gao, Yang [1 ]
Wang, Rui [2 ]
Huang, Heyan [1 ]
机构
[1] Beijing Inst Technol, Fac Comp Sci, Beijing 100081, Peoples R China
[2] Nanjing Univ Posts & Telecommun, Fac Comp Sci, Nanjing 210023, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Bit error rate; Semantics; Predictive models; Training; Context modeling; Social networking (online); Bidirectional encoder representations from transformers (BERT); pretrained neural language model; sentiment classification; topic-enhanced neural network;
D O I
10.1109/TNNLS.2021.3094987
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sentiment classification is a form of data analytics where people's feelings and attitudes toward a topic are mined from data. This tantalizing power to ``predict the zeitgeist'' means that sentiment classification has long attracted interest, but with mixed results. However, the recent development of the BERT framework and its pretrained neural language models is seeing new-found success for sentiment classification. BERT models are trained to capture word-level information via mask language modeling and sentence-level contexts via next sentence prediction tasks. Out of the box, they are adequate models for some natural language processing tasks. However, most models are further fine-tuned with domain-specific information to increase accuracy and usefulness. Motivated by the idea that a further fine-tuning step would improve the performance for downstream sentiment classification tasks, we developed TopicBERT--a BERT model fine-tuned to recognize topics at the corpus level in addition to the word and sentence levels. TopicBERT comprises two variants: TopicBERT-ATP (aspect topic prediction), which captures topic information via an auxiliary training task, and TopicBERT-TA, where topic representation is directly injected into a topic augmentation layer for sentiment classification. With TopicBERT-ATP, the topics are predetermined by an LDA mechanism and collapsed Gibbs sampling. With TopicBERT-TA, the topics can change dynamically during the training. Experimental results show that both approaches deliver the state-of-the-art performance in two different domains with SemEval 2014 Task 4. However, in a test of methods, direct augmentation outperforms further training. Comprehensive analyses in the form of ablation, parameter, and complexity studies accompany the results.
引用
下载
收藏
页码:380 / 393
页数:14
相关论文
共 50 条
  • [21] Robotic Arm Control By Fine-Tuned Convolutional Neural Network Model
    Bayraktar, Ertugrul
    Yigit, Cihat Bora
    Boyraz, Pinar
    2017 25TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2017,
  • [22] Automated classification of brain MRI reports using fine-tuned large language models
    Kanzawa, Jun
    Yasaka, Koichiro
    Fujita, Nana
    Fujiwara, Shin
    Abe, Osamu
    NEURORADIOLOGY, 2024, : 2177 - 2183
  • [23] Taiyi: a bilingual fine-tuned large language model for diverse biomedical tasks
    Luo, Ling
    Ning, Jinzhong
    Zhao, Yingwen
    Wang, Zhijun
    Ding, Zeyuan
    Chen, Peng
    Fu, Weiru
    Han, Qinyu
    Xu, Guangtao
    Qiu, Yunzhi
    Pan, Dinghao
    Li, Jiru
    Li, Hao
    Feng, Wenduo
    Tu, Senbo
    Liu, Yuqi
    Yang, Zhihao
    Wang, Jian
    Sun, Yuanyuan
    Lin, Hongfei
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 1865 - 1874
  • [24] Dermoscopic Image Classification Method Using an Ensemble of Fine-Tuned Convolutional Neural Networks
    Shen, Xin
    Wei, Lisheng
    Tang, Shaoyu
    SENSORS, 2022, 22 (11)
  • [25] Ensemble of fine-tuned convolutional neural networks for urine sediment microscopic image classification
    Liu, Wenqian
    Li, Weihong
    Gong, Weiguo
    IET COMPUTER VISION, 2020, 14 (01) : 18 - 25
  • [26] A Fine-Tuned Convolution Neural Network Based Approach For Phenotype Classification Of Zebrafish Embryo
    Tyagi, Gaurav
    Patel, Nilesh
    Sethi, Ishwar
    KNOWLEDGE-BASED AND INTELLIGENT INFORMATION & ENGINEERING SYSTEMS (KES-2018), 2018, 126 : 1138 - 1144
  • [27] Blending Ensemble of Fine-Tuned Convolutional Neural Networks Applied to Mammary Image Classification
    Zhang, Jingyi
    Pan, Shuwan
    Hong, Huichao
    Kong, Lingke
    JOURNAL OF MEDICAL IMAGING AND HEALTH INFORMATICS, 2019, 9 (06) : 1160 - 1166
  • [28] BERT’s sentiment score for portfolio optimization: a fine-tuned view in Black and Litterman model
    Francesco Colasanto
    Luca Grilli
    Domenico Santoro
    Giovanni Villani
    Neural Computing and Applications, 2022, 34 : 17507 - 17521
  • [29] BERT's sentiment score for portfolio optimization: a fine-tuned view in Black and Litterman model
    Colasanto, Francesco
    Grilli, Luca
    Santoro, Domenico
    Villani, Giovanni
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (20): : 17507 - 17521
  • [30] Understanding language-elicited EEG data by predicting it from a fine-tuned language model
    Schwartz, Dan
    Mitchell, Tom
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 43 - 57