Learning to share by masking the non-shared for multi-domain sentiment classification

被引:8
|
作者
Yuan, Jianhua [1 ]
Zhao, Yanyan [1 ]
Qin, Bing [1 ,2 ]
机构
[1] Harbin Inst Technol, Fac Comp, Harbin 150001, Peoples R China
[2] Pengcheng Lab, Shenzhen 518066, Peoples R China
基金
中国国家自然科学基金;
关键词
Natural language processing; Sentiment analysis; Cross domain; Masking;
D O I
10.1007/s13042-022-01556-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-domain sentiment classification deals with the scenario where labeled data exists for multiple domains but is insufficient for training effective sentiment classifiers that work across domains. Thus, fully exploiting sentiment knowledge shared across domains is crucial for real-world applications. While many existing works try to extract domain-invariant features in high-dimensional space, such models fail to explicitly distinguish between shared and private features at the text level, which to some extent lacks interpretability. Based on the assumption that removing domain-related tokens from texts would help improve their domain invariance, we instead first transform original sentences to be domain-agnostic. To this end, we propose the BERTMasker model which explicitly masks domain-related words from texts, learns domain-invariant sentiment features from these domain-agnostic texts and uses those masked words to form domain-aware sentence representations. Empirical experiments on the benchmark multiple domain sentiment classification datasets demonstrate the effectiveness of our proposed model, which improves the accuracy on multi-domain and cross-domain settings by 1.91% and 3.31% respectively. Further analysis on masking proves that removing those domain-related and sentiment irrelevant tokens decreases texts' domain separability, resulting in the performance degradation of a BERT-based domain classifier by over 12%.
引用
下载
收藏
页码:2711 / 2724
页数:14
相关论文
共 50 条
  • [1] Learning to share by masking the non-shared for multi-domain sentiment classification
    Jianhua Yuan
    Yanyan Zhao
    Bing Qin
    International Journal of Machine Learning and Cybernetics, 2022, 13 : 2711 - 2724
  • [2] Collaborative Multi-Domain Sentiment Classification
    Wu, Fangzhao
    Huang, Yongfeng
    2015 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2015, : 459 - 468
  • [3] Domain attention model for multi-domain sentiment classification
    Yuan, Zhigang
    Wu, Sixing
    Wu, Fangzhao
    Liu, Junxin
    Huang, Yongfeng
    KNOWLEDGE-BASED SYSTEMS, 2018, 155 : 1 - 10
  • [4] Multi-Domain Sentiment Classification with Classifier Combination
    Shou-Shan Li
    Chu-Ren Huang
    Cheng-Qing Zong
    Journal of Computer Science and Technology, 2011, 26 : 25 - 33
  • [5] Multi-Domain Sentiment Classification with Classifier Combination
    李寿山
    黄居仁
    宗成庆
    Journal of Computer Science & Technology, 2011, 26 (01) : 25 - 33
  • [6] Multi-Domain Sentiment Classification with Classifier Combination
    Li, Shou-Shan
    Huang, Chu-Ren
    Zong, Cheng-Qing
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2011, 26 (01) : 25 - 33
  • [7] A novel sentiment aware dictionary for multi-domain sentiment classification
    Jha, Vandana
    Savitha, R.
    Shenoy, P. Deepa
    Venugopal, K. R.
    Sangaiah, Arun Kumar
    COMPUTERS & ELECTRICAL ENGINEERING, 2018, 69 : 585 - 597
  • [8] REFORMIST: Hierarchical Attention Networks for Multi-Domain Sentiment Classification with Active Learning
    Katsarou, Katerina
    Douss, Nabil
    Stefanidis, Kostas
    38TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2023, 2023, : 919 - 928
  • [9] Are SentiWordNet Scores Suited for Multi-Domain Sentiment Classification?
    Denecke, Kerstin
    2009 FOURTH INTERNATIONAL CONFERENCE ON DIGITAL INFORMATION MANAGEMENT, 2009, : 247 - 252
  • [10] A Collaboration Multi-Domain Sentiment Classification on Specific Domain and Global Features
    He, Junping
    Teng, Shaohua
    Fei, Lunke
    Fang, Xiaozhao
    Zhang, Wei
    PROCEEDINGS OF THE 2021 IEEE 24TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN (CSCWD), 2021, : 323 - 328