A Curriculum Learning Approach for Multi-Domain Text Classification Using Keyword Weight Ranking

被引:1
|
作者
Yuan, Zilin [1 ]
Li, Yinghui [1 ]
Li, Yangning [1 ]
Zheng, Hai-Tao [1 ,2 ]
He, Yaobin [3 ,4 ]
Liu, Wenqiang [5 ]
Huang, Dongxiao [5 ]
Wu, Bei [5 ]
机构
[1] Tsinghua Univ, Shenzhen Int Grad Sch, Shenzhen 518055, Peoples R China
[2] Pengcheng Lab, Shenzhen 518055, Peoples R China
[3] Smart City Res Inst CETC, Shenzhen 518055, Peoples R China
[4] Natl Ctr Appl Math Shenzhen, Shenzhen 518055, Peoples R China
[5] Tencent Inc, Interact Entertainment Grp, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
multi-domain text classification; curriculum learning; keyword weight ranking;
D O I
10.3390/electronics12143040
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Text classification is a well-established task in NLP, but it has two major limitations. Firstly, text classification is heavily reliant on domain-specific knowledge, meaning that a classifier that is trained on a given corpus may not perform well when presented with text from another domain. Secondly, text classification models require substantial amounts of annotated data for training, and in certain domains, there may be an insufficient quantity of labeled data available. Consequently, it is essential to explore methods for efficiently utilizing text data from various domains to improve the performance of models across a range of domains. One approach for achieving this is through the use of multi-domain text classification models that leverage adversarial training to extract domain-shared features among all domains as well as the specific features of each domain. After observing the varying distinctness of domain-specific features, our paper introduces a curriculum learning approach using a ranking system based on keyword weight to enhance the effectiveness of multi-domain text classification models. The experimental data from Amazon reviews and FDU-MTL datasets show that our method significantly improves the efficacy of multi-domain text classification models adopting adversarial learning and reaching state-of-the-art outcomes on these two datasets.
引用
下载
收藏
页数:14
相关论文
共 50 条
  • [41] Learning to share by masking the non-shared for multi-domain sentiment classification
    Jianhua Yuan
    Yanyan Zhao
    Bing Qin
    International Journal of Machine Learning and Cybernetics, 2022, 13 : 2711 - 2724
  • [42] Multi-Domain Feature Fusion for Emotion Classification Using DEAP Dataset
    Khateeb, Muhammad
    Anwar, Syed Muhammad
    Alnowami, Majdi
    IEEE Access, 2021, 9 : 12134 - 12142
  • [43] Multi-Domain Sentiment Classification with Classifier Combination
    Shou-Shan Li
    Chu-Ren Huang
    Cheng-Qing Zong
    Journal of Computer Science and Technology, 2011, 26 : 25 - 33
  • [44] Multi-Domain Sentiment Classification with Classifier Combination
    李寿山
    黄居仁
    宗成庆
    Journal of Computer Science & Technology, 2011, 26 (01) : 25 - 33
  • [45] Multi-Domain Feature Fusion for Emotion Classification Using DEAP Dataset
    Khateeb, Muhammad
    Anwar, Syed Muhammad
    Alnowami, Majdi
    IEEE ACCESS, 2021, 9 : 12134 - 12142
  • [46] Multi-Domain Sentiment Classification with Classifier Combination
    Li, Shou-Shan
    Huang, Chu-Ren
    Zong, Cheng-Qing
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2011, 26 (01) : 25 - 33
  • [47] Conceptual Approach to Multi-Domain Operations
    Petras, Zdenek
    VOJENSKE ROZHLEDY-CZECH MILITARY REVIEW, 2023, 32 (04): : 66 - 85
  • [48] Ranking-Constrained Learning with Rationales for Text Classification
    Wang, Juanyan
    Sharma, Manali
    Bilgic, Mustafa
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 2034 - 2046
  • [49] Hypotheses Ranking and State Tracking for a Multi-Domain Dialog System using Multiple ASR Alternates
    Khan, Omar Zia
    Robichaud, Jean-Philippe
    Crook, Paul
    Sarikaya, Ruhi
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 2022 - 2026
  • [50] Building a multi-domain comparable corpus using a learning to rank method
    Rahimi, Razieh
    Shakery, Azadeh
    Dadashkarimi, Javid
    Ariannezhad, Mozhdeh
    Dehghani, Mostafa
    Esfahani, Hossein Nasr
    NATURAL LANGUAGE ENGINEERING, 2016, 22 (04) : 627 - 653