MII: A Novel Text Classification Model Combining Deep Active Learning with BERT

被引:11
|
作者
Zhang, Anman [1 ]
Li, Bohan [1 ,2 ,3 ]
Wang, Wenhuan [1 ]
Wan, Shuo [1 ]
Chen, Weitong [4 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing 211106, Peoples R China
[2] Minist Ind & Informat Technol, Key Lab Safety Crit Software, Nanjing 211106, Jiangsu, Peoples R China
[3] Collaborat Innovat Ctr Novel Software Technol & I, Nanjing 210046, Peoples R China
[4] Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Qld, Australia
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2020年 / 63卷 / 03期
基金
中国国家自然科学基金;
关键词
Active learning; instance selection; deep neural network; text classification;
D O I
10.32604/cmc.2020.09962
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Active learning has been widely utilized to reduce the labeling cost of supervised learning. By selecting specific instances to train the model, the performance of the model was improved within limited steps. However, rare work paid attention to the effectiveness of active learning on it. In this paper, we proposed a deep active learning model with bidirectional encoder representations from transformers (BERT) for text classification. BERT takes advantage of the self-attention mechanism to integrate contextual information, which is beneficial to accelerate the convergence of training. As for the process of active learning, we design an instance selection strategy based on posterior probabilities Margin, Intra-correlation and Inter-correlation (MII). Selected instances are characterized by small margin, low intra-cohesion and high inter-cohesion. We conduct extensive experiments and analytics with our methods. The effect of learner is compared while the effect of sampling strategy and text classification is assessed from three real datasets. The results show that our method outperforms the baselines in terms of accuracy.
引用
收藏
页码:1499 / 1514
页数:16
相关论文
共 50 条
  • [1] MII: A novel text classification model combining deep active learning with BERT
    Zhang, Anman
    Li, Bohan
    Wang, Wenhuan
    Wan, Shuo
    Chen, Weitong
    [J]. Computers, Materials and Continua, 2020, 63 (03): : 1499 - 1514
  • [2] A text classification network model combining machine learning and deep learning
    Chen, Hao
    Zhang, Haifei
    Yang, Yuwei
    He, Long
    [J]. INTERNATIONAL JOURNAL OF SENSOR NETWORKS, 2024, 44 (03) : 182 - 192
  • [3] Text Classification Model Based on BERT-Capsule with Integrated Deep Learning
    Tian, Yuwei
    Zhang, Zhi
    [J]. PROCEEDINGS OF THE 2021 IEEE 16TH CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA 2021), 2021, : 106 - 111
  • [4] Deep Active Learning for Text Classification
    An, Bang
    Wu, Wenjun
    Han, Huimin
    [J]. PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON VISION, IMAGE AND SIGNAL PROCESSING (ICVISP 2018), 2018,
  • [5] BertGCN: Transductive Text Classification by Combining GCN and BERT
    Lin, Yuxiao
    Meng, Yuxian
    Sun, Xiaofei
    Han, Qinghong
    Kuang, Kun
    Li, Jiwei
    Wu, Fei
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1456 - 1462
  • [6] An ensemble model for idioms and literal text classification using knowledge-enabled BERT in deep learning
    Abarna, S.
    Sheeba, J.I.
    Devaneyan, S. Pradeep
    [J]. Measurement: Sensors, 2022, 24
  • [7] Combining active learning and relevance vector machines for text classification
    Silva, C.
    Ribeiro, B.
    [J]. ICMLA 2007: SIXTH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, PROCEEDINGS, 2007, : 130 - +
  • [8] Deep Active Learning for Text Classification with Diverse Interpretations
    Liu, Qiang
    Zhu, Yanqiao
    Liu, Zhaocheng
    Zhang, Yufeng
    Wu, Shu
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 3263 - 3267
  • [9] A Hybrid Deep Learning Model for Text Classification
    Chen, Xianglong
    Ouyang, Chunping
    Liu, Yongbin
    Luo, Lingyun
    Yang, Xiaohua
    [J]. 2018 14TH INTERNATIONAL CONFERENCE ON SEMANTICS, KNOWLEDGE AND GRIDS (SKG), 2018, : 46 - 52