Deep Transfer Active Learning Method Combining Source Domain Difference and Target Domain Uncertainty

被引:0
|
作者
Liu D. [1 ]
Cao Y. [1 ]
Su C. [1 ]
Zhang L. [1 ]
机构
[1] School of Big Data and Computer Science, Guizhou Normal University, Guiyang
关键词
Deep Active Learning; Deep Transfer Learning; Information Extraction Ratio; Source Domain Difference; Target Domain Uncertainty;
D O I
10.16451/j.cnki.issn1003-6059.202110003
中图分类号
学科分类号
摘要
Training deep neural network models comes with a heavy labeling cost. To reduce the cost, a deep transfer active learning method combining source domain and target domain is proposed. With the initial model transferred from source task, the current task samples with larger contribution to the model performance improvement are labeled by using a dynamical weighting combination of source domain difference and target domain uncertainty. Information extraction ratio(IER) is concretely defined in the specific case. An IER-based batch training strategy and a T&N batch training strategy are proposed to deal with model training process. The proposed method is tested on two cross-dataset transfer learning experiments. The results show that the transfer active learning method achieves good performance and reduces the cost of annotation effectively and the proposed strategies optimize the distribution of computing resources during the active learning process. Thus, the model learns more times from samples in the early phases and less times in the later and end phases. © 2021, Science Press. All right reserved.
引用
收藏
页码:898 / 908
页数:10
相关论文
共 20 条
  • [1] LECUN Y, BENGIO Y, HINTON G., Deep Learning, Nature, 521, 7553, pp. 436-444, (2015)
  • [2] WEI Q D, SHAO F J, LIU J., Research Summary of Convolution Neural Network in Image Recognition, Proc of the International Conference on Data Processing and Applications, pp. 39-44, (2018)
  • [3] CHEN H S, LIU X R, YIN D W, Et al., A Survey on Dialogue Systems: Recent Advances and New Frontiers, ACM SIGKDD Explorations Newsletter, 19, 2, pp. 25-35, (2017)
  • [4] TAN C Q, SUN F C, KONG T, Et al., A Survey on Deep Transfer Learning, Proc of the International Conference on Artificial Neural Networks, pp. 270-279, (2018)
  • [5] HUANG S J, ZHAO J W, LIU Z Y., Cost-Effective Training of Deep CNNs with Active Model Adaptation, Proc of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mi-ning, pp. 1580-1588, (2018)
  • [6] OQUAB M, BOTTOU L, LAPTEV I, Et al., Learning and Transferring Mid-level Image Representations Using Convolutional Neural Networks, Proc of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1717-1724, (2014)
  • [7] DENG J, DONG W, SOCHER R, Et al., ImageNet: A Large-Scale Hierarchical Image Database, Proc of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 248-255, (2009)
  • [8] TZENG E, HOFFMAN J, ZHANG N, Et al., Deep Domain Confusion: Maximizing for Domain Invariance
  • [9] BORGWARDT K M, GRETTON A, RASCH M J, Et al., Integrating Structured Biological Data by Kernel Maximum Mean Discrepancy, Bioinformatics, 22, 14, pp. e49-e57, (2006)
  • [10] LONG M S, CAO Y, WANG J M, Et al., Learning Transferable Features with Deep Adaptation Networks, Proc of the 32nd International Conference on Machine Learning, pp. 97-105, (2015)