A Neural Expectation-Maximization Framework for Noisy Multi-Label Text Classification

被引:2
|
作者
Chen, Junfan [1 ,2 ]
Zhang, Richong [1 ,3 ]
Xu, Jie [4 ]
Hu, Chunming [1 ,3 ,5 ]
Mao, Yongyi [6 ]
机构
[1] Beihang Univ, Sch Comp Sci & Engn, SKLSDE, Beijing 100191, Peoples R China
[2] Beihang Univ, Sch Software, Beijing 100191, Peoples R China
[3] Zhongguancun Lab, Beijing 100190, Peoples R China
[4] Univ Leeds, Leeds LS2 9JT, England
[5] Beihang Univ, Sch Software, Beijing 100191, Peoples R China
[6] Univ Ottawa, Ottawa, ON K1N 6N5, Canada
基金
国家重点研发计划;
关键词
Multi-label text classification; noise label; expectation maximization; neural networks;
D O I
10.1109/TKDE.2022.3223067
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-label text classification (MLTC) has a wide range of real-world applications. Neural networks recently promoted the performance of MLTC models. Training these neural-network models relies on sufficient accurately labelled data. However, manually annotating large-scale multi-label text classification datasets is expensive and impractical for many applications. Weak supervision techniques have thus been developed to reduce the cost of annotating text corpus. However, these techniques introduce noisy labels into the training data and may degrade the model performance. This paper aims to deal with such noise-label problems in MLTC in both single-instance and multi-instance settings. We build a novel Neural Expectation-Maximization Framework (nEM) that combines neural networks with probabilistic modelling. The nEM framework produces text representations using neural-network text encoders and is optimized with the Expectation-Maximization algorithm. It naturally considers the noisy labels during learning by iteratively updating the model parameters and estimating the distribution of the ground-truth labels. We evaluate our nEM framework in multi-instance noisy MLTC on a benchmark relation extraction dataset constructed by distant supervision and in single-instance noisy MLTC on synthetic noisy datasets constructed by keywords supervision and label flipping. The experimental results demonstrate that nEM significantly improves upon baseline models in both single-instance and multi-instance noisy MLTC tasks. The experiment analysis suggests that our nEM framework efficiently reduces the noisy labels in MLTC datasets and significantly improves model performance.
引用
收藏
页码:10992 / 11003
页数:12
相关论文
共 50 条
  • [41] Deep Learning for Extreme Multi-label Text Classification
    Liu, Jingzhou
    Chang, Wei-Cheng
    Wu, Yuexin
    Yang, Yiming
    [J]. SIGIR'17: PROCEEDINGS OF THE 40TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2017, : 115 - 124
  • [42] A NEW INPUT REPRESENTATION FOR MULTI-LABEL TEXT CLASSIFICATION
    Alfaro, Rodrigo
    Allende, Hector
    [J]. 2011 INTERNATIONAL CONFERENCE ON INSTRUMENTATION, MEASUREMENT, CIRCUITS AND SYSTEMS (ICIMCS 2011), VOL 3: COMPUTER-AIDED DESIGN, MANUFACTURING AND MANAGEMENT, 2011, : 207 - 210
  • [43] Hierarchical Multi-Label Classification of Social Text Streams
    Ren, Zhaochun
    Peetz, Maria-Hendrike
    Liang, Shangsong
    van Dolen, Willemijn
    de Rijke, Maarten
    [J]. SIGIR'14: PROCEEDINGS OF THE 37TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2014, : 213 - 222
  • [44] An approach for multi-label classification by directed acyclic graph with label correlation maximization
    Lee, Jaedong
    Kim, Heera
    Kim, Noo-ri
    Lee, Jee-Hyong
    [J]. INFORMATION SCIENCES, 2016, 351 : 101 - 114
  • [45] Multi-label dataless text classification with topic modeling
    Zha, Daochen
    Li, Chenliang
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2019, 61 (01) : 137 - 160
  • [46] Hierarchical Transfer Learning for Multi-label Text Classification
    Banerjee, Siddhartha
    Akkaya, Cem
    Perez-Sorrosal, Francisco
    Tsioutsiouliklis, Kostas
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 6295 - 6300
  • [47] Effective Multi-Label Active Learning for Text Classification
    Yang, Bishan
    Sun, Jian-Tao
    Wang, Tengjiao
    Chen, Zheng
    [J]. KDD-09: 15TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2009, : 917 - 925
  • [48] A novel reasoning mechanism for multi-label text classification
    Wang, Ran
    Ridley, Robert
    Su, Xi'ao
    Qu, Weiguang
    Dai, Xinyu
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2021, 58 (02)
  • [49] Academic Resource Text Hierarchical Multi-Label Classification
    Wang, Yue
    Li, Yawen
    Li, Ang
    [J]. Computer Engineering and Applications, 2023, 59 (13): : 92 - 98
  • [50] A Combined Approach for Multi-Label Text Data Classification
    Strimaitis, Rokas
    Stefanovic, Pavel
    Ramanauskaite, Simona
    Slotkiene, Asta
    [J]. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022