Latent Class-Conditional Noise Model

被引:5
|
作者
Yao, Jiangchao [1 ,2 ]
Han, Bo [3 ]
Zhou, Zhihan [1 ,2 ]
Zhang, Ya [1 ,2 ]
Tsang, Ivor W. [4 ]
机构
[1] Shanghai Jiao Tong Univ, Cooperat Medianet Innovat Ctr, Shanghai 200240, Peoples R China
[2] Shanghai AI Lab, Shanghai 200030, Peoples R China
[3] Hong Kong Baptist Univ, Dept Comp Sci, Kowloon, Hong Kong, Peoples R China
[4] A STAR Ctr Frontier AI Res, Singapore 138632, Singapore
关键词
Noise measurement; Training; Optimization; Deep learning; Bayes methods; Robustness; Computational modeling; Bayesian modeling; deep learning; noisy supervision; semi-supervised learning; NETWORKS;
D O I
10.1109/TPAMI.2023.3247629
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning with noisy labels has become imperative in the Big Data era, which saves expensive human labors on accurate annotations. Previous noise-transition-based methods have achieved theoretically-grounded performance under the Class-Conditional Noisemodel (CCN). However, these approaches builds upon an ideal but impractical anchor set available to pre-estimate the noise transition. Even though subsequent works adapt the estimation as a neural layer, the ill-posed stochastic learning of its parameters in back-propagation easily falls into undesired local minimums. We solve this problem by introducing a Latent Class-Conditional Noise model (LCCN) to parameterize the noise transition under a Bayesian framework. By projecting the noise transition into the Dirichlet space, the learning is constrained on a simplex characterized by the complete dataset, instead of some ad-hoc parametric space wrapped by the neural layer. We then deduce a dynamic label regression method for LCCN, whose Gibbs sampler allows us efficiently infer the latent true labels to train the classifier and to model the noise. Our approach safeguards the stable update of the noise transition, which avoids previous arbitrarily tuning from a mini-batch of samples. We further generalize LCCN to different counterparts compatible with open-set noisy labels, semi-supervised learning as well as cross-model training. A range of experiments demonstrate the advantages of LCCN and its variants over the current state-of-the-art methods. The code is available at here.
引用
收藏
页码:9964 / 9980
页数:17
相关论文
共 50 条
  • [21] Advocacy Learning: Learning through Competition and Class-Conditional Representations
    Fox, Ian
    Wiens, Jenna
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2315 - 2321
  • [22] Multinomial classification with class-conditional overlapping sparse feature groups
    Li, Xiangrui
    Zhu, Dongxiao
    Dong, Ming
    PATTERN RECOGNITION LETTERS, 2018, 101 : 37 - 43
  • [23] Class-conditional Importance Weighting for Deep Learning with Noisy Labels
    Nagarajan, Bhalaji
    Marques, Ricardo
    Mejia, Marcos
    Radeva, Petia
    PROCEEDINGS OF THE 17TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS (VISAPP), VOL 5, 2022, : 679 - 686
  • [24] Conditional anomaly detection based on a latent class model
    Ohkubo, Masato
    Nagata, Yasushi
    TOTAL QUALITY MANAGEMENT & BUSINESS EXCELLENCE, 2019, 30 : S227 - S239
  • [25] Naive and Robust: Class-Conditional Independence in Human Classification Learning
    Jarecki, Jana B.
    Meder, Bjoern
    Nelson, Jonathan D.
    COGNITIVE SCIENCE, 2018, 42 (01) : 4 - 42
  • [26] Selection of Class-conditional Filters for Semantic Shifted OOD Detection
    Yu, Yeonguk
    Shin, Sungho
    Kim, Jongwon
    Lee, Kyoobin
    2021 18TH INTERNATIONAL CONFERENCE ON UBIQUITOUS ROBOTS (UR), 2021, : 43 - 47
  • [27] Robust Class-Conditional Distribution Alignment for Partial Domain Adaptation
    Choudhuri, Sandipan
    Sen, Arunabha
    FIFTY-SEVENTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, IEEECONF, 2023, : 530 - 536
  • [28] Unsupervised Keypoint Learning for Guiding Class-Conditional Video Prediction
    Kim, Yunji
    Nam, Seonghyeon
    Cho, In
    Kim, Seon Joo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [29] Quantifying quality of class-conditional generative models in time series domain
    Koochali, Alireza
    Walch, Maria
    Thota, Sankrutyayan
    Schichtel, Peter
    Dengel, Andreas
    Ahmed, Sheraz
    APPLIED INTELLIGENCE, 2023, 53 (20) : 24530 - 24550
  • [30] Multiclass object recognition using class-conditional independent component analysis
    Bressan, M
    Guillamet, D
    Vitrià, J
    CYBERNETICS AND SYSTEMS, 2004, 35 (01) : 35 - 61