Robust Loss functions for Learning Multi-Class Classifiers

被引:10
|
作者
Kumar, Himanshu [1 ]
Sastry, P. S. [1 ]
机构
[1] Indian Inst Sci, Dept Elect Engn, Bangalore, Karnataka, India
关键词
Label Noise; Classification; Multi-Class; Loss Functions; NOISE;
D O I
10.1109/SMC.2018.00125
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Robust learning in presence of label noise is an important problem of current interest. Training data often has label noise due to subjective biases of experts, crowd-sourced labelling or other automatic labelling processes. Recently, some sufficient conditions on a loss function are proposed so that risk minimization under such loss functions is provably tolerant to label noise. The standard loss functions such as cross-entropy or mean-squared error, used for learning neural network classifiers, do not satisfy these conditions. It was shown that a loss function based on mean absolute value of error satisfies the conditions and is also empirically seen to be robust to label noise. However, minimizing absolute value of error is a difficult optimization problem. In this paper we propose a new loss function, called robust log loss and show that it satisfies the sufficient conditions for robustness. The resulting optimization problem of minimizing empirical risk is well behaved. Through extensive empirical results we show that, in terms of accuracy and learning rate, the proposed loss function is as good as cross-entropy loss for learning neural network classifiers when there is no label noise and that it is better when the training data has label noise.
引用
收藏
页码:687 / 692
页数:6
相关论文
共 50 条
  • [1] Blind Multi-class Ensemble Learning with Dependent Classifiers
    Traganitis, Panagiotis A.
    Giannakis, Georgios B.
    [J]. 2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2018, : 2025 - 2029
  • [2] On reoptimizing multi-class classifiers
    Chris Bourke
    Kun Deng
    Stephen D. Scott
    Robert E. Schapire
    N. V. Vinodchandran
    [J]. Machine Learning, 2008, 71 : 219 - 242
  • [3] On reoptimizing multi-class classifiers
    Bourke, Chris
    Deng, Kun
    Scott, Stephen D.
    Schapire, Robert E.
    Vinodchandran, N. V.
    [J]. MACHINE LEARNING, 2008, 71 (2-3) : 219 - 242
  • [4] Construction of Multi-class Classifiers by Extreme Learning Machine Based One-class Classifiers
    Gautam, Chandan
    Tiwari, Aruna
    Ravindran, Sriram
    [J]. 2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 2001 - 2007
  • [5] Robust Multi-Class Transductive Learning with Graphs
    Liu, Wei
    Chang, Shih-Fu
    [J]. CVPR: 2009 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-4, 2009, : 381 - 388
  • [6] Using an Hebbian learning rule for multi-class SVM classifiers
    Viéville, T
    Crahay, S
    [J]. JOURNAL OF COMPUTATIONAL NEUROSCIENCE, 2004, 17 (03) : 271 - 287
  • [7] Using an Hebbian Learning Rule for Multi-Class SVM Classifiers
    Thierry Viéville
    Sylvie Crahay
    [J]. Journal of Computational Neuroscience, 2004, 17 : 271 - 287
  • [8] Experimental comparisons of multi-class classifiers
    Institute of Intelligent Computing and Information Technology, Chengdu Normal University, No.99 East Haike Road, Wenjiang District, Chengdu, China
    不详
    [J]. Informatica, 1 (71-85):
  • [9] Experimental Comparisons of Multi-class Classifiers
    Li, Lin
    Li, Lin
    Wu, Yue
    Ye, Mao
    [J]. INFORMATICA-JOURNAL OF COMPUTING AND INFORMATICS, 2015, 39 (01): : 71 - 85
  • [10] Multi-Class Classifiers and Their Underlying Shared Structure
    Vural, Volkan
    Fung, Glenn
    Rosales, Romer
    Dy, Jennifer G.
    [J]. 21ST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI-09), PROCEEDINGS, 2009, : 1267 - 1272