Leveraging Inter-rater Agreement for Classification in the Presence of Noisy Labels

被引:6
|
作者
Bucarelli, Maria Sofia [2 ]
Cassanol, Lucas [1 ]
Siciliano, Federico [2 ]
Mantrachl, Amin [1 ]
Silvestri, Fabrizio [2 ,3 ]
机构
[1] Amazon, Buenos Aires, DF, Argentina
[2] Sapienza Univ Rome, Rome, Italy
[3] CNR, ISTI, Pisa, Italy
基金
欧盟地平线“2020”;
关键词
D O I
10.1109/CVPR52729.2023.00335
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In practical settings, classification datasets are obtained through a labelling process that is usually done by humans. Labels can be noisy as they are obtained by aggregating the different individual labels assigned to the same sample by multiple, and possibly disagreeing, annotators. The inter-rater agreement on these datasets can be measured while the underlying noise distribution to which the labels are subject is assumed to be unknown. In this work, we: (i) show how to leverage the inter-annotator statistics to estimate the noise distribution to which labels are subject; (ii) introduce methods that use the estimate of the noise distribution to learn from the noisy dataset; and (iii) establish generalization bounds in the empirical risk minimization framework that depend on the estimated quantities. We conclude the paper by providing experiments that illustrate our findings.
引用
收藏
页码:3439 / 3448
页数:10
相关论文
共 50 条
  • [41] Gauging the Quality of Relevance Assessments using Inter-Rater Agreement
    Damessie, Tadele T.
    Nghiem, Thao P.
    Scholer, Falk
    Culpepper, J. Shane
    SIGIR'17: PROCEEDINGS OF THE 40TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2017, : 1089 - 1092
  • [42] Beyond Kappa: Estimating Inter-Rater Agreement with Nominal Classifications
    Bendermacher, Nol
    Souren, Pierre
    JOURNAL OF MODERN APPLIED STATISTICAL METHODS, 2009, 8 (01) : 110 - 121
  • [43] Inter-rater Agreement for the Diagnosis of Stroke Versus Stroke Mimic
    Liberman, Ava L.
    Rostanski, Sara K.
    Ruff, Ilana M.
    Meyer, Ashley N. D.
    Maas, Matthew B.
    Prabhakaran, Shyam
    NEUROLOGIST, 2018, 23 (04) : 118 - 121
  • [44] Inter-rater agreement in the diagnosis of mucositis and peri-implantitis
    Merli, Mauro
    Bernardelli, Francesco
    Giulianelli, Erica
    Toselli, Ivano
    Moscatelli, Marco
    Pagliaro, Umberto
    Nieri, Michele
    JOURNAL OF CLINICAL PERIODONTOLOGY, 2014, 41 (09) : 927 - 933
  • [45] Inter-rater agreement of the triage system RETTS-HEV
    Nissen, Louise
    Kirkegaard, Hans
    Perez, Noel
    Horlyk, Ulf
    Larsen, Louise P.
    EUROPEAN JOURNAL OF EMERGENCY MEDICINE, 2014, 21 (01) : 37 - 41
  • [46] Inter-rater agreement on assessment of outcome within a trauma registry
    Ekegren, C. L.
    Hart, M. J.
    Brown, A.
    Gabbe, B. J.
    INJURY-INTERNATIONAL JOURNAL OF THE CARE OF THE INJURED, 2016, 47 (01): : 130 - 134
  • [47] Inter-rater agreement in radiographic canine hip dysplasia evaluation
    Geissbuehler, U.
    Drazovic, S.
    Lang, J.
    Howard, J.
    VETERINARY RECORD, 2017, 180 (14) : 357 - +
  • [48] Inter-rater agreement in the evaluation of abdominal radiographs for necrotizing enterocolitis
    El-Kady, Sherif
    Petel, Dara
    Baird, Robert
    JOURNAL OF PEDIATRIC SURGERY, 2014, 49 (05) : 733 - 735
  • [49] Inter-rater Agreement in Scoring Home Cardiorespiratory Tracings • 1790
    Michael J Corwin
    G Lister
    J Silvestri
    M Peucker
    L Brooks
    S L Davidson Ward
    C E Hunt
    M R Neuman
    D H Crowell
    T Colton
    Pediatric Research, 1997, 41 (Suppl 4) : 301 - 301
  • [50] Risser sign inter-rater and intra-rater agreement: is the Risser sign reliable?
    Jennifer Reem
    Joseph Carney
    Mark Stanley
    Jeffrey Cassidy
    Skeletal Radiology, 2009, 38 : 371 - 375