Leveraging Inter-rater Agreement for Classification in the Presence of Noisy Labels

被引:6
|
作者
Bucarelli, Maria Sofia [2 ]
Cassanol, Lucas [1 ]
Siciliano, Federico [2 ]
Mantrachl, Amin [1 ]
Silvestri, Fabrizio [2 ,3 ]
机构
[1] Amazon, Buenos Aires, DF, Argentina
[2] Sapienza Univ Rome, Rome, Italy
[3] CNR, ISTI, Pisa, Italy
基金
欧盟地平线“2020”;
关键词
D O I
10.1109/CVPR52729.2023.00335
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In practical settings, classification datasets are obtained through a labelling process that is usually done by humans. Labels can be noisy as they are obtained by aggregating the different individual labels assigned to the same sample by multiple, and possibly disagreeing, annotators. The inter-rater agreement on these datasets can be measured while the underlying noise distribution to which the labels are subject is assumed to be unknown. In this work, we: (i) show how to leverage the inter-annotator statistics to estimate the noise distribution to which labels are subject; (ii) introduce methods that use the estimate of the noise distribution to learn from the noisy dataset; and (iii) establish generalization bounds in the empirical risk minimization framework that depend on the estimated quantities. We conclude the paper by providing experiments that illustrate our findings.
引用
收藏
页码:3439 / 3448
页数:10
相关论文
共 50 条
  • [31] THE AACHEN APHASIA TEST - RELIABILITY AND INTER-RATER AGREEMENT
    WENIGER, D
    WILLMES, K
    HUBER, W
    POECK, K
    NERVENARZT, 1981, 52 (05): : 269 - 277
  • [32] Debugging a Crowdsourced Task with Low Inter-Rater Agreement
    Alonso, Omar
    Marshall, Catherine C.
    Najork, Marc
    PROCEEDINGS OF THE 15TH ACM/IEEE-CS JOINT CONFERENCE ON DIGITAL LIBRARIES (JCDL'15), 2015, : 101 - 110
  • [33] Fragility Fractures of the Pelvis Classification A Multicenter Assessment of the Intra-Rater and Inter-Rater Reliabilities and Percentage of Agreement
    Pieroh, Philipp
    Hoech, Andreas
    Hohmann, Tim
    Gras, Florian
    Maerdian, Sven
    Pflug, Alexander
    Wittenberg, Silvan
    Ihle, Christoph
    Blankenburg, Notker
    Dallacker-Losensky, Kevin
    Schroeder, Tanja
    Herath, Steven C.
    Wagner, Daniel
    Palm, Hans-Georg
    Josten, Christoph
    Stuby, Fabian M.
    JOURNAL OF BONE AND JOINT SURGERY-AMERICAN VOLUME, 2019, 101 (11): : 987 - 994
  • [34] A Systematic Classification and Consistent Reporting Mechanism Yields High Inter-Rater EEG Agreement
    Burgess, Richard
    Wooledge, Gregory
    Nair, Dileep
    NEUROLOGY, 2016, 86
  • [35] DIGITISED BIOPSY REPORTING IN NAFLD/ NASH CLINICAL TRIALS: NAS INTER-RATER AGREEMENT ON DIGITAL SLIDES IS EQUIVALENT TO INTER-RATER AGREEMENT ON GLASS SLIDES
    Langford, Caitlin
    Kendall, Tim
    Fryer, Eve
    Telford, Alison
    Kazimianec, Arina
    Fleming, Kenneth A.
    HEPATOLOGY, 2022, 76 : S830 - S831
  • [36] Inter-rater agreement on decision making in glaucoma clinical encounters
    Oke, Isdin
    Eliassi-Rad, Babak
    Desai, Manishi
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2020, 61 (07)
  • [37] The Problem of Limited Inter-rater Agreement in Modelling Music Similarity
    Flexer, Arthur
    Grill, Thomas
    JOURNAL OF NEW MUSIC RESEARCH, 2016, 45 (03) : 239 - 251
  • [38] Inter-rater agreement on proximal humerus fracture treatment in Korea
    Jo, Chris Hyunchul
    CLINICS IN SHOULDER AND ELBOW, 2022, 25 (01): : 3 - 4
  • [39] Factors affecting inter-rater agreement in human classification of eye movements: a comparison of three datasets
    Lee Friedman
    Vladyslav Prokopenko
    Shagen Djanian
    Dmytro Katrychuk
    Oleg V. Komogortsev
    Behavior Research Methods, 2023, 55 : 417 - 427
  • [40] Factors affecting inter-rater agreement in human classification of eye movements: a comparison of three datasets
    Friedman, Lee
    Prokopenko, Vladyslav
    Djanian, Shagen
    Katrychuk, Dmytro
    Komogortsev, Oleg, V
    BEHAVIOR RESEARCH METHODS, 2023, 55 (01) : 417 - 427