Leveraging Inter-rater Agreement for Classification in the Presence of Noisy Labels

被引:6
|
作者
Bucarelli, Maria Sofia [2 ]
Cassanol, Lucas [1 ]
Siciliano, Federico [2 ]
Mantrachl, Amin [1 ]
Silvestri, Fabrizio [2 ,3 ]
机构
[1] Amazon, Buenos Aires, DF, Argentina
[2] Sapienza Univ Rome, Rome, Italy
[3] CNR, ISTI, Pisa, Italy
基金
欧盟地平线“2020”;
关键词
D O I
10.1109/CVPR52729.2023.00335
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In practical settings, classification datasets are obtained through a labelling process that is usually done by humans. Labels can be noisy as they are obtained by aggregating the different individual labels assigned to the same sample by multiple, and possibly disagreeing, annotators. The inter-rater agreement on these datasets can be measured while the underlying noise distribution to which the labels are subject is assumed to be unknown. In this work, we: (i) show how to leverage the inter-annotator statistics to estimate the noise distribution to which labels are subject; (ii) introduce methods that use the estimate of the noise distribution to learn from the noisy dataset; and (iii) establish generalization bounds in the empirical risk minimization framework that depend on the estimated quantities. We conclude the paper by providing experiments that illustrate our findings.
引用
收藏
页码:3439 / 3448
页数:10
相关论文
共 50 条
  • [1] Leveraging Inter-rater Agreement for Audio-Visual Emotion Recognition
    Kim, Yelin
    Provost, Emily Mower
    2015 INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION (ACII), 2015, : 553 - 559
  • [2] Inter-rater agreement on the protocol for care and risk classification in obstetrics
    Lima, Debora Rodrigues
    Magalhaes, Fernanda Jorge
    Felisbino-Mendes, Mariana Santos
    Bueno, Mariana
    Duarte, Elysangela Dittz
    ACTA PAULISTA DE ENFERMAGEM, 2025, 38
  • [3] Inter-rater agreement on the protocol for care and risk classification in obstetrics
    Lima, Debora Rodrigues
    Magalhaes, Fernanda Jorge
    Felisbino-Mendes, Mariana Santos
    Bueno, Mariana
    Duarte, Elysangela Dittz
    ACTA PAULISTA DE ENFERMAGEM, 2025, 38
  • [4] Comparison between Inter-rater Reliability and Inter-rater Agreement in Performance Assessment
    Liao, Shih Chieh
    Hunt, Elizabeth A.
    Chen, Walter
    ANNALS ACADEMY OF MEDICINE SINGAPORE, 2010, 39 (08) : 613 - 618
  • [5] Inter-rater agreement: a methodological issue
    Shahsavari, Meisam
    Shahsavari, Soodeh
    JOURNAL OF NEUROSURGERY, 2019, 131 (02) : 651 - 651
  • [6] Bayesian analysis for inter-rater agreement
    Broemeling, LD
    COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2001, 30 (03) : 437 - 446
  • [7] Computing inter-rater reliability and its variance in the presence of high agreement
    Gwet, Kilem Li
    BRITISH JOURNAL OF MATHEMATICAL & STATISTICAL PSYCHOLOGY, 2008, 61 : 29 - 48
  • [8] The Tulip classification of perinatal mortality: introduction and multidisciplinary inter-rater agreement
    Korteweg, FJ
    Gordijn, SJ
    Timmer, A
    Erwich, JJHM
    Bergman, KA
    Bouman, K
    Ravise, JM
    Heringa, MP
    Holm, JP
    BJOG-AN INTERNATIONAL JOURNAL OF OBSTETRICS AND GYNAECOLOGY, 2006, 113 (04) : 393 - 401
  • [9] Double Entropy Inter-Rater Agreement Indices
    Olenko, Andriy
    Tsyganok, Vitaliy
    APPLIED PSYCHOLOGICAL MEASUREMENT, 2016, 40 (01) : 37 - 55
  • [10] Inter-rater Agreement for Social Computing Studies
    Salminen, Joni O.
    Al-Merekhi, Hind A.
    Dey, Partha
    Jansen, Bernard J.
    2018 FIFTH INTERNATIONAL CONFERENCE ON SOCIAL NETWORKS ANALYSIS, MANAGEMENT AND SECURITY (SNAMS), 2018, : 80 - 87