Firebolt: Weak Supervision Under Weaker Assumptions

被引:0
|
作者
Kuang, Zhaobin [1 ]
Arachie, Chidubem [1 ,2 ]
Liang, Bangyong [1 ]
Narayana, Pradyumna [1 ]
DeSalvo, Giulia [1 ]
Quinn, Michael [1 ]
Huang, Bert [3 ]
Downs, Geoffrey [1 ]
Yang, Yang [1 ]
机构
[1] Google, Mountain View, CA 94043 USA
[2] Virginia Tech, Blacksburg, VA USA
[3] Tufts Univ, Medford, MA 02155 USA
关键词
FRAMEWORK;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Modern machine learning demands a large amount of training data. Weak supervision is a promising approach to meet this demand. It aggregates multiple labeling functions (LFs) noisy, user-provided labeling heuristics to rapidly and cheaply curate probabilistic labels for large-scale unlabeled data. However, standard assumptions in weak supervision such as user-specified class bal- ance, similar accuracy of an LF in classifying different classes, and full knowledge of LF de- pendency at inference time might be undesir- able in practice. In response, we present Firebolt, a new weak supervision framework that seeks to operate under weaker assumptions. In particular, Firebolt learns the class balance and class-specific accuracy of LFs jointly from unlabeled data. It carries out inference in an efficient and interpretable manner. We analyze the parameter estimation error of Firebolt and characterize its impact on downstream model performance. Furthermore, we show that on five publicly available datasets, Firebolt outperforms a state-of-the-art weak supervision method by up to 5.8 points in AUC. We also provide a case study in the production setting of a tech company, where a Firebolt-supervised model outperforms the existing weakly-supervised production model by 1.3 points in AUC and speeds up label model training and inference from one hour to three minutes.
引用
收藏
页数:46
相关论文
共 50 条
  • [21] Minimax signal detection under weak noise assumptions
    Marteau C.
    Sapatinas T.
    Mathematical Methods of Statistics, 2017, 26 (4) : 282 - 298
  • [22] ITERATIVE OPTIMIZATION METHODS CONVERGING UNDER WEAK ASSUMPTIONS
    SCHUMACHER, K
    NUMERISCHE MATHEMATIK, 1975, 24 (05) : 443 - 456
  • [23] Weak and weaker prepositional complements
    Ruys, E. G.
    Linguistics in the Netherlands 2005, 2005, 22 : 151 - 163
  • [24] Verifiable random functions from weaker assumptions
    Jager, Tibor
    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2015, 9015 : 121 - 143
  • [25] Verifiable Random Functions from Weaker Assumptions
    Jager, Tibor
    THEORY OF CRYPTOGRAPHY (TCC 2015), PT II, 2015, 9015 : 121 - 143
  • [26] Weak interest in the weaker language
    Bernardini, Petra
    BILINGUALISM-LANGUAGE AND COGNITION, 2017, 20 (01) : 29 - 30
  • [27] Yen: Weak and getting weaker
    Krieger, AJ
    FORBES, 1997, 159 (05): : 48 - 48
  • [28] Fast Collective Activity Recognition Under Weak Supervision
    Zhang, Peizhen
    Tang, Yongyi
    Hu, Jian-Fang
    Zheng, Wei-Shi
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 29 - 43
  • [29] Learning to Segment Under Various Forms of Weak Supervision
    Xu, Jia
    Schwing, Alexander G.
    Urtasun, Raquel
    2015 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2015, : 3781 - 3790
  • [30] WeText: Scene Text Detection under Weak Supervision
    Tian, Shangxuan
    Lu, Shijian
    Li, Chongshou
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 1501 - 1509