Crowdsourcing High Quality Labels with a Tight Budget

被引:32
|
作者
Li, Qi [1 ]
Ma, Fenglong [1 ]
Gao, Jing [1 ]
Su, Lu [1 ]
Quinn, Christopher J. [2 ]
机构
[1] SUNY Buffalo, Buffalo, NY 14620 USA
[2] Purdue Univ, W Lafayette, IN 47907 USA
关键词
DESIGN;
D O I
10.1145/2835776.2835797
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the past decade, commercial crowdsourcing platforms have revolutionized the ways of classifying and annotating data, especially for large datasets. Obtaining labels for a single instance can be inexpensive, but for large datasets, it is important to allocate budgets wisely. With limited budgets, requesters must trade-off between the quantity of labeled instances and the quality of the final results. Existing budget allocation methods can achieve good quantity but cannot guarantee high quality of individual instances under a tight budget. However, in some scenarios, requesters may be willing to label fewer instances but of higher quality. Moreover, they may have different requirements on quality for different tasks. To address these challenges, we propose a flexible budget allocation framework called Requallo. Requallo allows requesters to set their specific requirements on the labeling quality and maximizes the number of labeled instances that achieve the quality requirement under a tight budget. The budget allocation problem is modeled as a Markov decision process and a sequential labeling policy is produced. The proposed policy greedily searches for the instance to query next as the one that can provide the maximum reward for the goal. The Requallo framework is further extended to consider worker reliability so that the budget can be better allocated. Experiments on two real-world crowdsourcing tasks as well as a simulated task demonstrate that when the budget is tight, the proposed Requallo framework outperforms existing state-of-the-art budget allocation methods from both quantity and quality aspects.
引用
收藏
页码:237 / 246
页数:10
相关论文
共 50 条
  • [1] Optimal Budget Allocation for Crowdsourcing Labels for Graphs
    Kulkarni, Adithya
    Chakraborty, Mohna
    Xie, Sihong
    Li, Qi
    [J]. UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 1154 - 1163
  • [2] Balancing quality and budget considerations in mobile crowdsourcing
    Miao, Chunyan
    Yu, Han
    Shen, Zhiqi
    Leung, Cyril
    [J]. DECISION SUPPORT SYSTEMS, 2016, 90 : 56 - 64
  • [3] Quality and Budget Aware Task Allocation for Spatial Crowdsourcing
    Yu, Han
    Miao, Chunyan
    Shen, Zhiqi
    Leung, Cyril
    [J]. PROCEEDINGS OF THE 2015 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS (AAMAS'15), 2015, : 1689 - 1690
  • [4] Improving the Quality of Crowdsourcing Labels by Combination of Golden Data and Incentive
    Yang, Peijun
    Cai, Haibin
    Zheng, Zhiming
    [J]. PROCEEDINGS OF 2018 12TH IEEE INTERNATIONAL CONFERENCE ON ANTI-COUNTERFEITING, SECURITY, AND IDENTIFICATION (ASID), 2018, : 10 - 15
  • [5] BudgetFix: Budget Limited Crowdsourcing for Interdependent Task Allocation with Quality Guarantees
    Long Tran-Thanh
    Trung Dong Huynh
    Rosenfeld, Avi
    Ramchurn, Sarvapali
    Jennings, Nicholas R.
    [J]. AAMAS'14: PROCEEDINGS OF THE 2014 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2014, : 477 - 484
  • [6] Quality-Assure and Budget-Aware Task Assignment for Spatial Crowdsourcing
    Wang, Qing
    He, Wei
    Wang, Xinjun
    Cui, Lizhen
    [J]. COLLABORATE COMPUTING: NETWORKING, APPLICATIONS AND WORKSHARING, COLLABORATECOM 2016, 2017, 201 : 60 - 70
  • [7] TIGHT BUDGET DRIVE
    POTASZNIK, J
    [J]. BYTE, 1987, 12 (09): : 12 - 12
  • [8] SCIENCE ON A TIGHT BUDGET
    FLOWERS, BH
    [J]. SCIENCE, 1970, 170 (3965) : 1361 - &
  • [9] REINSTRUMENTING ON A TIGHT BUDGET
    LYTLE, TR
    [J]. INTECH, 1988, 35 (05) : 49 - 51
  • [10] INCENTIVES ON A TIGHT BUDGET
    HOLSCHUH, M
    [J]. AGRICULTURAL ENGINEERING, 1992, 73 (03): : 8 - 8