Finding High-Quality Unstructured Submissions in General Crowdsourcing Tasks

被引:1
|
作者
Lyu, Shanshan [1 ,2 ]
Ouyang, Wentao [1 ]
Shen, Huawei [1 ]
Cheng, Xueqi [1 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, CAS Key Lab Network Data Sci & Technol, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
来源
关键词
Crowdsourcing; Quality estimation; General tasks; Unstructured submissions;
D O I
10.1007/978-3-030-01012-6_16
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The quality of crowdsourced work varies drastically from superior to inferior. As a consequence, the problem of automatically finding high-quality crowdsourced work is of great importance. A variety of aggregation methods have been proposed for multiple-choice tasks such as item labeling with structured claims. However, they do not apply to more general tasks, such as article writing and brand design, with unstructured submissions that cannot be aggregated. Recent work tackles this problem by asking another set of crowd workers to review and grade each submission, essentially transforming unstructured submissions into structured ratings that can be aggregated. Nevertheless, such an approach incurs unnecessary monetary cost and delay. In this paper, we address this problem by exploiting task requesters' historical feedback and directly modeling the submission quality, without the need of additional crowdsourced ratings. We first propose three sets of features, which try to characterize the submission quality from various perspectives, including the submissions themselves, the workers who make the submissions, and the interactions between task requesters and workers. We then propose two quality models, where one judges the submission quality independently and the other judges comparatively. These models not only incorporate features, but also take worker-specific factors into consideration. Experimental results on three large-scale data sets demonstrate that our models outperform general-purpose learning-to-rank methods such as Logistic Regression, RankBoost, and ListNet for finding high-quality crowdsourced submissions.
引用
收藏
页码:198 / 210
页数:13
相关论文
共 50 条
  • [1] High-quality submissions
    Fisher, Greg
    [J]. BUSINESS HORIZONS, 2020, 63 (03) : 249 - 251
  • [2] Aggregating Unstructured Submissions for Reliable Answers in Crowdsourcing Systems
    Kurup, Ayswarya R.
    Sajeev, G. P.
    [J]. PROCEEDINGS OF THE 2019 9TH INTERNATIONAL SYMPOSIUM ON EMBEDDED COMPUTING AND SYSTEM DESIGN (ISED 2019), 2019, : 16 - 22
  • [3] Statistical Quality Estimation for General Crowdsourcing Tasks
    Baba, Yukino
    Kashima, Hisashi
    [J]. 19TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'13), 2013, : 554 - 562
  • [4] Finding New, High-Quality Compounds
    Segall, Matthew
    [J]. GENETIC ENGINEERING & BIOTECHNOLOGY NEWS, 2012, 32 (04): : 18 - 19
  • [5] Finding High-Quality Review Articles
    Slawson, David C.
    Reed, Sean W.
    [J]. AMERICAN FAMILY PHYSICIAN, 2009, 79 (10) : 875 - 877
  • [6] Obtaining High-Quality Relevance Judgments Using Crowdsourcing
    Vuurens, Jeroen B. P.
    de Vries, Arjen P.
    [J]. IEEE INTERNET COMPUTING, 2012, 16 (05) : 20 - 27
  • [7] Finding High-Quality Item Attributes for Recommendation
    Zheng, Xiaolin
    Tan, Yanchao
    Wang, Yan
    Wei, Xiangyu
    Zhang, Shengjia
    Chen, Chaochao
    Li, Longfei
    Yang, Carl
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (08) : 7980 - 7993
  • [8] Feedback Based High-Quality Task Assignment in Collaborative Crowdsourcing
    Qiao, Liang
    Tang, Feilong
    Liu, Jiacheng
    [J]. PROCEEDINGS 2018 IEEE 32ND INTERNATIONAL CONFERENCE ON ADVANCED INFORMATION NETWORKING AND APPLICATIONS (AINA), 2018, : 1139 - 1146
  • [9] Efficient generation of high-quality unstructured surface and volume grids
    Marcum, DL
    [J]. ENGINEERING WITH COMPUTERS, 2001, 17 (03) : 211 - 233
  • [10] Efficient Generation of High-Quality Unstructured Surface and Volume Grids
    D. L. Marcum
    [J]. Engineering with Computers, 2001, 17 : 211 - 233