IIRC: A Dataset of Incomplete Information Reading Comprehension Questions

被引:0
|
作者
Ferguson, James [1 ,2 ]
Gardner, Matt [2 ]
Hajishirzi, Hannaneh [1 ,2 ]
Khot, Tushar [2 ]
Dasigi, Pradeep [2 ]
机构
[1] Univ Washington, Seattle, WA USA
[2] Allen Inst AI, Seattle, WA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Humans often have to read multiple documents to address their information needs. However, most existing reading comprehension (RC) tasks only focus on questions for which the contexts provide all the information required to answer them, thus not evaluating a system's performance at identifying a potential lack of sufficient information and locating sources for that information. To fill this gap, we present a dataset, IIRC, with more than 13K questions over paragraphs from English Wikipedia that provide only partial information to answer them, with the missing information occurring in one or more linked documents. The questions were written by crowd workers who did not have access to any of the linked documents, leading to questions that have little lexical overlap with the contexts where the answers appear. This process also gave many questions without answers, and those that require discrete reasoning, increasing the difficulty of the task. We follow recent modeling work on various reading comprehension datasets to construct a baseline model for this dataset, finding that it achieves 31.1% F1 on this task, while estimated human performance is 88.4%. The dataset, code for the baseline system, and a leaderboard can be found at https://allennlp.org/iirc.
引用
下载
收藏
页码:1137 / 1147
页数:11
相关论文
共 50 条
  • [1] TORQUE: A Reading Comprehension Dataset of Temporal Ordering Questions
    Ning, Qiang
    Wu, Hao
    Han, Rujun
    Peng, Nanyun
    Gardner, Matt
    Roth, Dan
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1158 - 1172
  • [2] QUOREF: A Reading Comprehension Dataset with Questions Requiring Coreferential Reasoning
    Dasigi, Pradeep
    Liu, Nelson F.
    Marasovic, Ana
    Smith, Noah A.
    Gardner, Matt
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 5925 - 5932
  • [3] TibetanQA2.0: Dataset with Unanswerable Questions for Tibetan Machine Reading Comprehension
    Zhengcuo Dan
    Yuan Sun
    Data Intelligence, 2024, 6 (04) : 1158 - 1167
  • [4] Developing reading comprehension questions
    Day, Richard R.
    Park, Jeong-Suk
    READING IN A FOREIGN LANGUAGE, 2005, 17 (01): : 60 - 73
  • [5] Effects of advance questions on reading comprehension
    Kreiner, DS
    JOURNAL OF GENERAL PSYCHOLOGY, 1996, 123 (04): : 352 - 364
  • [6] Can LLMs Grade Open Response Reading Comprehension Questions? An Empirical Study Using the ROARs Dataset
    Henkel, Owen
    Hills, Libby
    Roberts, Bill
    McGrane, Joshua
    INTERNATIONAL JOURNAL OF ARTIFICIAL INTELLIGENCE IN EDUCATION, 2024,
  • [7] BIOMRC: A Dataset for Biomedical Machine Reading Comprehension
    Stavropoulos, Petros
    Pappas, Dimitris
    Androutsopoulos, Ion
    McDonald, Ryan
    19TH SIGBIOMED WORKSHOP ON BIOMEDICAL LANGUAGE PROCESSING (BIONLP 2020), 2020, : 140 - 149
  • [8] BioRead: A New Dataset for Biomedical Reading Comprehension
    Pappas, Dimitris
    Androutsopoulos, Ion
    Papageorgiou, Haris
    PROCEEDINGS OF THE ELEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2018), 2018, : 2771 - 2776
  • [9] PolicyQA: A Reading Comprehension Dataset for Privacy Policies
    Ahmad, Wasi Uddin
    Chi, Jianfeng
    Tian, Yuan
    Chang, Kai-Wei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 743 - 749
  • [10] Enhance Machine Reading Comprehension on Multiple Sentence Questions with Gated and Dense Coreference Information
    Tretasayuth, Nattachai
    Vateekul, Peerapon
    Boonkwan, Prachya
    2018 15TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER SCIENCE AND SOFTWARE ENGINEERING (JCSSE), 2018, : 279 - 284