Benchmarking question answering systems

被引:15
|
作者
Usbeck, Ricardo [1 ]
Roeder, Michael [1 ]
Hoffmann, Michael [3 ]
Conrads, Felix [1 ]
Huthmann, Jonathan [3 ]
Ngonga-Ngomo, Axel-Cyrille [1 ]
Demmler, Christian [3 ]
Unger, Christina [2 ]
机构
[1] Paderborn Univ, DICE Data Sci Grp, Paderborn, Germany
[2] Univ Bielefeld, CITEC, Bielefeld, Germany
[3] Univ Leipzig, AKSW Grp, Leipzig, Germany
基金
欧盟地平线“2020”;
关键词
Factoid question answering; benchmarking; repeatable open research;
D O I
10.3233/SW-180312
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The necessity of making the Semantic Web more accessible for lay users, alongside the uptake of interactive systems and smart assistants for the Web, have spawned a new generation of RDF-based question answering systems. However, fair evaluation of these systems remains a challenge due to the different type of answers that they provide. Hence, repeating current published experiments or even benchmarking on the same datasets remains a complex and time-consuming task. We present a novel online benchmarking platform for question answering (QA) that relies on the FAIR principles to support the fine-grained evaluation of question answering systems. We detail how the platform addresses the fair benchmarking platform of question answering systems through the rewriting of URIs and URLs. In addition, we implement different evaluation metrics, measures, datasets and pre-implemented systems as well as methods to work with novel formats for interactive and non-interactive benchmarking of question answering systems. Our analysis of current frameworks shows that most of the current frameworks are tailored towards particular datasets and challenges but do not provide generic models. In addition, while most frameworks perform well in the annotation of entities and properties, the generation of SPARQL queries from annotated text remains a challenge.
引用
收藏
页码:293 / 304
页数:12
相关论文
共 50 条
  • [1] Question/Answering Systems
    Visser, Ubbo
    [J]. KUNSTLICHE INTELLIGENZ, 2012, 26 (02): : 191 - 195
  • [2] QUESTION ANSWERING SYSTEMS
    Tomljanovic, Jasminka
    Krsnik, Marina
    Pavlic, Mile
    [J]. ZBORNIK VELEUCILISTA U RIJECI-JOURNAL OF THE POLYTECHNICS OF RIJEKA, 2014, 2 (01): : 177 - 195
  • [3] A Review of Question Answering Systems
    Ojokoh, Bolanle
    Adebisi, Emmanuel
    [J]. JOURNAL OF WEB ENGINEERING, 2018, 17 (08): : 717 - 758
  • [4] Comparison of Question Answering Systems
    Dodiya, Tripti
    Jain, Sonal
    [J]. INTELLIGENT INFORMATICS, 2013, 182 : 99 - +
  • [5] Question Classification for Arabic Question Answering Systems
    Al Chalabi, Hani Maluf
    Ray, Santosh Kumar
    Shaalan, Khaled
    [J]. 2015 INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGY RESEARCH (ICTRC), 2015, : 310 - 313
  • [6] Benchmarking Entity Linking for Question Answering over Knowledge Graphs
    Echegoyen, Guillermo
    Rodrigo, Alvaro
    Penas, Anselmo
    [J]. PROCESAMIENTO DEL LENGUAJE NATURAL, 2019, (63): : 121 - 128
  • [7] Beyond Accuracy: A Consolidated Tool for Visual Question Answering Benchmarking
    Vath, Dirk
    Tilli, Pascal
    Ngoc Thang Vu
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021): PROCEEDINGS OF SYSTEM DEMONSTRATIONS, 2021, : 114 - 123
  • [8] Question Recommendation for Collaborative Question Answering Systems with RankSLDA
    San Pedro, Jose
    Karatzoglou, Alexandros
    [J]. PROCEEDINGS OF THE 8TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'14), 2014, : 193 - 200
  • [9] Improving the Robustness of Question Answering Systems to Question Paraphrasing
    Gan, Wee Chung
    Ng, Hwee Tou
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 6065 - 6075
  • [10] Connecting Question Answering and Conversational AgentsContextualizing German Questions for Interactive Question Answering Systems
    Ulli Waltinger
    Alexa Breuing
    Ipke Wachsmuth
    [J]. KI - Künstliche Intelligenz, 2012, 26 (4) : 381 - 390