Benchmarking question answering systems

被引:16
|
作者
Usbeck, Ricardo [1 ]
Roeder, Michael [1 ]
Hoffmann, Michael [3 ]
Conrads, Felix [1 ]
Huthmann, Jonathan [3 ]
Ngonga-Ngomo, Axel-Cyrille [1 ]
Demmler, Christian [3 ]
Unger, Christina [2 ]
机构
[1] Paderborn Univ, DICE Data Sci Grp, Paderborn, Germany
[2] Univ Bielefeld, CITEC, Bielefeld, Germany
[3] Univ Leipzig, AKSW Grp, Leipzig, Germany
基金
欧盟地平线“2020”;
关键词
Factoid question answering; benchmarking; repeatable open research;
D O I
10.3233/SW-180312
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The necessity of making the Semantic Web more accessible for lay users, alongside the uptake of interactive systems and smart assistants for the Web, have spawned a new generation of RDF-based question answering systems. However, fair evaluation of these systems remains a challenge due to the different type of answers that they provide. Hence, repeating current published experiments or even benchmarking on the same datasets remains a complex and time-consuming task. We present a novel online benchmarking platform for question answering (QA) that relies on the FAIR principles to support the fine-grained evaluation of question answering systems. We detail how the platform addresses the fair benchmarking platform of question answering systems through the rewriting of URIs and URLs. In addition, we implement different evaluation metrics, measures, datasets and pre-implemented systems as well as methods to work with novel formats for interactive and non-interactive benchmarking of question answering systems. Our analysis of current frameworks shows that most of the current frameworks are tailored towards particular datasets and challenges but do not provide generic models. In addition, while most frameworks perform well in the annotation of entities and properties, the generation of SPARQL queries from annotated text remains a challenge.
引用
收藏
页码:293 / 304
页数:12
相关论文
共 50 条
  • [31] Using contradictions improves question answering systems
    Fortier-Dubois, Etienne
    Rosati, Domenic
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 827 - 840
  • [32] Versatile question answering systems: Seeing in synthesis
    Mittal S.
    Mittal A.
    International Journal of Intelligent Information and Database Systems, 2011, 5 (02) : 119 - 142
  • [33] Intelligent Question - Answering Systems: Review of research
    Tomljanovic, J.
    Pavlic, M.
    Katic, M. Asenbrener
    2014 37TH INTERNATIONAL CONVENTION ON INFORMATION AND COMMUNICATION TECHNOLOGY, ELECTRONICS AND MICROELECTRONICS (MIPRO), 2014, : 1228 - 1233
  • [34] Combining semantic information in question answering systems
    Moreda, Paloma
    Llorens, Hector
    Saquete, Estela
    Palomar, Manuel
    INFORMATION PROCESSING & MANAGEMENT, 2011, 47 (06) : 870 - 885
  • [35] Question-Answering Systems: Development and Prospects
    Lapshin, V. A.
    AUTOMATIC DOCUMENTATION AND MATHEMATICAL LINGUISTICS, 2012, 46 (03) : 138 - 145
  • [36] A survey on legal question-answering systems
    Martinez-Gil, Jorge
    COMPUTER SCIENCE REVIEW, 2023, 48
  • [37] Question-answering systems: Development and prospects
    V. A. Lapshin
    Automatic Documentation and Mathematical Linguistics, 2012, 46 (3) : 138 - 145
  • [38] Toward a document model for question answering systems
    Pérez-Coutiño, M
    Solorio, T
    Montes-y-Gómez, M
    López-López, A
    Villaseñor-Pineda, L
    ADVANCES IN WEB INTELLIGENCE, PROCEEDINGS, 2004, 3034 : 145 - 154
  • [39] Evolution of Reading Comprehension and Question Answering Systems
    Krishnamoorthy, Venkatesh
    BIG DATA, IOT, AND AI FOR A SMARTER FUTURE, 2021, 185 : 231 - 238
  • [40] A Framework of Evaluation for Question-Answering Systems
    El Ayari, Sarra
    Grau, Brigitte
    ADVANCES IN INFORMATION RETRIEVAL, PROCEEDINGS, 2009, 5478 : 744 - 748