The SciQA Scientific Question Answering Benchmark for Scholarly Knowledge

被引:4
|
作者
Auer, Soeren [1 ,2 ]
Barone, Dante A. C. [3 ]
Bartz, Cassiano [3 ]
Cortes, Eduardo G. [3 ]
Jaradeh, Mohamad Yaser [1 ,2 ]
Karras, Oliver [1 ]
Koubarakis, Manolis [4 ]
Mouromtsev, Dmitry [5 ]
Pliukhin, Dmitrii [5 ]
Radyush, Daniil [5 ]
Shilin, Ivan [5 ]
Stocker, Markus [1 ,2 ]
Tsalapati, Eleni [4 ]
机构
[1] TIB Leibniz Informat Ctr Sci & Technol, Hannover, Germany
[2] Leibniz Univ Hannover, L3S Res Ctr, Hannover, Germany
[3] Univ Fed Rio Grande do Sul, Inst Informat, Porto Alegre, Brazil
[4] Natl & Kapodistrian Univ Athens, Dept Informat & Telecommun, Athens, Greece
[5] ITMO Univ, Lab Informat Sci & Semant Technol, St Petersburg, Russia
基金
欧洲研究理事会; 欧盟地平线“2020”;
关键词
D O I
10.1038/s41598-023-33607-z
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Knowledge graphs have gained increasing popularity in the last decade in science and technology. However, knowledge graphs are currently relatively simple to moderate semantic structures that are mainly a collection of factual statements. Question answering (QA) benchmarks and systems were so far mainly geared towards encyclopedic knowledge graphs such as DBpedia and Wikidata. We present SciQA a scientific QA benchmark for scholarly knowledge. The benchmark leverages the Open Research Knowledge Graph (ORKG) which includes almost 170,000 resources describing research contributions of almost 15,000 scholarly articles from 709 research fields. Following a bottom-up methodology, we first manually developed a set of 100 complex questions that can be answered using this knowledge graph. Furthermore, we devised eight question templates with which we automatically generated further 2465 questions, that can also be answered with the ORKG. The questions cover a range of research fields and question types and are translated into corresponding SPARQL queries over the ORKG. Based on two preliminary evaluations, we show that the resulting SciQA benchmark represents a challenging task for next-generation QA systems. This task is part of the open competitions at the 22nd International Semantic Web Conference 2023 as the Scholarly Question Answering over Linked Data (QALD) Challenge.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] The SciQA Scientific Question Answering Benchmark for Scholarly Knowledge
    Sören Auer
    Dante A. C. Barone
    Cassiano Bartz
    Eduardo G. Cortes
    Mohamad Yaser Jaradeh
    Oliver Karras
    Manolis Koubarakis
    Dmitry Mouromtsev
    Dmitrii Pliukhin
    Daniil Radyush
    Ivan Shilin
    Markus Stocker
    Eleni Tsalapati
    [J]. Scientific Reports, 13
  • [2] Large Language Models for Scientific Question Answering: An Extensive Analysis of the SciQA Benchmark
    Lehmann, Jens
    Meloni, Antonello
    Motta, Enrico
    Osborne, Francesco
    Recupero, Diego Reforgiato
    Salatino, Angelo Antonio
    Vandati, Sahar
    [J]. SEMANTIC WEB, PT I, ESWC 2024, 2024, 14664 : 199 - 217
  • [3] Question Answering on Scholarly Knowledge Graphs
    Jaradeh, Mohamad Yaser
    Stocker, Markus
    Auer, Soeren
    [J]. DIGITAL LIBRARIES FOR OPEN KNOWLEDGE, TPDL 2020, 2020, 12246 : 19 - 32
  • [4] A-OKVQA: A Benchmark for Visual Question Answering Using World Knowledge
    Schwenk, Dustin
    Khandelwal, Apoorv
    Clark, Christopher
    Marino, Kenneth
    Mottaghi, Roozbeh
    [J]. COMPUTER VISION, ECCV 2022, PT VIII, 2022, 13668 : 146 - 162
  • [5] TempQuestions: A Benchmark for Temporal Question Answering
    Jia, Zhen
    Abujabal, Abdalghani
    Roy, Rishiraj Saha
    Stroetgen, Jannik
    Weikum, Gerhard
    [J]. COMPANION PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2018 (WWW 2018), 2018, : 1057 - 1062
  • [6] StreamingQA: A Benchmark for Adaptation to New Knowledge over Time in Question Answering Models
    Liska, Adam
    Kocisky, Tomas
    Gribovskaya, Elena
    Terzi, Tayfun
    Sezener, Eren
    Agrawal, Devang
    d'Autume, Cyprien de Masson
    Scholtes, Tim
    Zaheer, Manzil
    Young, Susannah
    Gilsenan-McMahon, Ellen
    Austin, Sophia
    Blunsom, Phil
    Lazaridou, Angeliki
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [7] FORECASTTKGQUESTIONS: A Benchmark for Temporal Question Answering and Forecasting over Temporal Knowledge Graphs
    Ding, Zifeng
    Li, Zongyue
    Qi, Ruoxia
    Wu, Jingpei
    He, Bailan
    Ma, Yunpu
    Meng, Zhao
    Chen, Shuo
    Liao, Ruotong
    Han, Zhen
    Tresp, Volker
    [J]. SEMANTIC WEB, ISWC 2023, PART I, 2023, 14265 : 541 - 560
  • [8] OK-VQA: A Visual Question Answering Benchmark Requiring External Knowledge
    Marino, Kenneth
    Rastegari, Mohammad
    Farhadi, Ali
    Mottaghi, Roozbeh
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 3190 - 3199
  • [9] A Cooking Knowledge Graph and Benchmark for Question Answering Evaluation in Lifelong Learning Scenarios
    Veron, Mathilde
    Penas, Anselmo
    Echegoyen, Guillermo
    Banerjee, Somnath
    Ghannay, Sahar
    Rosset, Sophie
    [J]. NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2020), 2020, 12089 : 94 - 101
  • [10] Select, Substitute, Search: A New Benchmark for Knowledge-Augmented Visual Question Answering
    Jain, Aman
    Kothyari, Mayank
    Kumar, Vishwajeet
    Jyothi, Preethi
    Ramakrishnan, Ganesh
    Chakrabarti, Soumen
    [J]. SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 2491 - 2498