The SciQA Scientific Question Answering Benchmark for Scholarly Knowledge

被引:0
|
作者
Sören Auer
Dante A. C. Barone
Cassiano Bartz
Eduardo G. Cortes
Mohamad Yaser Jaradeh
Oliver Karras
Manolis Koubarakis
Dmitry Mouromtsev
Dmitrii Pliukhin
Daniil Radyush
Ivan Shilin
Markus Stocker
Eleni Tsalapati
机构
[1] TIB—Leibniz Information Centre for Science and Technology,L3S Research Center
[2] Leibniz University Hannover,Institute of Informatics
[3] Federal University of Rio Grande do Sul,Department of Informatics and Telecommunications
[4] National and Kapodistrian University of Athens,Laboratory of Information Science and Semantic Technologies
[5] ITMO University,undefined
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Knowledge graphs have gained increasing popularity in the last decade in science and technology. However, knowledge graphs are currently relatively simple to moderate semantic structures that are mainly a collection of factual statements. Question answering (QA) benchmarks and systems were so far mainly geared towards encyclopedic knowledge graphs such as DBpedia and Wikidata. We present SciQA a scientific QA benchmark for scholarly knowledge. The benchmark leverages the Open Research Knowledge Graph (ORKG) which includes almost 170,000 resources describing research contributions of almost 15,000 scholarly articles from 709 research fields. Following a bottom-up methodology, we first manually developed a set of 100 complex questions that can be answered using this knowledge graph. Furthermore, we devised eight question templates with which we automatically generated further 2465 questions, that can also be answered with the ORKG. The questions cover a range of research fields and question types and are translated into corresponding SPARQL queries over the ORKG. Based on two preliminary evaluations, we show that the resulting SciQA benchmark represents a challenging task for next-generation QA systems. This task is part of the open competitions at the 22nd International Semantic Web Conference 2023 as the Scholarly Question Answering over Linked Data (QALD) Challenge.
引用
收藏
相关论文
共 50 条
  • [1] The SciQA Scientific Question Answering Benchmark for Scholarly Knowledge
    Auer, Soeren
    Barone, Dante A. C.
    Bartz, Cassiano
    Cortes, Eduardo G.
    Jaradeh, Mohamad Yaser
    Karras, Oliver
    Koubarakis, Manolis
    Mouromtsev, Dmitry
    Pliukhin, Dmitrii
    Radyush, Daniil
    Shilin, Ivan
    Stocker, Markus
    Tsalapati, Eleni
    [J]. SCIENTIFIC REPORTS, 2023, 13 (01)
  • [2] Large Language Models for Scientific Question Answering: An Extensive Analysis of the SciQA Benchmark
    Lehmann, Jens
    Meloni, Antonello
    Motta, Enrico
    Osborne, Francesco
    Recupero, Diego Reforgiato
    Salatino, Angelo Antonio
    Vandati, Sahar
    [J]. SEMANTIC WEB, PT I, ESWC 2024, 2024, 14664 : 199 - 217
  • [3] Question Answering on Scholarly Knowledge Graphs
    Jaradeh, Mohamad Yaser
    Stocker, Markus
    Auer, Soeren
    [J]. DIGITAL LIBRARIES FOR OPEN KNOWLEDGE, TPDL 2020, 2020, 12246 : 19 - 32
  • [4] A-OKVQA: A Benchmark for Visual Question Answering Using World Knowledge
    Schwenk, Dustin
    Khandelwal, Apoorv
    Clark, Christopher
    Marino, Kenneth
    Mottaghi, Roozbeh
    [J]. COMPUTER VISION, ECCV 2022, PT VIII, 2022, 13668 : 146 - 162
  • [5] TempQuestions: A Benchmark for Temporal Question Answering
    Jia, Zhen
    Abujabal, Abdalghani
    Roy, Rishiraj Saha
    Stroetgen, Jannik
    Weikum, Gerhard
    [J]. COMPANION PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2018 (WWW 2018), 2018, : 1057 - 1062
  • [6] FORECASTTKGQUESTIONS: A Benchmark for Temporal Question Answering and Forecasting over Temporal Knowledge Graphs
    Ding, Zifeng
    Li, Zongyue
    Qi, Ruoxia
    Wu, Jingpei
    He, Bailan
    Ma, Yunpu
    Meng, Zhao
    Chen, Shuo
    Liao, Ruotong
    Han, Zhen
    Tresp, Volker
    [J]. SEMANTIC WEB, ISWC 2023, PART I, 2023, 14265 : 541 - 560
  • [7] StreamingQA: A Benchmark for Adaptation to New Knowledge over Time in Question Answering Models
    Liska, Adam
    Kocisky, Tomas
    Gribovskaya, Elena
    Terzi, Tayfun
    Sezener, Eren
    Agrawal, Devang
    d'Autume, Cyprien de Masson
    Scholtes, Tim
    Zaheer, Manzil
    Young, Susannah
    Gilsenan-McMahon, Ellen
    Austin, Sophia
    Blunsom, Phil
    Lazaridou, Angeliki
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [8] OK-VQA: A Visual Question Answering Benchmark Requiring External Knowledge
    Marino, Kenneth
    Rastegari, Mohammad
    Farhadi, Ali
    Mottaghi, Roozbeh
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 3190 - 3199
  • [9] A Cooking Knowledge Graph and Benchmark for Question Answering Evaluation in Lifelong Learning Scenarios
    Veron, Mathilde
    Penas, Anselmo
    Echegoyen, Guillermo
    Banerjee, Somnath
    Ghannay, Sahar
    Rosset, Sophie
    [J]. NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2020), 2020, 12089 : 94 - 101
  • [10] Select, Substitute, Search: A New Benchmark for Knowledge-Augmented Visual Question Answering
    Jain, Aman
    Kothyari, Mayank
    Kumar, Vishwajeet
    Jyothi, Preethi
    Ramakrishnan, Ganesh
    Chakrabarti, Soumen
    [J]. SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 2491 - 2498