An empirical study of pre-trained language models in simple knowledge graph question answering

被引:9
|
作者
Hu, Nan [1 ]
Wu, Yike [1 ]
Qi, Guilin [1 ]
Min, Dehai [1 ]
Chen, Jiaoyan [2 ]
Pan, Jeff Z. [3 ]
Ali, Zafar [1 ]
机构
[1] Southeast Univ, Sch Comp Sci & Engn, 2 Dongda Rd, Nanjing 211189, Jiangsu, Peoples R China
[2] Univ Manchester, Dept Comp Sci, Oxford Rd, Manchester M13 9PL, England
[3] Univ Edinburgh, Sch Informat, 10 Crichton St, Edinburgh 2EH8 9AB, Scotland
关键词
Knowledge graph question answering; Pretrained language models; Accuracy and efficiency; Scalability;
D O I
10.1007/s11280-023-01166-y
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large-scale pre-trained language models (PLMs) such as BERT have recently achieved great success and become a milestone in natural language processing (NLP). It is now the consensus of the NLP community to adopt PLMs as the backbone for downstream tasks. In recent works on knowledge graph question answering (KGQA), BERT or its variants have become necessary in their KGQA models. However, there is still a lack of comprehensive research and comparison of the performance of different PLMs in KGQA. To this end, we summarize two basic KGQA frameworks based on PLMs without additional neural network modules to compare the performance of nine PLMs in terms of accuracy and efficiency. In addition, we present three benchmarks for larger-scale KGs based on the popular SimpleQuestions benchmark to investigate the scalability of PLMs. We carefully analyze the results of all PLMs-based KGQA basic frameworks on these benchmarks and two other popular datasets, WebQuestionSP and FreebaseQA, and find that knowledge distillation techniques and knowledge enhancement methods in PLMs are promising for KGQA. Furthermore, we test ChatGPT (https://chat.openai.com/), which has drawn a great deal of attention in the NLP community, demonstrating its impressive capabilities and limitations in zero-shot KGQA. We have released the code and benchmarks to promote the use of PLMs on KGQA (https://github.com/aannonymouuss/PLMs-in-Practical-KBQA).
引用
收藏
页码:2855 / 2886
页数:32
相关论文
共 50 条
  • [21] Probing Pre-Trained Language Models for Disease Knowledge
    Alghanmi, Israa
    Espinosa-Anke, Luis
    Schockaert, Steven
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 3023 - 3033
  • [22] Audio-Aware Spoken Multiple-Choice Question Answering with Pre-Trained Language Models
    Kuo, Chia-Chih
    Chen, Kuan-Yu
    Luo, Shang-Bao
    IEEE/ACM Transactions on Audio Speech and Language Processing, 2021, 29 : 3170 - 3179
  • [23] Augmenting Pre-trained Language Models with QA-Memory for Open-Domain Question Answering
    Chen, Wenhu
    Verga, Pat
    de Jong, Michiel
    Wieting, John
    Cohen, William W.
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1597 - 1610
  • [24] Audio-Aware Spoken Multiple-Choice Question Answering With Pre-Trained Language Models
    Kuo, Chia-Chih
    Chen, Kuan-Yu
    Luo, Shang-Bao
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 : 3170 - 3179
  • [25] Dynamic Knowledge Distillation for Pre-trained Language Models
    Li, Lei
    Lin, Yankai
    Ren, Shuhuai
    Li, Peng
    Zhou, Jie
    Sun, Xu
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 379 - 389
  • [26] A Survey of Knowledge Enhanced Pre-Trained Language Models
    Hu, Linmei
    Liu, Zeyi
    Zhao, Ziwang
    Hou, Lei
    Nie, Liqiang
    Li, Juanzi
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (04) : 1413 - 1430
  • [27] Commonsense Knowledge Transfer for Pre-trained Language Models
    Zhou, Wangchunshu
    Le Bras, Ronan
    Choi, Yejin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5946 - 5960
  • [28] BERT-MK: Integrating Graph Contextualized Knowledge into Pre-trained Language Models
    He, Bin
    Zhou, Di
    Xiao, Jinghui
    Jiang, Xin
    Liu, Qun
    Yuan, Nicholas Jing
    Xu, Tong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 2281 - 2290
  • [29] KG-prompt: Interpretable knowledge graph prompt for pre-trained language models
    Chen, Liyi
    Liu, Jie
    Duan, Yutai
    Wang, Runze
    KNOWLEDGE-BASED SYSTEMS, 2025, 311
  • [30] Scalable Educational Question Generation with Pre-trained Language Models
    Bulathwela, Sahan
    Muse, Hamze
    Yilmaz, Emine
    ARTIFICIAL INTELLIGENCE IN EDUCATION, AIED 2023, 2023, 13916 : 327 - 339