An empirical study of pre-trained language models in simple knowledge graph question answering

被引:9
|
作者
Hu, Nan [1 ]
Wu, Yike [1 ]
Qi, Guilin [1 ]
Min, Dehai [1 ]
Chen, Jiaoyan [2 ]
Pan, Jeff Z. [3 ]
Ali, Zafar [1 ]
机构
[1] Southeast Univ, Sch Comp Sci & Engn, 2 Dongda Rd, Nanjing 211189, Jiangsu, Peoples R China
[2] Univ Manchester, Dept Comp Sci, Oxford Rd, Manchester M13 9PL, England
[3] Univ Edinburgh, Sch Informat, 10 Crichton St, Edinburgh 2EH8 9AB, Scotland
关键词
Knowledge graph question answering; Pretrained language models; Accuracy and efficiency; Scalability;
D O I
10.1007/s11280-023-01166-y
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large-scale pre-trained language models (PLMs) such as BERT have recently achieved great success and become a milestone in natural language processing (NLP). It is now the consensus of the NLP community to adopt PLMs as the backbone for downstream tasks. In recent works on knowledge graph question answering (KGQA), BERT or its variants have become necessary in their KGQA models. However, there is still a lack of comprehensive research and comparison of the performance of different PLMs in KGQA. To this end, we summarize two basic KGQA frameworks based on PLMs without additional neural network modules to compare the performance of nine PLMs in terms of accuracy and efficiency. In addition, we present three benchmarks for larger-scale KGs based on the popular SimpleQuestions benchmark to investigate the scalability of PLMs. We carefully analyze the results of all PLMs-based KGQA basic frameworks on these benchmarks and two other popular datasets, WebQuestionSP and FreebaseQA, and find that knowledge distillation techniques and knowledge enhancement methods in PLMs are promising for KGQA. Furthermore, we test ChatGPT (https://chat.openai.com/), which has drawn a great deal of attention in the NLP community, demonstrating its impressive capabilities and limitations in zero-shot KGQA. We have released the code and benchmarks to promote the use of PLMs on KGQA (https://github.com/aannonymouuss/PLMs-in-Practical-KBQA).
引用
收藏
页码:2855 / 2886
页数:32
相关论文
共 50 条
  • [41] VQAttack: Transferable Adversarial Attacks on Visual Question Answering via Pre-trained Models
    Yin, Ziyi
    Ye, Muchao
    Zhang, Tianrong
    Wang, Jiaqi
    Liu, Han
    Chen, Jinghui
    Wang, Ting
    Ma, Fenglong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 7, 2024, : 6755 - 6763
  • [42] Textual Pre-Trained Models for Age Screening Across Community Question-Answering
    Figueroa, Alejandro
    Timilsina, Mohan
    IEEE ACCESS, 2024, 12 : 30030 - 30038
  • [43] A Study of Pre-trained Language Models in Natural Language Processing
    Duan, Jiajia
    Zhao, Hui
    Zhou, Qian
    Qiu, Meikang
    Liu, Meiqin
    2020 IEEE INTERNATIONAL CONFERENCE ON SMART CLOUD (SMARTCLOUD 2020), 2020, : 116 - 121
  • [44] DeFormer: Decomposing Pre-trained Transformers for Faster Question Answering
    Cao, Qingqing
    Trivedi, Harsh
    Balasubramanian, Aruna
    Balasubramanian, Niranjan
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 4487 - 4497
  • [45] An Empirical Study on Hyperparameter Optimization for Fine-Tuning Pre-trained Language Models
    Liu, Xueqing
    Wang, Chi
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 2286 - 2300
  • [46] Pre-Trained Language Models and Their Applications
    Wang, Haifeng
    Li, Jiwei
    Wu, Hua
    Hovy, Eduard
    Sun, Yu
    ENGINEERING, 2023, 25 : 51 - 65
  • [47] An Empirical Survey of the Effectiveness of Debiasing Techniques for Pre-trained Language Models
    Meade, Nicholas
    Poole-Dayan, Elinor
    Reddy, Siva
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 1878 - 1898
  • [48] Exploiting Pre-Trained Language Models for Black-Box Attack against Knowledge Graph Embeddings
    Yang, Guangqian
    Zhang, Lei
    Liu, Yi
    Xie, Hongtao
    Mao, Zhendong
    ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2024, 19 (01)
  • [49] Prompting disentangled embeddings for knowledge graph completion with pre-trained language model
    Geng, Yuxia
    Chen, Jiaoyan
    Zeng, Yuhang
    Chen, Zhuo
    Zhang, Wen
    Pan, Jeff Z.
    Wang, Yuxiang
    Xu, Xiaoliang
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 268
  • [50] NMT Enhancement based on Knowledge Graph Mining with Pre-trained Language Model
    Yang, Hao
    Qin, Ying
    Deng, Yao
    Wang, Minghan
    2020 22ND INTERNATIONAL CONFERENCE ON ADVANCED COMMUNICATION TECHNOLOGY (ICACT): DIGITAL SECURITY GLOBAL AGENDA FOR SAFE SOCIETY!, 2020, : 185 - 189