Large language models can better understand knowledge graphs than we thought

被引:0
|
作者
Dai, Xinbang [1 ]
Hua, Yuncheng [2 ]
Wu, Tongtong [3 ]
Sheng, Yang [4 ]
Ji, Qiu [4 ]
Qi, Guilin [1 ]
机构
[1] Southeast Univ, Nanjing, Jiangsu, Peoples R China
[2] Univ New South Wales, Sydney, NSW, Australia
[3] Monash Univ, Melbourne, Vic, Australia
[4] Nanjing Univ Posts & Telecommun, Nanjing, Jiangsu, Peoples R China
关键词
Knowledge graph; Large language model;
D O I
10.1016/j.knosys.2025.113060
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
When we integrate factual knowledge from knowledge graphs (KGs) into large language models (LLMs) to enhance their performance, the cost of injection through training increases with the scale of the models. Consequently, there is significant interest in developing prompt strategies that effectively incorporate KG information into LLMs. However, the community has not yet comprehensively understood how LLMs process and interpret KG information in different input formats and organizations within prompts, and researchers often rely on trial and error. To address this gap, we design extensive experiments to empirically study LLMs' comprehension of different KG prompts. At the literal level, we reveal LLMs' preferences for various input formats (from linearized triples to fluent natural language text). At the attention distribution level, we discuss the underlying mechanisms driving these preferences. We then investigate how the organization of structured knowledge impacts LLMs and evaluate LLMs' robustness in processing and utilizing KG information in practical scenarios. Our experiments show that (1) linearized triples are more effective than fluent NL text in helping LLMs understand KG information and answer fact-intensive questions; (2) Different LLMs exhibit varying preferences for different organizational formats of triples; (3) LLMs with larger scales are more susceptible to noisy, incomplete subgraphs.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] A medical question answering system using large language models and knowledge graphs
    Guo, Quan
    Cao, Shuai
    Yi, Zhang
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (11) : 8548 - 8564
  • [42] From language models to large-scale food and biomedical knowledge graphs
    Cenikj, Gjorgjina
    Strojnik, Lidija
    Angelski, Risto
    Ogrinc, Nives
    Seljak, Barbara Korousic
    Eftimov, Tome
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [43] Automated Retrosynthesis Planning of Macromolecules Using Large Language Models and Knowledge Graphs
    Ma, Qinyu
    Zhou, Yuhao
    Li, Jianfeng
    MACROMOLECULAR RAPID COMMUNICATIONS, 2025,
  • [44] Can Large Language Models Truly Understand Prompts? A Case Study with Negated Prompts
    Jang, Joel
    Ye, Seongheyon
    Seo, Minjoon
    TRANSFER LEARNING FOR NATURAL LANGUAGE PROCESSING WORKSHOP, VOL 203, 2022, 203 : 52 - 62
  • [45] WE CAN DO BETTER TO HELP STUDENTS UNDERSTAND PHYSICS
    BAUMAN, RP
    PHYSICS TEACHER, 1985, 23 (06): : 377 - 378
  • [46] Do Large Language Models Understand Us?
    Aguera y Arcas, Blaise
    DAEDALUS, 2022, 151 (02) : 183 - 197
  • [47] Implantable cardioverter defibrillators: even better than we thought?
    Buckley, Una
    Shivkumar, Kalyanam
    EUROPEAN HEART JOURNAL, 2015, 36 (26) : 1646 - 1648
  • [48] BODIES OF KNOWLEDGE AND KNOWLEDGE OF BODIES: "WE CAN KNOW MORE THAN WE CAN TELL"
    Machin, Amanda
    EPISTEMOLOGY & PHILOSOPHY OF SCIENCE-EPISTEMOLOGIYA I FILOSOFIYA NAUKI, 2018, 55 (04): : 84 - 97
  • [49] We can understand complex models (at least, sometimes)
    Hodges, James S.
    ENVIRONMETRICS, 2013, 24 (08) : 527 - 528
  • [50] Temporal Knowledge Graph Link Prediction Using Synergized Large Language Models and Temporal Knowledge Graphs
    Chen, Yao
    Shen, Yuming
    NEURAL COMPUTING FOR ADVANCED APPLICATIONS, NCAA 2024, PT III, 2025, 2183 : 33 - 45