GenKP: generative knowledge prompts for enhancing large language modelsGenKP: generative knowledge prompts for enhancing large language modelsX. Li et al.

被引:0
|
作者
Xinbai Li [1 ]
Shaowen Peng [1 ]
Shuntaro Yada [1 ]
Shoko Wakamiya [2 ]
Eiji Aramaki [1 ]
机构
[1] Nara Institute of Science and Technology,
[2] University of Tsukuba,undefined
关键词
Large language models; Knowledge graph; Knowledge prompts; In-context learning;
D O I
10.1007/s10489-025-06318-3
中图分类号
学科分类号
摘要
Large language models (LLMs) have demonstrated extensive capabilities across various natural language processing (NLP) tasks. Knowledge graphs (KGs) harbor vast amounts of facts, furnishing external knowledge for language models. The structured knowledge extracted from KGs must undergo conversion into sentences to align with the input format required by LLMs. Previous research has commonly utilized methods such as triple conversion and template-based conversion. However, sentences converted using existing methods frequently encounter issues such as semantic incoherence, ambiguity, and unnaturalness, which distort the original intent, and deviate the sentences from the facts. Meanwhile, despite the improvement that knowledge-enhanced pre-training and prompt-tuning methods have achieved in small-scale models, they are difficult to implement for LLMs in the absence of computational resources. The advanced comprehension of LLMs facilitates in-context learning (ICL), thereby enhancing their performance without the need for additional training. In this paper, we propose a knowledge prompts generation method, GenKP, which injects knowledge into LLMs by ICL. Compared to inserting triple-conversion or templated-conversion knowledge without selection, GenKP entails generating knowledge samples using LLMs in conjunction with KGs and makes a trade-off of knowledge samples through weighted verification and BM25 ranking, reducing knowledge noise. Experimental results illustrate that incorporating knowledge prompts enhances the performance of LLMs. Furthermore, LLMs augmented with GenKP exhibit superior improvements compared to the methods utilizing triple and template-based knowledge injection.
引用
收藏
相关论文
共 28 条
  • [1] GenKP: generative knowledge prompts for enhancing large language models
    Li, Xinbai
    Peng, Shaowen
    Yada, Shuntaro
    Wakamiya, Shoko
    Aramaki, Eiji
    APPLIED INTELLIGENCE, 2025, 55 (06)
  • [3] Enhancing Orthopedic Knowledge Assessments: The Performance of Specialized Generative Language Model Optimization
    Zhou, Hong
    Wang, Hong-lin
    Duan, Yu-yu
    Yan, Zi-neng
    Luo, Rui
    Lv, Xiang-xin
    Xie, Yi
    Zhang, Jia-yao
    Yang, Jia-ming
    Xue, Ming-di
    Fang, Ying
    Lu, Lin
    Liu, Peng-ran
    Ye, Zhe-wei
    CURRENT MEDICAL SCIENCE, 2024, : 1001 - 1005
  • [4] Generative Multi-Modal Knowledge Retrieval with Large Language Models
    Long, Xinwei
    Zeng, Jiali
    Meng, Fandong
    Ma, Zhiyuan
    Zhang, Kaiyan
    Zhou, Bowen
    Zhou, Jie
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 18733 - 18741
  • [5] Enhancing Large Language Models Through External Domain Knowledge
    Welz, Laslo
    Lanquillon, Carsten
    ARTIFICIAL INTELLIGENCE IN HCI, PT III, AI-HCI 2024, 2024, 14736 : 135 - 146
  • [6] Better Together: Enhancing Generative Knowledge Graph Completion with Language Models and Neighborhood Information
    Chepurova, Alla
    Bulatov, Aydar
    Kuratov, Yuri
    Burtsev, Mikhail
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 5306 - 5316
  • [7] PU-GEN: Enhancing generative commonsense reasoning for language models with human-centered knowledge
    Seo, Jaehyung
    Oh, Dongsuk
    Eo, Sugyeong
    Park, Chanjun
    Yang, Kisu
    Moon, Hyeonseok
    Park, Kinam
    Lim, Heuiseok
    KNOWLEDGE-BASED SYSTEMS, 2022, 256
  • [8] Enhancing Bottleneck Analysis in Ship Manufacturing with Knowledge Graphs and Large Language Models
    Ma, Yanjun
    Wu, Tao
    Zhou, Bin
    Liang, Xiaoyang
    Du, Jiwang
    Bao, Jinsong
    MACHINES, 2025, 13 (03)
  • [9] Enhancing Sequential Recommenders with Augmented Knowledge from Aligned Large Language Models
    Ren, Yankun
    Chen, Zhongde
    Yang, Xinxing
    Li, Longfei
    Jiang, Cong
    Cheng, Lei
    Zhang, Bo
    Mo, Linjian
    Zhou, Jun
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 345 - 354
  • [10] Enhancing emergency decision-making with knowledge graphs and large language models
    Chen, Minze
    Tao, Zhenxiang
    Tang, Weitong
    Qin, Tingxin
    Yang, Rui
    Zhu, Chunli
    INTERNATIONAL JOURNAL OF DISASTER RISK REDUCTION, 2024, 113