A few-shot learning method based on knowledge graph in large language models

被引:0
|
作者
Wang, Feilong [1 ,2 ]
Shi, Donghui [1 ,2 ]
Aguilar, Jose [3 ,4 ,5 ]
Cui, Xinyi [1 ]
机构
[1] Anhui Jianzhu Univ, Sch Elect & Informat Engn, Dept Comp Engn, Hefei 230601, Peoples R China
[2] Mass Spectrometry Key Technol R&D&Clin Applicat An, Hefei 230601, Peoples R China
[3] Univ EAFIT, Grp Invest IDI T, Medellin, Colombia
[4] Univ Los Andes, Ctr Estudios Microelect & Sistemas Distribuidos, Merida, Venezuela
[5] IMDEA Networks Inst, Madrid, Spain
关键词
Large language model; Few-shot learning; Fine-tuning; Knowledge-driven dialog generation; Knowledge graph;
D O I
10.1007/s41060-024-00699-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emergence of large language models has significantly transformed natural language processing and text generation. Fine-tuning these models for specific domains enables them to generate answers tailored to the unique requirements of those fields, such as in legal or medical domains. However, these models often perform poorly in few-shot scenarios. Herein, the challenges of data scarcity in fine-tuning large language models in low-sample scenarios were addressed by proposing three different KDGI (Knowledge-Driven Dialog Generation Instances) generation strategies, including entity-based KDGI generation, relation-based KDGI generation, and semantic-based multi-level KDGI generation. These strategies aimed to enhance few-shot datasets to address the issue of low fine-tuning metrics caused by insufficient data. Specifically, knowledge graphs were utilized to define the distinct KDGI generation strategies for enhancing few-shot data. Subsequently, these KDGI data were employed to fine-tune the large language model using the P-tuning v2 approach. Through multiple experiments, the effectiveness of the three KDGI generation strategies was validated using BLEU and ROUGE metrics, and the fine-tuning benefits of few-shot learning on large language models were confirmed. To further evaluate the effectiveness of KDGI, additional experiments were conducted, including LoRA-based fine-tuning in the medical domain and comparative studies leveraging Mask Language Model augmentation, back-translation, and noise injection methods. Consequently, the paper proposes a reference method for leveraging knowledge graphs in prompt data engineering, which shows potential in facilitating few-shot learning for fine-tuning large language models.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] Knowledge transfer based hierarchical few-shot learning via tree-structured knowledge graph
    Zhong Zhang
    Zhiping Wu
    Hong Zhao
    Minjie Hu
    International Journal of Machine Learning and Cybernetics, 2023, 14 : 281 - 294
  • [32] Knowledge transfer based hierarchical few-shot learning via tree-structured knowledge graph
    Zhang, Zhong
    Wu, Zhiping
    Zhao, Hong
    Hu, Minjie
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2023, 14 (01) : 281 - 294
  • [33] Graph Few-shot Learning with Attribute Matching
    Wang, Ning
    Luo, Minnan
    Ding, Kaize
    Zhang, Lingling
    Li, Jundong
    Zheng, Qinghua
    CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, : 1545 - 1554
  • [34] Meta-Learning Based Dynamic Adaptive Relation Learning for Few-Shot Knowledge Graph Completion
    Cai, Linqin
    Wang, Lingjun
    Yuan, Rongdi
    Lai, Tingjie
    BIG DATA RESEARCH, 2023, 33
  • [35] Fairness-guided Few-shot Prompting for Large Language Models
    Ma, Huan
    Zhang, Changqing
    Bian, Yatao
    Liu, Lemao
    Zhang, Zhirui
    Zhao, Peilin
    Zhang, Shu
    Fu, Huazhu
    Hu, Qinghua
    Wu, Bingzhe
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [36] LLaFS: When Large Language Models Meet Few-Shot Segmentation
    Zhu, Lanyun
    Chen, Tianrun
    Ji, Deyi
    Ye, Jieping
    Liu, Jun
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 3065 - 3075
  • [37] Political Bias of Large Language Models in Few-Shot News Summarization
    Onishi, Takeshi
    Caverlee, James
    ADVANCES IN BIAS AND FAIRNESS IN INFORMATION RETRIEVAL, BIAS 2024, 2025, 2227 : 32 - 45
  • [38] Refactoring Programs Using Large Language Models with Few-Shot Examples
    Shirafuji, Atsushi
    Oda, Yusuke
    Suzuki, Jun
    Morishita, Makoto
    Watanobe, Yutaka
    PROCEEDINGS OF THE 2023 30TH ASIA-PACIFIC SOFTWARE ENGINEERING CONFERENCE, APSEC 2023, 2023, : 151 - 160
  • [39] TabLLM: Few-shot Classification of Tabular Data with Large Language Models
    Hegselmann, Stefan
    Buendia, Alejandro
    Lang, Hunter
    Agrawal, Monica
    Jiang, Xiaoyi
    Sontag, David
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 206, 2023, 206
  • [40] Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm
    Reynolds, Laria
    McDonell, Kyle
    EXTENDED ABSTRACTS OF THE 2021 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI'21), 2021,