InstructEdit: Instruction-Based Knowledge Editing for Large Language Models

被引:0
|
作者
Zhang, Ningyu [1 ]
Tian, Bozhong [1 ]
Cheng, Siyuan [2 ]
Liang, Xiaozhuan [2 ]
Hu, Yi [2 ]
Xue, Kouying [2 ]
Gou, Yanjie [2 ]
Chen, Xi [2 ]
Chen, Huajun [1 ]
机构
[1] Zhejiang Univ, Hangzhou, Zhejiang, Peoples R China
[2] Tencent, Shenzhen, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge editing for large language models can offer an efficient solution to alter a model's behavior without negatively impacting the overall performance. However, the current approaches encounter issues with limited generalizability across tasks, necessitating one distinct editor for each task, significantly hindering the broader applications. To address this, we take the first step to analyze the multi-task generalization issue in knowledge editing. Specifically, we develop an instruction-based editing technique, termed InstructEdit, which facilitates the editor's adaptation to various task performances simultaneously using simple instructions. With only one unified editor for each LLM, we empirically demonstrate that InstructEdit can improve the editor's control, leading to an average 14.86% increase in Reliability in multi-task editing setting. Furthermore, experiments involving holdout unseen task illustrate that InstructEdit consistently surpass previous strong baselines. To further investigate the underlying mechanisms of instruction-based knowledge editing, we analyze the principal components of the editing gradient directions, which unveils that instructions can help control optimization direction with stronger OOD generalization.
引用
收藏
页码:6633 / 6641
页数:9
相关论文
共 50 条
  • [1] SmartEdit: Exploring Complex Instruction-based Image Editing with Multimodal Large Language Models
    Huang, Yuzhou
    Xie, Liangbin
    Wang, Xintao
    Yuan, Ziyang
    Cun, Xiaodong
    Ge, Yixiao
    Zhou, Jiantao
    Dong, Chao
    Huang, Rui
    Zhang, Ruimao
    Shan, Ying
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 8362 - 8371
  • [2] Knowledge Editing for Large Language Models: A Survey
    Wang, Song
    Zhu, Yaochen
    Liu, Haochen
    Zheng, Zaiyi
    Chen, Chen
    Li, Jundong
    ACM COMPUTING SURVEYS, 2025, 57 (03)
  • [3] INSTRUCTION-BASED KNOWLEDGE ACQUISITION AND MODIFICATION - THE OPERATIONAL KNOWLEDGE FOR A FUNCTIONAL, VISUAL PROGRAMMING LANGUAGE
    SCHRODER, O
    FRANK, KD
    KOHNERT, K
    MOBUS, C
    RAUTERBERG, M
    COMPUTERS IN HUMAN BEHAVIOR, 1990, 6 (01) : 31 - 49
  • [4] Detoxifying Large Language Models via Knowledge Editing
    Wang, Mengru
    Zhang, Ningyu
    Xu, Ziwen
    Xi, Zekun
    Deng, Shumin
    Yao, Yunzhi
    Zhang, Qishen
    Yang, Linyi
    Wang, Jindong
    Chen, Huajun
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 3093 - 3118
  • [5] Unmasking large language models by means of OpenAI GPT-4 and Google AI: A deep instruction-based analysis
    Zahid, Idrees A.
    Joudar, Shahad Sabbar
    Albahri, A. S.
    Albahri, O. S.
    Alamoodi, A. H.
    Santamaria, Jose
    Alzubaidi, Laith
    INTELLIGENT SYSTEMS WITH APPLICATIONS, 2024, 23
  • [6] Cross-Lingual Knowledge Editing in Large Language Models
    Wang, Jiaan
    Liang, Yunlong
    Sun, Zengkui
    Cao, Yuxuan
    Xu, Jiarong
    Meng, Fandong
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 11676 - 11686
  • [7] Knowledge Editing of Large Language Models Unconstrained by Word Order
    Ishigaki, Ryoma
    Suzuki, Jundai
    Shuzo, Masaki
    Maeda, Eisaku
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 4: STUDENT RESEARCH WORKSHOP, 2024, : 177 - 187
  • [8] DistillMIKE: Editing Distillation of Massive In-Context Knowledge Editing in Large Language Models
    Qiao, Shanbao
    Liu, Xuebing
    Na, Seung-Hoon
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 7639 - 7654
  • [9] Instruction-based communications
    Tschudin, C
    MOBILE OBJECT SYSTEMS: TOWARDS THE PROGRAMMABLE INTERNET, 1997, 1222 : 67 - 90
  • [10] Editing Factual Knowledge in Language Models
    De Cao, Nicola
    Aziz, Wilker
    Titov, Ivan
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 6491 - 6506