Knowledge Editing of Large Language Models Unconstrained by Word Order

被引:0
|
作者
Ishigaki, Ryoma [1 ]
Suzuki, Jundai [1 ]
Shuzo, Masaki [1 ]
Maeda, Eisaku [1 ]
机构
[1] Tokyo Denki Univ, Tokyo, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large Language Models (LLMs) are considered to have potentially extensive knowledge, but because their internal processing is black-boxed, it has been difficult to directly edit the knowledge held by the LLMs themselves. To address this issue, a method called local modification-based knowledge editing has been developed. This method identifies the knowledge neurons that encode the target knowledge and adjusts the parameters associated with these neurons to update the knowledge. Knowledge neurons are identified by masking the o part from sentences representing relational triplets (s, r, o), having the LLM predict the masked part, and observing the LLM's activation during the prediction. When the architecture is decoder-based, the predicted o needs to be located at the end of the sentence. Previous local modification-based knowledge editing methods for decoder-based models have assumed SVO languages and faced challenges when applied to SOV languages such as Japanese. In this study, we propose a knowledge editing method that eliminates the need for word order constraints by converting the input for identifying knowledge neurons into a question where o is the answer. We conducted validation experiments on 500 examples and confirmed that the proposed method is effective for Japanese, a non-SVO language. We also applied this method to English, an SVO language, and demonstrated that it outperforms conventional methods.
引用
收藏
页码:177 / 187
页数:11
相关论文
共 50 条
  • [31] Correction: Leveraging large language models for word sense disambiguation
    Jung H. Yae
    Nolan C. Skelly
    Neil C. Ranly
    Phillip M. LaCasse
    Neural Computing and Applications, 2025, 37 (10) : 7449 - 7450
  • [32] TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space
    Zhang, Shaolei
    Yu, Tian
    Feng, Yang
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 8908 - 8949
  • [33] Benchmarking Biomedical Relation Knowledge in Large Language Models
    Zhang, Fenghui
    Yang, Kuo
    Zhao, Chenqian
    Li, Haixu
    Dong, Xin
    Tian, Haoyu
    Zhou, Xuezhong
    BIOINFORMATICS RESEARCH AND APPLICATIONS, PT II, ISBRA 2024, 2024, 14955 : 482 - 495
  • [34] Updating knowledge in Large Language Models: an Empirical Evaluation
    Marinelli, Alberto Roberto
    Carta, Antonio
    Passaro, Lucia C.
    IEEE CONFERENCE ON EVOLVING AND ADAPTIVE INTELLIGENT SYSTEMS 2024, IEEE EAIS 2024, 2024, : 289 - 296
  • [35] ALCUNA: Large Language Models Meet New Knowledge
    Yin, Xunjian
    Huang, Baizhou
    Wan, Xiaojun
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 1397 - 1414
  • [36] Knowledge of cultural moral norms in large language models
    Ramezani, Aida
    Xu, Yang
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 428 - 446
  • [37] Systematic Assessment of Factual Knowledge in Large Language Models
    Luo, Linhao
    Thuy-Trang Vu
    Phung, Dinh
    Haffari, Gholamreza
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 13272 - 13286
  • [38] Poisoning medical knowledge using large language models
    Yang, Junwei
    Xu, Hanwen
    Mirzoyan, Srbuhi
    Chen, Tong
    Liu, Zixuan
    Liu, Zequn
    Ju, Wei
    Liu, Luchen
    Xiao, Zhiping
    Zhang, Ming
    Wang, Sheng
    NATURE MACHINE INTELLIGENCE, 2024, 6 (10) : 1156 - 1168
  • [39] Knowledge Graph Treatments for Hallucinating Large Language Models
    Collarana, Diego
    Busch, Moritz
    Lange, Christoph
    ERCIM NEWS, 2024, (136): : 35 - 36
  • [40] SKILL: Structured Knowledge Infusion for Large Language Models
    Moiseev, Fedor
    Dong, Zhe
    Alfonseca, Enrique
    Jaggi, Martin
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 1581 - 1588