Knowledge Editing of Large Language Models Unconstrained by Word Order

被引:0
|
作者
Ishigaki, Ryoma [1 ]
Suzuki, Jundai [1 ]
Shuzo, Masaki [1 ]
Maeda, Eisaku [1 ]
机构
[1] Tokyo Denki Univ, Tokyo, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large Language Models (LLMs) are considered to have potentially extensive knowledge, but because their internal processing is black-boxed, it has been difficult to directly edit the knowledge held by the LLMs themselves. To address this issue, a method called local modification-based knowledge editing has been developed. This method identifies the knowledge neurons that encode the target knowledge and adjusts the parameters associated with these neurons to update the knowledge. Knowledge neurons are identified by masking the o part from sentences representing relational triplets (s, r, o), having the LLM predict the masked part, and observing the LLM's activation during the prediction. When the architecture is decoder-based, the predicted o needs to be located at the end of the sentence. Previous local modification-based knowledge editing methods for decoder-based models have assumed SVO languages and faced challenges when applied to SOV languages such as Japanese. In this study, we propose a knowledge editing method that eliminates the need for word order constraints by converting the input for identifying knowledge neurons into a question where o is the answer. We conducted validation experiments on 500 examples and confirmed that the proposed method is effective for Japanese, a non-SVO language. We also applied this method to English, an SVO language, and demonstrated that it outperforms conventional methods.
引用
收藏
页码:177 / 187
页数:11
相关论文
共 50 条
  • [41] Unifying Large Language Models and Knowledge Graphs: A Roadmap
    Pan, Shirui
    Luo, Linhao
    Wang, Yufei
    Chen, Chen
    Wang, Jiapu
    Wu, Xindong
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (07) : 3580 - 3599
  • [42] (sic) UHGEval: Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained Generation
    Liang, Xun
    Song, Shichao
    Niu, Simin
    Li, Zhiyu
    Xiong, Feiyu
    Tang, Bo
    Wang, Yezhaohui
    He, Dawei
    Cheng, Peng
    Wang, Zhonghao
    Deng, Haiying
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 5266 - 5293
  • [43] A word-order database for testing computational models of language acquisition
    Sakas, WG
    41ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE, 2003, : 415 - 422
  • [44] SSS: Editing Factual Knowledge in Language Models towards Semantic Sparse Space
    Wang, Huazheng
    Sun, Haifeng
    Wang, Jingyu
    Qi, Qi
    Xia, Zixuan
    Zhang, Menghao
    Liao, Jianxin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5559 - 5570
  • [45] Reasoning in Large Language Models Through Symbolic Math Word Problems
    Gaur, Vedant
    Saunshi, Nikunj
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5889 - 5903
  • [46] Large Language Models and Multimodal Retrieval for Visual Word Sense Disambiguation
    Kritharoula, Anastasia
    Lymperaiou, Maria
    Stamou, Giorgos
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 13053 - 13077
  • [47] Utility of word embeddings from large language models in medical diagnosis
    Yazdani, Shahram
    Henry, Ronald Claude
    Byrne, Avery
    Henry, Isaac Claude
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2025, 32 (03) : 526 - 534
  • [48] The Use of Large Language Models in Scientific Manuscript Writing and Editing in Interventional Radiology
    Goh, Gerard S.
    Clements, Warren
    Bilhim, Tiago
    Hausegger, Klaus
    CARDIOVASCULAR AND INTERVENTIONAL RADIOLOGY, 2025, 48 (03) : 286 - 287
  • [49] DAFNet: Dynamic Auxiliary Fusion for Sequential Model Editing in Large Language Models
    Zhang, Taolin
    Chen, Qizhou
    Li, Dongyang
    Wang, Chengyu
    He, Xiaofeng
    Huang, Longtao
    Xue, Hui
    Huang, Jun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 1588 - 1602
  • [50] Distilling Script Knowledge from Large Language Models for Constrained Language Planning
    Yuan, Siyu
    Chen, Jiangjie
    Fu, Ziquan
    Ge, Xuyang
    Shah, Soham
    Jankowski, Charles Robert
    Xiao, Yanghua
    Yang, Deqing
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 4303 - 4325