Cross-Lingual Knowledge Editing in Large Language Models

被引:0
|
作者
Wang, Jiaan [1 ]
Liang, Yunlong [2 ]
Sun, Zengkui [3 ]
Cao, Yuxuan [4 ]
Xu, Jiarong [1 ]
Meng, Fandong [2 ]
机构
[1] Fudan Univ, Shanghai, Peoples R China
[2] Tencent Inc, Pattern Recognit Ctr, WeChat AI, Shenzhen, Peoples R China
[3] Beijing Jiaotong Univ, Beijing, Peoples R China
[4] Zhejiang Univ, Hangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge editing aims to change language models' performance on several special cases (i.e., editing scope) by infusing the corresponding expected knowledge into them. With the recent advancements in large language models (LLMs), knowledge editing has been shown as a promising technique to adapt LLMs to new knowledge without retraining from scratch. However, most of the previous studies neglect the multi-lingual nature of some main-stream LLMs (e.g., LLaMA, ChatGPT and GPT-4), and typically focus on monolingual scenarios, where LLMs are edited and evaluated in the same language. As a result, it is still unknown the effect of source language editing on a different target language. In this paper, we aim to figure out this cross-lingual effect in knowledge editing. Specifically, we first collect a largescale cross-lingual synthetic dataset by translating ZsRE from English to Chinese. Then, we conduct English editing on various knowledge editing methods covering different paradigms, and evaluate their performance in Chinese, and vice versa. To give deeper analyses of the crosslingual effect, the evaluation includes four aspects, i.e., reliability, generality, locality and portability. Furthermore, we analyze the inconsistent behaviors of the edited models and discuss their specific challenges.
引用
收藏
页码:11676 / 11686
页数:11
相关论文
共 50 条
  • [1] Language Anisotropic Cross-Lingual Model Editing
    Xu, Yang
    Hou, Yutai
    Che, Wanxiang
    Zhang, Min
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5554 - 5569
  • [2] Cross-Lingual Consistency of Factual Knowledge in Multilingual Language Models
    Qi, Jirui
    Fernandez, Raquel
    Bisazza, Arianna
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 10650 - 10666
  • [3] Steering Large Language Models for Cross-lingual Information Retrieval
    Guo, Ping
    Ren, Yubing
    Hu, Yue
    Cao, Yanan
    Li, Yunpeng
    Huang, Heyan
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 585 - 596
  • [4] Cross-lingual transfer of knowledge in distributional language models: Experiments in Hungarian
    Novak, Attila
    Novak, Borbala
    ACTA LINGUISTICA ACADEMICA, 2022, 69 (04): : 405 - 449
  • [5] Analyzing the Evaluation of Cross-Lingual Knowledge Transfer in Multilingual Language Models
    Rajaee, Sara
    Monz, Christof
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 2895 - 2914
  • [6] Exploring Cross-lingual Textual Style Transfer with Large Multilingual Language Models
    Moskovskiy, Daniil
    Dementieva, Daryna
    Panchenko, Alexander
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): STUDENT RESEARCH WORKSHOP, 2022, : 346 - 354
  • [7] cViL: Cross-Lingual Training of Vision-Language Models using Knowledge Distillation
    Gupta, Kshitij
    Gautam, Devansh
    Mamidi, Radhika
    Proceedings - International Conference on Pattern Recognition, 2022, 2022-August : 1734 - 1741
  • [8] cViL: Cross-Lingual Training of Vision-Language Models using Knowledge Distillation
    Gupta, Kshitij
    Gautam, Devansh
    Mamidi, Radhika
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 1734 - 1741
  • [9] Cross-lingual projection for class-based language models
    Gfeller, Beat
    Schogol, Vlad
    Hall, Keith
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2016), VOL 2, 2016, : 83 - 88
  • [10] Mass-Editing Memory with Attention in Transformers: A cross-lingual exploration of knowledge
    Tamayo, Daniel
    Gonzalez-Agirre, Aitor
    Hernando, Javier
    Villegas, Marta
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5831 - 5847