Medical large language models are susceptible to targeted misinformation attacks

被引:3
|
作者
Han, Tianyu [1 ]
Nebelung, Sven [1 ]
Khader, Firas [1 ]
Wang, Tianci [1 ]
Mueller-Franzes, Gustav [1 ]
Kuhl, Christiane [1 ]
Foersch, Sebastian [2 ]
Kleesiek, Jens [3 ]
Haarburger, Christoph [4 ]
Bressem, Keno K. [5 ,6 ,7 ,8 ]
Kather, Jakob Nikolas [9 ,10 ,11 ]
Truhn, Daniel [1 ]
机构
[1] Univ Hosp Aachen, Dept Diag & Intervent Radiol, Aachen, Germany
[2] Univ Med Ctr Johannes Gutenberg, Inst Pathol, Mainz, Germany
[3] Univ Med Essen, Inst AI Med, Essen, Germany
[4] Ocumeda GmbH, Munich, Germany
[5] Charite Univ Med Berlin, Dept Radiol, Berlin, Germany
[6] Free Univ Berlin, Berlin, Germany
[7] Humboldt Univ, Berlin, Germany
[8] Charite Univ Med Berlin, Berlin Inst Hlth, Berlin, Germany
[9] Tech Univ Dresden, Else Kroener Fresenius Ctr Digital Hlth EKFZ, Dresden, Germany
[10] Univ Hosp Dresden, Dept Med 1, Dresden, Germany
[11] Univ Hosp Heidelberg, Natl Ctr Tumor Dis NCT, Med Oncol, Heidelberg, Germany
来源
NPJ DIGITAL MEDICINE | 2024年 / 7卷 / 01期
基金
美国国家卫生研究院; 欧洲研究理事会;
关键词
All Open Access; Gold;
D O I
10.1038/s41746-024-01282-7
中图分类号
R19 [保健组织与事业(卫生事业管理)];
学科分类号
摘要
Large language models (LLMs) have broad medical knowledge and can reason about medical information across many domains, holding promising potential for diverse medical applications in the near future. In this study, we demonstrate a concerning vulnerability of LLMs in medicine. Through targeted manipulation of just 1.1% of the weights of the LLM, we can deliberately inject incorrect biomedical facts. The erroneous information is then propagated in the model's output while maintaining performance on other biomedical tasks. We validate our findings in a set of 1025 incorrect biomedical facts. This peculiar susceptibility raises serious security and trustworthiness concerns for the application of LLMs in healthcare settings. It accentuates the need for robust protective measures, thorough verification mechanisms, and stringent management of access to these models, ensuring their reliable and safe use in medical practice.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] FramedTruth: A Frame-Based Model Utilising Large Language Models for Misinformation Detection
    Wang, Guan
    Frederick, Rebecca
    Haghighi, Boshra Talebi
    Wong, B. L. William
    Rupar, Verica
    Li, Weihua
    Bai, Quan
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS, PT I, ACIIDS 2024, 2024, 14795 : 135 - 146
  • [22] Bilingual witnesses are more susceptible to the misinformation effect in their less proficient language
    Calvillo, Dustin P.
    Mills, Nicole, V
    CURRENT PSYCHOLOGY, 2020, 39 (02) : 673 - 680
  • [23] Bilingual witnesses are more susceptible to the misinformation effect in their less proficient language
    Dustin P. Calvillo
    Nicole V. Mills
    Current Psychology, 2020, 39 : 673 - 680
  • [24] Data Stealing Attacks against Large Language Models via Backdooring
    He, Jiaming
    Hou, Guanyu
    Jia, Xinyue
    Chen, Yangyang
    Liao, Wenqi
    Zhou, Yinhang
    Zhou, Rang
    ELECTRONICS, 2024, 13 (14)
  • [25] JailbreakLens: Visual Analysis of Jailbreak Attacks Against Large Language Models
    Feng, Yingchaojie
    Chen, Zhizhang
    Kang, Zhining
    Wang, Sijia
    Zhu, Minfeng
    Zhang, Wei
    Chen, Wei
    arXiv,
  • [26] HARNESSING TASK OVERLOAD FOR SCALABLE JAILBREAK ATTACKS ON LARGE LANGUAGE MODELS
    Dong, Yiting
    Shen, Guobin
    Zhao, Dongcheng
    He, Xiang
    Zeng, Yi
    arXiv,
  • [27] Adversarial Attacks and Defenses in Large Language Models: Old and New Threats
    Schwinn, Leo
    Dobre, David
    Guennemann, Stephan
    Gidel, Gauthier
    PROCEEDINGS ON I CAN'T BELIEVE IT'S NOT BETTER: FAILURE MODES IN THE AGE OF FOUNDATION MODELS AT NEURIPS 2023 WORKSHOPS, 2023, 239 : 103 - 117
  • [28] A paradigm shift?-On the ethics of medical large language models
    Grote, Thomas
    Berens, Philipp
    BIOETHICS, 2024, 38 (05) : 383 - 390
  • [29] Conformal Prediction and Large Language Models for Medical Coding
    Snyder, Christopher
    Brodsky, Victor
    AMERICAN JOURNAL OF CLINICAL PATHOLOGY, 2024, 162
  • [30] Teaching the Limitations of Large Language Models in Medical School
    Gunawardene, Araliya N.
    Schmuter, Gabriella
    JOURNAL OF SURGICAL EDUCATION, 2024, 81 (05) : 625 - 625