In-context learning enables multimodal large language models to classify cancer pathology images

被引:0
|
作者
Dyke Ferber [1 ]
Georg Wölflein [2 ]
Isabella C. Wiest [3 ]
Marta Ligero [4 ]
Srividhya Sainath [3 ]
Narmin Ghaffari Laleh [5 ]
Omar S. M. El Nahhas [3 ]
Gustav Müller-Franzes [3 ]
Dirk Jäger [3 ]
Daniel Truhn [3 ]
Jakob Nikolas Kather [6 ]
机构
[1] Heidelberg University Hospital,National Center for Tumor Diseases (NCT)
[2] Heidelberg University Hospital,Department of Medical Oncology
[3] Technical University Dresden,Else Kroener Fresenius Center for Digital Health
[4] University of St Andrews,School of Computer Science
[5] Heidelberg University,Department of Medicine II, Medical Faculty Mannheim
[6] University Hospital Aachen,Department of Diagnostic and Interventional Radiology
[7] University Hospital Dresden,Department of Medicine I
关键词
D O I
10.1038/s41467-024-51465-9
中图分类号
学科分类号
摘要
Medical image classification requires labeled, task-specific datasets which are used to train deep learning networks de novo, or to fine-tune foundation models. However, this process is computationally and technically demanding. In language processing, in-context learning provides an alternative, where models learn from within prompts, bypassing the need for parameter updates. Yet, in-context learning remains underexplored in medical image analysis. Here, we systematically evaluate the model Generative Pretrained Transformer 4 with Vision capabilities (GPT-4V) on cancer image processing with in-context learning on three cancer histopathology tasks of high importance: Classification of tissue subtypes in colorectal cancer, colon polyp subtyping and breast tumor detection in lymph node sections. Our results show that in-context learning is sufficient to match or even outperform specialized neural networks trained for particular tasks, while only requiring a minimal number of samples. In summary, this study demonstrates that large vision language models trained on non-domain specific data can be applied out-of-the box to solve medical image-processing tasks in histopathology. This democratizes access of generalist AI models to medical experts without technical background especially for areas where annotated data is scarce.
引用
收藏
相关论文
共 50 条
  • [21] Diluie: constructing diverse demonstrations of in-context learning with large language model for unified information extraction
    Guo Q.
    Guo Y.
    Zhao J.
    Neural Computing and Applications, 2024, 36 (22) : 13491 - 13512
  • [22] Capability of multimodal large language models to interpret pediatric radiological images
    Reith, Thomas P.
    D'Alessandro, Donna M.
    D'Alessandro, Michael P.
    PEDIATRIC RADIOLOGY, 2024, : 1729 - 1737
  • [23] Multimodal large language models for inclusive collaboration learning tasks
    Lewis, Armanda
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES: PROCEEDINGS OF THE STUDENT RESEARCH WORKSHOP, 2022, : 202 - 210
  • [24] Meta-in-context learning in large language models
    Coda-Forno, Julian
    Binz, Marcel
    Akata, Zeynep
    Botvinick, Matthew
    Wang, Jane X.
    Schulz, Eric
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [25] Meta-learning via Language Model In-context Tuning
    Chen, Yanda
    Zhong, Ruiqi
    Zha, Sheng
    Karypis, George
    He, He
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 719 - 730
  • [26] Cultural Understanding Using In-context Learning and Masked Language Modeling
    Qian, Ming
    Newton, Charles
    Qian, Davis
    HCI INTERNATIONAL 2021 - LATE BREAKING PAPERS: MULTIMODALITY, EXTENDED REALITY, AND ARTIFICIAL INTELLIGENCE, 2021, 13095 : 500 - 508
  • [27] Generating Images with Multimodal Language Models
    Koh, Jing Yu
    Fried, Daniel
    Salakhutdinov, Ruslan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [28] Investigating large language model (LLM) performance using in-context learning (ICL) for interpretation of ESMO and NCCN guidelines for lung cancer
    Iivanainen, Sanna
    Lagus, Jarkko
    Viertolahti, Henri
    Sippola, Lauri
    Koivunen, Jussi
    JOURNAL OF CLINICAL ONCOLOGY, 2024, 42 (16)
  • [29] Stabilized In-Context Learning with Pre-trained Language Models for Few Shot Dialogue State Tracking
    Chen, Derek
    Qian, Kun
    Yu, Zhou
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1551 - 1564
  • [30] Jailbreak and Guard Aligned Language Models with Only Few In-Context Demonstrations
    Wei, Zeming
    Wang, Yifei
    Li, Ang
    Mo, Yichuan
    Wang, Yisen
    arXiv, 2023,