In-context learning enables multimodal large language models to classify cancer pathology images

被引:0
|
作者
Dyke Ferber [1 ]
Georg Wölflein [2 ]
Isabella C. Wiest [3 ]
Marta Ligero [4 ]
Srividhya Sainath [3 ]
Narmin Ghaffari Laleh [5 ]
Omar S. M. El Nahhas [3 ]
Gustav Müller-Franzes [3 ]
Dirk Jäger [3 ]
Daniel Truhn [3 ]
Jakob Nikolas Kather [6 ]
机构
[1] Heidelberg University Hospital,National Center for Tumor Diseases (NCT)
[2] Heidelberg University Hospital,Department of Medical Oncology
[3] Technical University Dresden,Else Kroener Fresenius Center for Digital Health
[4] University of St Andrews,School of Computer Science
[5] Heidelberg University,Department of Medicine II, Medical Faculty Mannheim
[6] University Hospital Aachen,Department of Diagnostic and Interventional Radiology
[7] University Hospital Dresden,Department of Medicine I
关键词
D O I
10.1038/s41467-024-51465-9
中图分类号
学科分类号
摘要
Medical image classification requires labeled, task-specific datasets which are used to train deep learning networks de novo, or to fine-tune foundation models. However, this process is computationally and technically demanding. In language processing, in-context learning provides an alternative, where models learn from within prompts, bypassing the need for parameter updates. Yet, in-context learning remains underexplored in medical image analysis. Here, we systematically evaluate the model Generative Pretrained Transformer 4 with Vision capabilities (GPT-4V) on cancer image processing with in-context learning on three cancer histopathology tasks of high importance: Classification of tissue subtypes in colorectal cancer, colon polyp subtyping and breast tumor detection in lymph node sections. Our results show that in-context learning is sufficient to match or even outperform specialized neural networks trained for particular tasks, while only requiring a minimal number of samples. In summary, this study demonstrates that large vision language models trained on non-domain specific data can be applied out-of-the box to solve medical image-processing tasks in histopathology. This democratizes access of generalist AI models to medical experts without technical background especially for areas where annotated data is scarce.
引用
收藏
相关论文
共 50 条
  • [1] Adaptive In-Context Learning with Large Language Models for Bundle
    Sun, Zhu
    Feng, Kaidong
    Yang, Jie
    Qu, Xinghua
    Fang, Hui
    Ong, Yew-Soon
    Liu, Wenyuan
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 966 - 976
  • [2] Adapting Large Multimodal Models to Distribution Shifts: The Role of In-Context Learning
    University of New South Wales, Australia
    不详
    不详
    不详
    不详
    不详
    不详
    arXiv,
  • [3] Meta In-Context Learning: Harnessing Large Language Models for Electrical Data Classification
    Zhou, Mi
    Li, Fusheng
    Zhang, Fan
    Zheng, Junhao
    Ma, Qianli
    ENERGIES, 2023, 16 (18)
  • [4] Large Language Models Are Latent Variable Models: Explaining and Finding Good Demonstrations for In-Context Learning
    Wang, Xinyi
    Zhu, Wanrong
    Saxon, Michael
    Steyvers, Mark
    Wang, William Yang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Query-focused Submodular Demonstration Selection for In-context Learning in Large Language Models
    Trust, Paul
    Minghim, Rosane
    2023 31ST IRISH CONFERENCE ON ARTIFICIAL INTELLIGENCE AND COGNITIVE SCIENCE, AICS, 2023,
  • [6] Automatic smart contract comment generation via large language models and in-context learning
    Zhao, Junjie
    Chen, Xiang
    Yang, Guang
    Shen, Yiheng
    INFORMATION AND SOFTWARE TECHNOLOGY, 2024, 168
  • [7] In-Context Impersonation Reveals Large Language Models' Strengths and Biases
    Salewski, Leonard
    Alaniz, Stephan
    Rio-Torto, Isabel
    Schulz, Eric
    Akata, Zeynep
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [8] MetaVL: Transferring In-Context Learning Ability From Language Models to Vision-Language Models
    Monajatipoor, Masoud
    Li, Liunian Harold
    Rouhsedaghat, Mozhdeh
    Yang, Lin F.
    Chang, Kai-Wei
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 495 - 508
  • [9] IN-CONTEXT LANGUAGE LEARNING: ARCHITECTURES AND ALGORITHMS
    Akyürek, Ekin
    Wang, Bailin
    Kim, Yoon
    Andreas, Jacob
    arXiv,
  • [10] In-Context Learning Unlocked for Diffusion Models
    Wang, Zhendong
    Jiang, Yifan
    Lu, Yadong
    Shen, Yelong
    He, Pengcheng
    Chen, Weizhu
    Wang, Zhangyang
    Zhou, Mingyuan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,