CancerGPT for few shot drug pair synergy prediction using large pretrained language models

被引:0
|
作者
Tianhao Li
Sandesh Shetty
Advaith Kamath
Ajay Jaiswal
Xiaoqian Jiang
Ying Ding
Yejin Kim
机构
[1] University of Texas at Austin,School of Information
[2] University of Massachusetts Amherst,Manning College of Information and Computer Sciences
[3] University of Texas at Austin,Department of Chemical Engineering
[4] University of Texas Health Science Center at Houston,McWilliams School of Biomedical Informatics
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Large language models (LLMs) have been shown to have significant potential in few-shot learning across various fields, even with minimal training data. However, their ability to generalize to unseen tasks in more complex fields, such as biology and medicine has yet to be fully evaluated. LLMs can offer a promising alternative approach for biological inference, particularly in cases where structured data and sample size are limited, by extracting prior knowledge from text corpora. Here we report our proposed few-shot learning approach, which uses LLMs to predict the synergy of drug pairs in rare tissues that lack structured data and features. Our experiments, which involved seven rare tissues from different cancer types, demonstrate that the LLM-based prediction model achieves significant accuracy with very few or zero samples. Our proposed model, the CancerGPT (with ~ 124M parameters), is comparable to the larger fine-tuned GPT-3 model (with ~ 175B parameters). Our research contributes to tackling drug pair synergy prediction in rare tissues with limited data, and also advancing the use of LLMs for biological and medical inference tasks.
引用
收藏
相关论文
共 50 条
  • [31] The Goldilocks paradigm: comparing classical machine learning, large language models, and few-shot learning for drug discovery applications
    Snyder, Scott H.
    Vignaux, Patricia A.
    Ozalp, Mustafa Kemal
    Gerlach, Jacob
    Puhl, Ana C.
    Lane, Thomas R.
    Corbett, John
    Urbina, Fabio
    Ekins, Sean
    COMMUNICATIONS CHEMISTRY, 2024, 7 (01):
  • [32] True Few-Shot Learning with Language Models
    Perez, Ethan
    Kiela, Douwe
    Cho, Kyunghyun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [33] An Empirical Evaluation of the Zero-Shot, Few-Shot, and Traditional Fine-Tuning Based Pretrained Language Models for Sentiment Analysis in Software Engineering
    Shafikuzzaman, Md
    Islam, Md Rakibul
    Rolli, Alex C.
    Akhter, Sharmin
    Seliya, Naeem
    IEEE ACCESS, 2024, 12 : 109714 - 109734
  • [34] Normalized difference vegetation index prediction using reservoir computing and pretrained language models
    Olamofe, John
    Ray, Ram
    Dong, Xishuang
    Qian, Lijun
    ARTIFICIAL INTELLIGENCE IN AGRICULTURE, 2025, 15 (01): : 116 - 129
  • [35] PivotFEC: Enhancing Few-shot Factual Error Correction with a Pivot Task Approach using Large Language Models
    He, Xingwei
    Jini, A-Long
    Ma, Jun
    Yuan, Yuan
    Yin, Siu Ming
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 9960 - 9976
  • [36] Using Large Pretrained Language Models for Answering User Queries from Product Specifications
    Roy, Kalyani
    Shah, Smit
    Pai, Nithish
    Ramtej, Jaidam
    Nadkarn, Prajit Prashant
    Banerjee, Jyotirmoy
    Goyal, Pawan
    Kumar, Surender
    WORKSHOP ON E-COMMERCE AND NLP (ECNLP 3), 2020, : 35 - 39
  • [37] Empowering Few-Shot Recommender Systems With Large Language Models-Enhanced Representations
    Wang, Zhoumeng
    IEEE ACCESS, 2024, 12 : 29144 - 29153
  • [38] Application of Pretrained Large Language Models in Embodied Artificial Intelligence
    A. K. Kovalev
    A. I. Panov
    Doklady Mathematics, 2022, 106 : S85 - S90
  • [39] Evaluation of Pretrained Large Language Models in Embodied Planning Tasks
    Sarkisyan, Christina
    Korchemnyi, Alexandr
    Kovalev, Alexey K.
    Panov, Aleksandr, I
    ARTIFICIAL GENERAL INTELLIGENCE, AGI 2023, 2023, 13921 : 222 - 232
  • [40] Generalized Planning in PDDL Domains with Pretrained Large Language Models
    Silver, Tom
    Dan, Soham
    Srinivas, Kavitha
    Tenenbaum, Joshua B.
    Kaelbling, Leslie
    Katz, Michael
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 18, 2024, : 20256 - 20264