Transfer learning enables predictions in network biology

被引:223
|
作者
Theodoris, Christina V. [1 ,2 ,3 ,4 ]
Xiao, Ling [2 ,5 ]
Chopra, Anant [6 ]
Chaffin, Mark D. [2 ]
Al Sayed, Zeina R. [2 ]
Hill, Matthew C. [2 ,5 ]
Mantineo, Helene [2 ,5 ]
Brydon, Elizabeth M. [6 ]
Zeng, Zexian [1 ,7 ]
Liu, X. Shirley [1 ,7 ,8 ]
Ellinor, Patrick T. [2 ,5 ]
机构
[1] Dana Farber Canc Inst, Dept Data Sci, Boston, MA 02215 USA
[2] Broad Inst MIT & Harvard, Cardiovasc Dis Initiat & Precis Cardiol Lab, Cambridge, MA 02142 USA
[3] Boston Childrens Hosp, Div Genet & Genom, Boston, MA 02115 USA
[4] Harvard Med Sch, Genet Training Program, Boston, MA 02115 USA
[5] Massachusetts Gen Hosp, Cardiovasc Res Ctr, Boston, MA 02114 USA
[6] Bayer US LLC, Precis Cardiol Lab, Cambridge, MA USA
[7] Harvard TH Chan Sch Publ Hlth, Dept Biostat, Boston, MA USA
[8] Dana Farber Canc Inst, Ctr Funct Canc Epigenet, Boston, MA USA
基金
美国国家卫生研究院;
关键词
SINGLE-CELL TRANSCRIPTOME; IN-VITRO; DIFFERENTIATION; MUTATIONS; GENES; HETEROGENEITY; TRAJECTORIES; LANDSCAPE; ORGANOIDS; SUBSETS;
D O I
10.1038/s41586-023-06139-9
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Mapping gene networks requires large amounts of transcriptomic data to learn the connections between genes, which impedes discoveries in settings with limited data, including rare diseases and diseases affecting clinically inaccessible tissues. Recently, transfer learning has revolutionized fields such as natural language understanding1,2 and computer vision3 by leveraging deep learning models pretrained on large-scale general datasets that can then be fine-tuned towards a vast array of downstream tasks with limited task-specific data. Here, we developed a context-aware, attention-based deep learning model, Geneformer, pretrained on a large-scale corpus of about 30 million single-cell transcriptomes to enable context-specific predictions in settings with limited data in network biology. During pretraining, Geneformer gained a fundamental understanding of network dynamics, encoding network hierarchy in the attention weights of the model in a completely self-supervised manner. Fine-tuning towards a diverse panel of downstream tasks relevant to chromatin and network dynamics using limited task-specific data demonstrated that Geneformer consistently boosted predictive accuracy. Applied to disease modelling with limited patient data, Geneformer identified candidate therapeutic targets for cardiomyopathy. Overall, Geneformer represents a pretrained deep learning model from which fine-tuning towards a broad range of downstream applications can be pursued to accelerate discovery of key network regulators and candidate therapeutic targets.
引用
收藏
页码:616 / 624
页数:32
相关论文
共 50 条
  • [31] EMUNE: Architecture for Mobile Data Transfer Scheduling with Network Availability Predictions
    Upendra Rathnayake
    Henrik Petander
    Maximilian Ott
    Aruna Seneviratne
    Mobile Networks and Applications, 2012, 17 : 216 - 233
  • [32] A transfer learning approach for network modeling
    Huang, Shuai
    Li, Jing
    Chen, Kewei
    Wu, Teresa
    Ye, Jieping
    Wu, Xia
    Yao, Li
    IIE TRANSACTIONS, 2012, 44 (11) : 915 - 931
  • [33] NETWORK BIOLOGY Illuminating the dark side of machine learning
    Burgess, Darren J.
    NATURE REVIEWS GENETICS, 2019, 20 (07) : 374 - 375
  • [34] *Rethinking Data Center Networks: Machine Learning Enables Network Intelligence
    Li B.
    Wang T.
    Yang P.
    Chen M.
    Hamdi M.
    Journal of Communications and Information Networks, 2022, 7 (02) : 157 - 169
  • [35] Learning Predictions for Algorithms with Predictions
    Khodak, Mikhail
    Balcan, Maria-Florina
    Talwalkar, Ameet
    Vassilvitskii, Sergei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [36] Developed Models Based on Transfer Learning for Improving Fake News Predictions
    Wotaifi, Tahseen A.
    Dhannoon, Ban N.
    JOURNAL OF UNIVERSAL COMPUTER SCIENCE, 2023, 29 (05) : 491 - 507
  • [37] A Transfer Learning Approach for Microstructure Reconstruction and Structure-property Predictions
    Li, Xiaolin
    Zhang, Yichi
    Zhao, He
    Burkhart, Craig
    Brinson, L. Catherine
    Chen, Wei
    SCIENTIFIC REPORTS, 2018, 8
  • [38] Transfer Learning Study of Motion Transformer-based Trajectory Predictions
    Ullrich, Lars
    McMaster, Alex
    Graichen, Knut
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 110 - 117
  • [39] A Transfer Learning Approach for Microstructure Reconstruction and Structure-property Predictions
    Xiaolin Li
    Yichi Zhang
    He Zhao
    Craig Burkhart
    L. Catherine Brinson
    Wei Chen
    Scientific Reports, 8
  • [40] Improving Across Dataset Brain Age Predictions Using Transfer Learning
    Dular, Lara
    Spiclin, Ziga
    PREDICTIVE INTELLIGENCE IN MEDICINE, PRIME 2021, 2021, 12928 : 243 - 254