Fine-tuning of conditional Transformers improves in silico enzyme prediction and generation

被引:0
|
作者
Nicolini, Marco [1 ]
Saitto, Emanuele [1 ]
Jimenez Franco, Ruben Emilio [4 ]
Cavalleri, Emanuele [1 ]
Galeano Alfonso, Aldo Javier [4 ]
Malchiodi, Dario [1 ]
Paccanaro, Alberto [4 ,5 ]
Robinson, Peter N. [2 ,3 ]
Casiraghi, Elena [1 ,2 ]
Valentini, Giorgio [1 ,2 ]
机构
[1] AnacletoLab, Dipartimento di Informatica, Universita degli Studi di Milano, Italy
[2] ELLIS - European Laboratory for Learning and Intelligent Systems, Milan Unit, Italy
[3] Berlin Institute of Health at Charite (BIH), Berlin, Germany
[4] School of Applied Mathematics (EMAp) - FGV, Rio de Janeiro, Brazil
[5] Department of Computer Science, Bioinformatics Centre for Systems and Synthetic Biology, Royal Holloway, University of London, United Kingdom
关键词
Prediction models;
D O I
10.1016/j.csbj.2025.03.037
中图分类号
学科分类号
摘要
We introduce Finenzyme, a Protein Language Model (PLM) that employs a multifaceted learning strategy based on transfer learning from a decoder-based Transformer, conditional learning using specific functional keywords, and fine-tuning for the in silico modeling of enzymes. Our experiments show that Finenzyme significantly enhances generalist PLMs like ProGen for the in silico prediction and generation of enzymes belonging to specific Enzyme Commission (EC) categories. Our in silico experiments demonstrate that Finenzyme generated sequences can diverge from natural ones, while retaining similar predicted tertiary structure, predicted functions and the active sites of their natural counterparts. We show that embedded representations of the generated sequences obtained from the embeddings computed by both Finenzyme and ESMFold closely resemble those of natural ones, thus making them suitable for downstream tasks, including e.g. EC classification. Clustering analysis based on the primary and predicted tertiary structure of sequences reveals that the generated enzymes form clusters that largely overlap with those of natural enzymes. These overall in silico validation experiments indicate that Finenzyme effectively captures the structural and functional properties of target enzymes, and can in perspective support targeted enzyme engineering tasks. © 2025
引用
收藏
页码:1318 / 1334
相关论文
共 50 条
  • [21] Patent claim generation by fine-tuning OpenAI GPT-2
    Lee, Jieh-Sheng
    Hsiang, Jieh
    WORLD PATENT INFORMATION, 2020, 62
  • [22] Goal-directed molecule generation with fine-tuning by policy gradient
    Sha, Chunli
    Zhu, Fei
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 246
  • [23] Fine-tuning Pipeline for Hand Image Generation Using Diffusion Model
    Bai, Bingyuan
    Xie, Haoran
    Miyata, Kazunori
    2024 NICOGRAPH INTERNATIONAL, NICOINT 2024, 2024, : 58 - 63
  • [24] Goal-directed molecule generation with fine-tuning by policy gradient
    Sha, Chunli
    Zhu, Fei
    Expert Systems with Applications, 2024, 246
  • [25] Surveil and Prediction of Pandemic Disease by Fine-Tuning Hyperparameters in Deep Learning
    Shanthini, A.
    Vinodhini, G.
    INTERNATIONAL JOURNAL OF UNCERTAINTY FUZZINESS AND KNOWLEDGE-BASED SYSTEMS, 2021, 29 (06) : 855 - 873
  • [26] Prediction of Author's Profile Basing on Fine-Tuning BERT Model
    Bsir B.
    Khoufi N.
    Zrigui M.
    Informatica (Slovenia), 2024, 48 (01): : 69 - 78
  • [27] From pixels to sentiment: Fine-tuning CNNs for visual sentiment prediction
    Campos, Victor
    Jou, Brendan
    Giro-i-Nieto, Xavier
    IMAGE AND VISION COMPUTING, 2017, 65 : 15 - 22
  • [28] A Generative Approach for Script Event Prediction via Contrastive Fine-Tuning
    Zhu, Fangqi
    Gao, Jun
    Yu, Changlong
    Wang, Wei
    Xu, Chen
    Mu, Xin
    Yang, Min
    Xu, Ruifeng
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 14056 - 14064
  • [29] Stage-wise Fine-tuning for Graph-to-Text Generation
    Wang, Qingyun
    Yavuz, Semih
    Lin, Xi Victoria
    Ji, Heng
    Rajani, Nazneen Fatema
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING: PROCEEDINGS OF THE STUDENT RESEARCH WORKSHOP, 2021, : 16 - 22
  • [30] Fine-tuning gibberellin improves rice alkali-thermal tolerance and yield
    Guo, Shuang-Qin
    Chen, Ya-Xin
    Ju, Ya-Lin
    Pan, Chen-Yang
    Shan, Jun-Xiang
    Ye, Wang-Wei
    Dong, Nai-Qian
    Kan, Yi
    Yang, Yi-Bing
    Zhao, Huai-Yu
    Yu, Hong-Xiao
    Lu, Zi-Qi
    Lei, Jie-Jie
    Liao, Ben
    Mu, Xiao-Rui
    Cao, Ying-Jie
    Guo, Liangxing
    Gao, Jin
    Zhou, Ji-Fu
    Yang, Kai-Yang
    Lin, Hong-Xuan
    Lin, Youshun
    NATURE, 2025, 639 (8053) : 162 - 171