Fine-tuning of conditional Transformers improves in silico enzyme prediction and generation

被引:0
|
作者
Nicolini, Marco [1 ]
Saitto, Emanuele [1 ]
Jimenez Franco, Ruben Emilio [4 ]
Cavalleri, Emanuele [1 ]
Galeano Alfonso, Aldo Javier [4 ]
Malchiodi, Dario [1 ]
Paccanaro, Alberto [4 ,5 ]
Robinson, Peter N. [2 ,3 ]
Casiraghi, Elena [1 ,2 ]
Valentini, Giorgio [1 ,2 ]
机构
[1] AnacletoLab, Dipartimento di Informatica, Universita degli Studi di Milano, Italy
[2] ELLIS - European Laboratory for Learning and Intelligent Systems, Milan Unit, Italy
[3] Berlin Institute of Health at Charite (BIH), Berlin, Germany
[4] School of Applied Mathematics (EMAp) - FGV, Rio de Janeiro, Brazil
[5] Department of Computer Science, Bioinformatics Centre for Systems and Synthetic Biology, Royal Holloway, University of London, United Kingdom
关键词
Prediction models;
D O I
10.1016/j.csbj.2025.03.037
中图分类号
学科分类号
摘要
We introduce Finenzyme, a Protein Language Model (PLM) that employs a multifaceted learning strategy based on transfer learning from a decoder-based Transformer, conditional learning using specific functional keywords, and fine-tuning for the in silico modeling of enzymes. Our experiments show that Finenzyme significantly enhances generalist PLMs like ProGen for the in silico prediction and generation of enzymes belonging to specific Enzyme Commission (EC) categories. Our in silico experiments demonstrate that Finenzyme generated sequences can diverge from natural ones, while retaining similar predicted tertiary structure, predicted functions and the active sites of their natural counterparts. We show that embedded representations of the generated sequences obtained from the embeddings computed by both Finenzyme and ESMFold closely resemble those of natural ones, thus making them suitable for downstream tasks, including e.g. EC classification. Clustering analysis based on the primary and predicted tertiary structure of sequences reveals that the generated enzymes form clusters that largely overlap with those of natural enzymes. These overall in silico validation experiments indicate that Finenzyme effectively captures the structural and functional properties of target enzymes, and can in perspective support targeted enzyme engineering tasks. © 2025
引用
收藏
页码:1318 / 1334
相关论文
共 50 条
  • [1] Fine-tuning transformers: Vocabulary transfer
    Mosin, Vladislav
    Samenko, Igor
    Kozlovskii, Borislav
    Tikhonov, Alexey
    Yamshchikov, Ivan P.
    ARTIFICIAL INTELLIGENCE, 2023, 317
  • [2] Gradient Sparsification For Masked Fine-Tuning of Transformers
    O'Neill, James
    Dutta, Sourav
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [3] On the Interplay Between Fine-tuning and Composition in Transformers
    Yu, Lang
    Ettinger, Allyson
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 2279 - 2293
  • [4] Multi-resolution Fine-Tuning of Vision Transformers
    Fitzgerald, Kerr
    Law, Meng
    Seah, Jarrel
    Tang, Jennifer
    Matuszewski, Bogdan
    MEDICAL IMAGE UNDERSTANDING AND ANALYSIS, MIUA 2022, 2022, 13413 : 535 - 546
  • [5] MMSFT: Multilingual Multimodal Summarization by Fine-Tuning Transformers
    Phani, Siginamsetty
    Abdul, Ashu
    Krishna Siva Prasad, M.
    Kumar Deva Sarma, Hiren
    IEEE ACCESS, 2024, 12 : 129673 - 129689
  • [6] Fine-tuning Image Transformers using Learnable Memory
    Sandler, Mark
    Zhmoginov, Andrey
    Vladymyrov, Max
    Jackson, Andrew
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12145 - 12154
  • [7] IMHO Fine-Tuning Improves Claim Detection
    Chakrabarty, Tuhin
    Hidey, Christopher
    McKeown, Kathleen
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 558 - 563
  • [8] Dissecting the Allosteric Fine-Tuning of Enzyme Catalysis
    Yao, Xin-Qiu
    Hamelberg, Donald
    JACS AU, 2024, 4 (02): : 837 - 846
  • [9] Breast Cancer Prediction Empowered with Fine-Tuning
    Nasir, Muhammad Umar
    Ghazal, Taher M.
    Khan, Muhammad Adnan
    Zubair, Muhammad
    Rahman, Atta-ur
    Ahmed, Rashad
    Hamadi, Hussam Al
    Yeun, Chan Yeob
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [10] Contrastive Fine-tuning Improves Robustness for Neural Rankers
    Ma, Xiaofei
    dos Santos, Cicero Nogueira
    Arnold, Andrew O.
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 570 - 582