BatmanNet: bi-branch masked graph transformer autoencoder for molecular representation

被引:5
|
作者
Wang, Zhen [3 ]
Feng, Zheng [4 ]
Li, Yanjun [5 ]
Li, Bowen [2 ]
Wang, Yongrui [6 ]
Sha, Chulin [2 ]
He, Min [1 ,2 ,7 ]
Li, Xiaolin [2 ]
机构
[1] Hunan Univ, Coll Elect & Informat Engn, Changsha 410082, Hunan, Peoples R China
[2] Chinese Acad Sci, Hangzhou Inst Med, Hangzhou 310018, Zhejiang, Peoples R China
[3] Hunan Univ, Sch Elect & Informat Engn, Dept Elect Sci & Technol, Changsha, Peoples R China
[4] Univ Florida, Coll Med, Dept Hlth Outcomes & Biomed Informat, Gainesville, FL USA
[5] Univ Florida, Ctr Nat Prod Drug Discovery & Dev, Dept Med Chem, Gainesville, FL USA
[6] Univ Chinese Acad Sci, Hangzhou Inst Adv Study, Sch Mol Med, Beijing, Peoples R China
[7] Hunan Univ, Changsha, Peoples R China
关键词
molecular representation; deep learning; graph neural network; self-supervised learning; GENERATION; PREDICTION; KNOWLEDGE;
D O I
10.1093/bib/bbad400
中图分类号
Q5 [生物化学];
学科分类号
071010 ; 081704 ;
摘要
Although substantial efforts have been made using graph neural networks (GNNs) for artificial intelligence (AI)-driven drug discovery, effective molecular representation learning remains an open challenge, especially in the case of insufficient labeled molecules. Recent studies suggest that big GNN models pre-trained by self-supervised learning on unlabeled datasets enable better transfer performance in downstream molecular property prediction tasks. However, the approaches in these studies require multiple complex self-supervised tasks and large-scale datasets , which are time-consuming, computationally expensive and difficult to pre-train end-to-end. Here, we design a simple yet effective self-supervised strategy to simultaneously learn local and global information about molecules, and further propose a novel bi-branch masked graph transformer autoencoder (BatmanNet) to learn molecular representations. BatmanNet features two tailored complementary and asymmetric graph autoencoders to reconstruct the missing nodes and edges, respectively, from a masked molecular graph. With this design, BatmanNet can effectively capture the underlying structure and semantic information of molecules, thus improving the performance of molecular representation. BatmanNet achieves state-of-the-art results for multiple drug discovery tasks, including molecular properties prediction, drug-drug interaction and drug-target interaction, on 13 benchmark datasets, demonstrating its great potential and superiority in molecular representation learning.
引用
收藏
页数:13
相关论文
共 13 条
  • [1] Bi-Branch Vision Transformer Network for EEG Emotion Recognition
    Lu, Wei
    Tan, Tien-Ping
    Ma, Hua
    IEEE ACCESS, 2023, 11 : 36233 - 36243
  • [2] Advancing Latent Representation Ranking for Masked Graph Autoencoder
    Hu, Yulan
    Chen, Ge
    Ouyang, Sheng
    Yang, Zhirui
    Wan, Junchen
    Zhang, Fuzheng
    Wang, Zhongyuan
    Cao, Zhao
    Wu, Shangquan
    Liu, Yong
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, PT VI, DASFAA 2024, 2024, 14855 : 385 - 394
  • [3] BCTNet: Bi-Branch Cross-Fusion Transformer for Building Footprint Extraction
    Xu, Lele
    Li, Ye
    Xu, Jinzhong
    Zhang, Yue
    Guo, Lili
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [4] A Novel Bi-Branch Graph Convolutional Neural Network for Aspect Level Sentiment Classification
    Chen, Bingliang
    Lu, Guojun
    Xue, Yun
    Cai, Qianhua
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [5] GMAEEG: A Self-Supervised Graph Masked Autoencoder for EEG Representation Learning
    Fu, Zanhao
    Zhu, Huaiyu
    Zhao, Yisheng
    Huan, Ruohong
    Zhang, Yi
    Chen, Shuohui
    Pan, Yun
    IEEE Journal of Biomedical and Health Informatics, 2024, 28 (11): : 6486 - 6497
  • [6] GMAEEG: A Self-Supervised Graph Masked Autoencoder for EEG Representation Learning
    Fu, Zanhao
    Zhu, Huaiyu
    Zhao, Yisheng
    Huan, Ruohong
    Zhang, Yi
    Chen, Shuohui
    Pan, Yun
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (11) : 6486 - 6497
  • [7] Convolution-Enhanced Bi-Branch Adaptive Transformer With Cross-Task Interaction for Food Category and Ingredient Recognition
    Liu, Yuxin
    Min, Weiqing
    Jiang, Shuqiang
    Rui, Yong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 2572 - 2586
  • [8] Pretraining graph transformer for molecular representation with fusion of multimodal information
    Chen, Ruizhe
    Li, Chunyan
    Wang, Longyue
    Liu, Mingquan
    Chen, Shugao
    Yang, Jiahao
    Zeng, Xiangxiang
    INFORMATION FUSION, 2025, 115
  • [9] Yet Another Traffic Classifier: A Masked Autoencoder Based Traffic Transformer with Multi-Level Flow Representation
    Zhao, Ruijie
    Zhan, Mingwei
    Deng, Xianwen
    Wang, Yanhao
    Wang, Yijun
    Gui, Guan
    Xue, Zhi
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 4, 2023, : 5420 - 5427
  • [10] Integrating transformer and autoencoder techniques with spectral graph algorithms for the prediction of scarcely labeled molecular data
    Hayes, Nicole
    Merkurjev, Ekaterina
    Wei, Guo-Wei
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 153