Energy-Motivated Equivariant Pretraining for 3D Molecular Graphs

被引:0
|
作者
Jiao, Rui [1 ,2 ]
Han, Jiaqi [1 ,2 ]
Huang, Wenbing [4 ,5 ]
Rong, Yu [6 ]
Liu, Yang [1 ,2 ,3 ]
机构
[1] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol BNRis, Dept Comp Sci & Technol, Beijing, Peoples R China
[2] Tsinghua Univ, Inst Ind Res AIR, Beijing, Peoples R China
[3] Beijing Acad Artificial Intelligence, Beijing, Peoples R China
[4] Renmin Univ China, Gaoling Sch Artificial Intelligence, Beijing, Peoples R China
[5] Beijing Key Lab Big Data Management & Anal Method, Beijing, Peoples R China
[6] Tencent AI Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pretraining molecular representation models without labels is fundamental to various applications. Conventional methods mainly process 2D molecular graphs and focus solely on 2D tasks, making their pretrained models incapable of characterizing 3D geometry and thus defective for downstream 3D tasks. In this work, we tackle 3D molecular pretraining in a complete and novel sense. In particular, we first propose to adopt an equivariant energy-based model as the backbone for pretraining, which enjoys the merits of fulfilling the symmetry of 3D space. Then we develop a node-level pretraining loss for force prediction, where we further exploit the Riemann-Gaussian distribution to ensure the loss to be E(3)-invariant, enabling more robustness. Moreover, a graph-level noise scale prediction task is also leveraged to further pro-mote the eventual performance. We evaluate our model pretrained from a large-scale 3D dataset GEOM-QM9 on two challenging 3D benchmarks: MD17 and QM9. Experimental results demonstrate the efficacy of our method against current state-of-the-art pretraining approaches, and verify the validity of our design for each proposed component. Code is available at https://github.com/jiaor17/3D-EMGP.
引用
收藏
页码:8096 / 8104
页数:9
相关论文
共 50 条
  • [31] Rethinking Masked-Autoencoder-Based 3D Point Cloud Pretraining
    Cheng, Nuo
    Luo, Chuanyu
    Li, Xinzhe
    Hu, Ruizhi
    Li, Han
    Ma, Sikun
    Ren, Zhong
    Jiang, Haipeng
    Li, Xiaohan
    Lei, Shengguang
    Li, Pu
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 2763 - 2768
  • [32] Learning 3D Semantic Scene Graphs from 3D Indoor Reconstructions
    Wald, Johanna
    Dhamo, Helisa
    Navab, Nassir
    Tombari, Federico
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 3960 - 3969
  • [33] DEVIANT: Depth EquiVarIAnt NeTwork for Monocular 3D Object Detection
    Kumar, Abhinav
    Brazil, Garrick
    Corona, Enrique
    Parchami, Armin
    Liu, Xiaoming
    COMPUTER VISION, ECCV 2022, PT IX, 2022, 13669 : 664 - 683
  • [34] Shadow graphs and 3D texture reconstruction
    Yu, YH
    Chang, JT
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2005, 62 (1-2) : 35 - 60
  • [35] Validating a 3D layout for memorable graphs
    Regan, B
    OZCHI 98 - 1998 AUSTRALASIAN COMPUTER HUMAN INTERACTION CONFERENCE, PROCEEDINGS, 1998, : 334 - 334
  • [36] Transformation-Equivariant 3D Object Detection for Autonomous Driving
    Wu, Hai
    Wen, Chenglu
    Li, Wei
    Li, Xin
    Yang, Ruigang
    Wang, Cheng
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 3, 2023, : 2795 - +
  • [37] Querying 3D Data by Adjacency Graphs
    Bore, Nils
    Jensfelt, Patric
    Folkesson, John
    COMPUTER VISION SYSTEMS (ICVS 2015), 2015, 9163 : 243 - 252
  • [38] 3D Proportional Contact Representations of Graphs
    Alam, Md. Jawaherul
    Kobourov, Stephen G.
    Liotta, Giuseppe
    Pupyrev, Sergey
    Veeramoni, Sankar
    5TH INTERNATIONAL CONFERENCE ON INFORMATION, INTELLIGENCE, SYSTEMS AND APPLICATIONS, IISA 2014, 2014, : 27 - 32
  • [39] Rectangle and box visibility graphs in 3D
    Fekete, SP
    Meijer, H
    INTERNATIONAL JOURNAL OF COMPUTATIONAL GEOMETRY & APPLICATIONS, 1999, 9 (01) : 1 - 27
  • [40] Shadow Graphs and 3D Texture Reconstruction
    Yizhou Yu
    Johnny T. Chang
    International Journal of Computer Vision, 2005, 62 : 35 - 60