Relative molecule self-attention transformer

被引:2
|
作者
Maziarka, Lukasz [1 ]
Majchrowski, Dawid [2 ]
Danel, Tomasz [1 ]
Gainski, Piotr [1 ,3 ]
Tabor, Jacek [1 ]
Podolak, Igor [1 ]
Morkisz, Pawel [2 ]
Jastrzebski, Stanislaw [4 ]
机构
[1] Jagiellonian Univ, Fac Math & Comp Sci, Lojasiewicza 6, PL-30348 Krakow, Poland
[2] NVIDIA, 2788 San Tomas Expy, Santa Clara, CA 95051 USA
[3] Ardigen, Podole 76, PL-30394 Krakow, Poland
[4] Molecule one, Al Jerozolimskie 96, PL-00807 Warsaw, Poland
关键词
Molecular property prediction; Molecular self-attention; Neural networks pre-training; MACHINE LEARNING-METHODS;
D O I
10.1186/s13321-023-00789-7
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
The prediction of molecular properties is a crucial aspect in drug discovery that can save a lot of money and time during the drug design process. The use of machine learning methods to predict molecular properties has become increasingly popular in recent years. Despite advancements in the field, several challenges remain that need to be addressed, like finding an optimal pre-training procedure to improve performance on small datasets, which are common in drug discovery. In our paper, we tackle these problems by introducing Relative Molecule Self-Attention Transformer for molecular representation learning. It is a novel architecture that uses relative self-attention and 3D molecular representation to capture the interactions between atoms and bonds that enrich the backbone model with domain-specific inductive biases. Furthermore, our two-step pretraining procedure allows us to tune only a few hyperparameter values to achieve good performance comparable with state-of-the-art models on a wide selection of downstream tasks. A novel graph transformer architecture for molecular property prediction is introduced. The task-agnostic methodology for pre-training this model is presented, improving target task performance with minimal hyperparameter tuning. A rigorous exploration of the design space for the self-attention layer is conducted to identify the optimal architecture.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Attention Guided CAM: Visual Explanations of Vision Transformer Guided by Self-Attention
    Leem, Saebom
    Seo, Hyunseok
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 4, 2024, : 2956 - 2964
  • [22] Self-Attention Attribution: Interpreting Information Interactions Inside Transformer
    Hao, Yaru
    Dong, Li
    Wei, Furu
    Xu, Ke
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 12963 - 12971
  • [23] ET: Re -Thinking Self-Attention for Transformer Models on GPUs
    Chen, Shiyang
    Huang, Shaoyi
    Pandey, Santosh
    Li, Bingbing
    Gao, Guang R.
    Zheng, Long
    Ding, Caiwen
    Liu, Hang
    [J]. SC21: INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2021,
  • [24] Nucleic Transformer: Classifying DNA Sequences with Self-Attention and Convolutions
    He, Shujun
    Gao, Baizhen
    Sabnis, Rushant
    Sun, Qing
    [J]. ACS SYNTHETIC BIOLOGY, 2023, 12 (11): : 3205 - 3214
  • [25] Additional Self-Attention Transformer With Adapter for Thick Haze Removal
    Cai, Zhenyang
    Ning, Jin
    Ding, Zhiheng
    Duo, Bin
    [J]. IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [26] Lightweight Vision Transformer with Spatial and Channel Enhanced Self-Attention
    Zheng, Jiahao
    Yang, Longqi
    Li, Yiying
    Yang, Ke
    Wang, Zhiyuan
    Zhou, Jun
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 1484 - 1488
  • [27] Spectral Superresolution Using Transformer with Convolutional Spectral Self-Attention
    Liao, Xiaomei
    He, Lirong
    Mao, Jiayou
    Xu, Meng
    [J]. REMOTE SENSING, 2024, 16 (10)
  • [28] ENHANCING TONGUE REGION SEGMENTATION THROUGH SELF-ATTENTION AND TRANSFORMER BASED
    Song, Yihua
    Li, Can
    Zhang, Xia
    Liu, Zhen
    Song, Ningning
    Zhou, Zuojian
    [J]. JOURNAL OF MECHANICS IN MEDICINE AND BIOLOGY, 2024, 24 (02)
  • [29] MULTI-VIEW SELF-ATTENTION BASED TRANSFORMER FOR SPEAKER RECOGNITION
    Wang, Rui
    Ao, Junyi
    Zhou, Long
    Liu, Shujie
    Wei, Zhihua
    Ko, Tom
    Li, Qing
    Zhang, Yu
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6732 - 6736
  • [30] Wavelet Frequency Division Self-Attention Transformer Image Deraining Network
    Fang, Siyan
    Liu, Bin
    [J]. Computer Engineering and Applications, 2024, 60 (06) : 259 - 273