MetaTransformer: deep metagenomic sequencing read classification using self-attention models

被引:6
|
作者
Wichmann, Alexander [1 ]
Buschong, Etienne [1 ]
Mueller, Andre [1 ]
Juenger, Daniel [1 ]
Hildebrandt, Andreas [1 ]
Hankeln, Thomas [2 ]
Schmidt, Bertil [1 ]
机构
[1] Johannes Gutenberg Univ Mainz, Inst Comp Sci, Staudingerweg 9, D-55128 Mainz, Rhineland Palat, Germany
[2] Johannes Gutenberg Univ Mainz, Inst Organ & Mol Evolut iomE, J-J Becher Weg 30A, D-55128 Mainz, Rhineland Palat, Germany
关键词
MICROBIOME; GENOMES;
D O I
10.1093/nargab/lqad082
中图分类号
Q3 [遗传学];
学科分类号
071007 ; 090102 ;
摘要
Deep learning has emerged as a paradigm that revolutionizes numerous domains of scientific research. Transformers have been utilized in language modeling outperforming previous approaches. Therefore, the utilization of deep learning as a tool for analyzing the genomic sequences is promising, yielding convincing results in fields such as motif identification and variant calling. DeepMicrobes, a machine learning-based classifier, has recently been introduced for taxonomic prediction at species and genus level. However, it relies on complex models based on bidirectional long short-term memory cells resulting in slow runtimes and excessive memory requirements, hampering its effective usability. We present MetaTransformer, a self-attention-based deep learning metagenomic analysis tool. Our transformer-encoder-based models enable efficient parallelization while outperforming DeepMicrobes in terms of species and genus classification abilities. Furthermore, we investigate approaches to reduce memory consumption and boost performance using different embedding schemes. As a result, we are able to achieve 2x to 5x speedup for inference compared to DeepMicrobes while keeping a significantly smaller memory footprint. MetaTransformer can be trained in 9 hours for genus and 16 hours for species prediction. Our results demonstrate performance improvements due to self-attention models and the impact of embedding schemes in deep learning on metagenomic sequencing data.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Deep Alternate Kernel Fused Self-Attention Model-Based Lung Nodule Classification
    Saritha, R. Rani
    Sangeetha, V.
    JOURNAL OF ADVANCES IN INFORMATION TECHNOLOGY, 2024, 15 (11) : 1242 - 1251
  • [42] Self-attention Adversarial Based Deep Subspace Clustering
    Yin M.
    Wu H.-Y.
    Xie S.-L.
    Yang Q.-Y.
    Zidonghua Xuebao/Acta Automatica Sinica, 2022, 48 (01): : 271 - 281
  • [43] Reconstructing computational spectra using deep learning’s self-attention method
    Wu, Hao
    Wu, Hui
    Su, Xinyu
    Wu, Jingjun
    Liu, Shuangli
    Optica Applicata, 2024, 54 (03) : 383 - 394
  • [44] A Dynamic Self-Attention Mechanism for Improving Deep Learning-based Plant Disease Classification
    Akella, Gopi Krishna
    Wibowo, Santoso
    Grandhi, Srimannarayana
    Sabrina, Fariza
    Mubarak, Sameera
    27TH IEEE/ACIS INTERNATIONAL SUMMER CONFERENCE ON SOFTWARE ENGINEERING ARTIFICIAL INTELLIGENCE NETWORKING AND PARALLEL/DISTRIBUTED COMPUTING, SNPD 2024-SUMMER, 2024, : 222 - 227
  • [45] SELF-ATTENTION GUIDED DEEP FEATURES FOR ACTION RECOGNITION
    Xiao, Renyi
    Hou, Yonghong
    Guo, Zihui
    Li, Chuankun
    Wang, Pichao
    Li, Wanqing
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1060 - 1065
  • [46] A Deep Learning Method Based Self-Attention and Bi-directional LSTM in Emotion Classification
    Fei, Rong
    Zhu, Yuanbo
    Yao, Quanzhu
    Xu, Qingzheng
    Hu, Bo
    JOURNAL OF INTERNET TECHNOLOGY, 2020, 21 (05): : 1447 - 1461
  • [47] Sentence Matching with Deep Self-attention and Co-attention Features
    Wang, Zhipeng
    Yan, Danfeng
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2021, PT II, 2021, 12816 : 550 - 561
  • [48] Homogeneous Learning: Self-Attention Decentralized Deep Learning
    Sun, Yuwei
    Ochiai, Hideya
    IEEE ACCESS, 2022, 10 : 7695 - 7703
  • [49] FCN-Attention: A deep learning UWB NLOS/LOS classification algorithm using fully convolution neural network with self-attention mechanism
    Pei, Yu
    Chen, Ruizhi
    Li, Deren
    Xiao, Xiongwu
    Zheng, Xingyu
    GEO-SPATIAL INFORMATION SCIENCE, 2024, 27 (04): : 1162 - 1181
  • [50] Continuous Self-Attention Models with Neural ODE Networks
    Zhang, Jing
    Zhang, Peng
    Kong, Baiwen
    Wei, Junqiu
    Jiang, Xin
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 14393 - 14401