MetaTransformer: deep metagenomic sequencing read classification using self-attention models

被引:6
|
作者
Wichmann, Alexander [1 ]
Buschong, Etienne [1 ]
Mueller, Andre [1 ]
Juenger, Daniel [1 ]
Hildebrandt, Andreas [1 ]
Hankeln, Thomas [2 ]
Schmidt, Bertil [1 ]
机构
[1] Johannes Gutenberg Univ Mainz, Inst Comp Sci, Staudingerweg 9, D-55128 Mainz, Rhineland Palat, Germany
[2] Johannes Gutenberg Univ Mainz, Inst Organ & Mol Evolut iomE, J-J Becher Weg 30A, D-55128 Mainz, Rhineland Palat, Germany
关键词
MICROBIOME; GENOMES;
D O I
10.1093/nargab/lqad082
中图分类号
Q3 [遗传学];
学科分类号
071007 ; 090102 ;
摘要
Deep learning has emerged as a paradigm that revolutionizes numerous domains of scientific research. Transformers have been utilized in language modeling outperforming previous approaches. Therefore, the utilization of deep learning as a tool for analyzing the genomic sequences is promising, yielding convincing results in fields such as motif identification and variant calling. DeepMicrobes, a machine learning-based classifier, has recently been introduced for taxonomic prediction at species and genus level. However, it relies on complex models based on bidirectional long short-term memory cells resulting in slow runtimes and excessive memory requirements, hampering its effective usability. We present MetaTransformer, a self-attention-based deep learning metagenomic analysis tool. Our transformer-encoder-based models enable efficient parallelization while outperforming DeepMicrobes in terms of species and genus classification abilities. Furthermore, we investigate approaches to reduce memory consumption and boost performance using different embedding schemes. As a result, we are able to achieve 2x to 5x speedup for inference compared to DeepMicrobes while keeping a significantly smaller memory footprint. MetaTransformer can be trained in 9 hours for genus and 16 hours for species prediction. Our results demonstrate performance improvements due to self-attention models and the impact of embedding schemes in deep learning on metagenomic sequencing data.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] SABDM: A self-attention based bidirectional-RNN deep model for requirements classification
    Kaur, Kamaljit
    Kaur, Parminder
    JOURNAL OF SOFTWARE-EVOLUTION AND PROCESS, 2024, 36 (02)
  • [22] Denoising adaptive deep clustering with self-attention mechanism on single-cell sequencing data
    Su, Yansen
    Lin, Rongxin
    Wang, Jing
    Tan, Dayu
    Zheng, Chunhou
    BRIEFINGS IN BIOINFORMATICS, 2023, 24 (02)
  • [23] Improving Sample Quality of Diffusion Models Using Self-Attention Guidance
    Hong, Susung
    Lee, Gyuseong
    Jang, Wooseok
    Kim, Seungryong
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 7428 - 7437
  • [24] Synthesizer: Rethinking Self-Attention for Transformer Models
    Tay, Yi
    Bahri, Dara
    Metzler, Donald
    Juan, Da-Cheng
    Zhao, Zhe
    Zheng, Che
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139 : 7192 - 7203
  • [25] Fine-grained entity type classification using GRU with self-attention
    Dhrisya K.
    Remya G.
    Mohan A.
    International Journal of Information Technology, 2020, 12 (3) : 869 - 878
  • [26] Assessing the Impact of Attention and Self-Attention Mechanisms on the Classification of Skin Lesions
    Pedro, Rafael
    Oliveira, Arlindo L.
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [27] Fake news detection and classification using hybrid BiLSTM and self-attention model
    Mohapatra, Asutosh
    Thota, Nithin
    Prakasam, P.
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (13) : 18503 - 18519
  • [28] Multi-Scale Self-Attention for Text Classification
    Guo, Qipeng
    Qiu, Xipeng
    Liu, Pengfei
    Xue, Xiangyang
    Zhang, Zheng
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7847 - 7854
  • [29] TESANet: Self-attention network for olfactory EEG classification
    Tong, Chengxuan
    Ding, Yi
    Liang, Kevin Lim Jun
    Zhang, Zhuo
    Zhang, Haihong
    Guan, Cuntai
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [30] Grassmannian Manifold Self-Attention Network for Signal Classification
    Wang, Rui
    Hu, Chen
    Chen, Ziheng
    Wu, Xiao-Jun
    Song, Xiaoning
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 5099 - 5107