Predicting transcription factor binding sites by a multi-modal representation learning method based on cross-attention network

被引:0
|
作者
Wei, Yuxiao [1 ]
Zhang, Qi [2 ]
Liu, Liwei [2 ]
机构
[1] Dalian Jiaotong Univ, Coll Software, Dalian 116028, Peoples R China
[2] Dalian Jiaotong Univ, Coll Sci, Dalian 116028, Peoples R China
关键词
Transcription factor binding sites; Deep learning; Cross-attention mechanism; Model interpretability; CHIP-SEQ; DNA; SPECIFICITIES;
D O I
10.1016/j.asoc.2024.112134
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The prediction of transcription factor binding sites (TFBS) plays a crucial role in studying cellular functions and understanding transcriptional regulatory processes. With the development of chromatin immunoprecipitation sequencing (ChIP-seq) technology, an increasing number of computer-aided TFBS prediction models have emerged. However, how to integrate multi-modal information of DNA and obtain efficient features to improve prediction accuracy remains a major challenge. Here, we propose MultiTF, a multi-modal representation learning method based on a cross-attention network for predicting transcription factor binding sites. Among TFBS prediction methods, we are the first to use graph neural networks and cross-attention networks for representation learning. MultiTF uses dna2vec to extract global contextual features of DNA sequences, DNAshapeR to extract shape features, and the CDPfold model and graph attention network for learning and representation of DNA structural features. Finally, with the help of our cross-attention module, we successfully combine sequence, structural, and shape features to achieve interactive fusion. When comparing MultiTF to other state-of-the-art methods using 165 ENCODE ChIP-seq datasets, we find that MultiTF exhibits average ACC, ROC-AUC, and PR-AUC values of 0.911, 0.978, and 0.982, respectively. The results show that MultiTF achieves unprecedented prediction accuracy compared to previous TFBS prediction models. In addition, our visual analysis of structural features provides interpretability for the prediction results.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Multi-modal cross-attention network for Alzheimer's disease diagnosis with multi data
    Zhang, Jin
    He, Xiaohai
    Liu, Yan
    Cai, Qingyan
    Chen, Honggang
    Qing, Linbo
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 162
  • [2] Multi-Level Multi-Modal Cross-Attention Network for Fake News Detection
    Ying, Long
    Yu, Hui
    Wang, Jinguang
    Ji, Yongze
    Qian, Shengsheng
    IEEE ACCESS, 2021, 9 : 132363 - 132373
  • [3] Multi-modal Network Representation Learning
    Zhang, Chuxu
    Jiang, Meng
    Zhang, Xiangliang
    Ye, Yanfang
    Chawla, Nitesh, V
    KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 3557 - 3558
  • [4] IS CROSS-ATTENTION PREFERABLE TO SELF-ATTENTION FOR MULTI-MODAL EMOTION RECOGNITION?
    Rajan, Vandana
    Brutti, Alessio
    Cavallaro, Andrea
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4693 - 4697
  • [5] Cross-Attention Model for Multi-modal Bio-Signal Processing
    Heesoo, Son
    Sangseok, Lee
    Sael, Lee
    2022 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (IEEE BIGCOMP 2022), 2022, : 43 - 46
  • [6] A joint hierarchical cross-attention graph convolutional network for multi-modal facial expression recognition
    Xu, Chujie
    Du, Yong
    Wang, Jingzi
    Zheng, Wenjie
    Li, Tiejun
    Yuan, Zhansheng
    COMPUTATIONAL INTELLIGENCE, 2024, 40 (01)
  • [7] Multi-Modal Sentiment Analysis Based on Image and Text Fusion Based on Cross-Attention Mechanism
    Li, Hongchan
    Lu, Yantong
    Zhu, Haodong
    ELECTRONICS, 2024, 13 (11)
  • [8] MutualFormer: Multi-modal Representation Learning via Cross-Diffusion Attention
    Wang, Xixi
    Wang, Xiao
    Jiang, Bo
    Tang, Jin
    Luo, Bin
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (09) : 3867 - 3888
  • [9] Mineral: Multi-modal Network Representation Learning
    Kefato, Zekarias T.
    Sheikh, Nasrullah
    Montresor, Alberto
    MACHINE LEARNING, OPTIMIZATION, AND BIG DATA, MOD 2017, 2018, 10710 : 286 - 298
  • [10] Multi-modal Perception Fusion Method Based on Cross Attention
    Zhang B.-L.
    Pan Z.-H.
    Jiang J.-Z.
    Zhang C.-B.
    Wang Y.-X.
    Yang C.-L.
    Zhongguo Gonglu Xuebao/China Journal of Highway and Transport, 2024, 37 (03): : 181 - 193