CvFormer: Cross-view transFormers with pre-training for fMRI analysis of human brain

被引:0
|
作者
Meng, Xiangzhu [1 ,3 ]
Wei, Wei [4 ]
Liu, Qiang [1 ,2 ]
Wang, Yu [3 ]
Li, Min [3 ]
Wang, Liang [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Automat, Ctr Res Intelligent Percept & Comp, 95 Zhongguancun East Rd, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, 1 Yanqihu East Rd, Beijing 101408, Peoples R China
[3] Jing Dong Retail, Dept User Growth & Operat, 18,Kechuang 11th St, Beijing 100176, Peoples R China
[4] Zhengzhou Univ, Sch Management, 100 Sci Ave, Zhengzhou 450001, Henan, Peoples R China
基金
中国国家自然科学基金;
关键词
Functional MRI; Human brain; Cross-view modeling; Transformers; Self-supervised learning; CONVOLUTIONAL NEURAL-NETWORKS;
D O I
10.1016/j.patrec.2024.09.010
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, functional magnetic resonance imaging (fMRI) has been widely utilized to diagnose neurological disease, by exploiting the region of interest (RoI) nodes as well as their connectivities in human brain. However, most of existing works only rely on either RoIs or connectivities, neglecting the potential for complementary information between them. To address this issue, we study how to discover the rich cross-view information in fMRI data of human brain. This paper presents a novel method for cross-view analysis of fMRI data of the human brain, called Cross-view transFormers (CvFormer). CvFormer employs RoI and connectivity encoder modules to generate two separate views of the human brain, represented as RoI and sub-connectivity tokens. Then, basic transformer modules can be used to process the RoI and sub-connectivity tokens, and cross-view modules integrate the complement information across two views. Furthermore, CvFormer uses a global token for each branch as a query to exchange information with other branches in cross-view modules, which only requires linear time for both computational and memory complexity instead of quadratic time. To enhance the robustness of the proposed CvFormer, we propose a two-stage strategy to train its parameters. To be specific, RoI and connectivity views can be firstly utilized as self-supervised information to pre-train the CvFormer by combining it with contrastive learning and then fused to finetune the CvFormer using label information. Experiment results on two public ABIDE and ADNI datasets can show clear improvements by the proposed CvFormer, which can validate its effectiveness and superiority.
引用
收藏
页码:85 / 90
页数:6
相关论文
共 50 条
  • [41] SatMAE: Pre-training Transformers for Temporal and Multi-Spectral Satellite Imagery
    Cong, Yezhen
    Khanna, Samar
    Meng, Chenlin
    Liu, Patrick
    Rozi, Erik
    He, Yutong
    Burke, Marshall
    Lobell, David B.
    Ermon, Stefano
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [42] Mask and Reason: Pre-Training Knowledge Graph Transformers for Complex Logical Queries
    Liu, Xiao
    Zhao, Shiyu
    Su, Kai
    Cen, Yukuo
    Qiu, Jiezhong
    Zhang, Mengdi
    Wu, Wei
    Dong, Yuxiao
    Tang, Jie
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 1120 - 1130
  • [43] Beyond masking: Demystifying token-based pre-training for vision transformers
    Tian, Yunjie
    Xie, Lingxi
    Fang, Jiemin
    Jiao, Jianbin
    Tian, Qi
    PATTERN RECOGNITION, 2025, 162
  • [44] Transferring Training Instances for Convenient Cross-View Object Classification in Surveillance
    Zhang, Zhaoxiang
    Zhao, Yuhang
    Wang, Yunhong
    Liu, Jianyun
    Yao, Zhenjun
    Tang, Jun
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2013, 8 (10) : 1632 - 1641
  • [45] An Efficient Self-Supervised Cross-View Training For Sentence Embedding
    Limkonchotiwat, Peerat
    Ponwitayarat, Wuttikorn
    Lowphansirikul, Lalita
    Udomcharoenchaikit, Can
    Chuangsuwanich, Ekapol
    Nutanong, Sarana
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2023, 11 : 1572 - 1587
  • [46] Heterogeneous discriminant analysis for cross-view action recognition
    Sui, Wanchen
    Wu, Xinxiao
    Feng, Yang
    Jia, Yunde
    NEUROCOMPUTING, 2016, 191 : 286 - 295
  • [47] Cross-View Human Intention Recognition for Human-Robot Collaboration
    Ni, Shouxiang
    Zhao, Lindong
    Li, Ang
    Wu, Dan
    Zhou, Liang
    IEEE WIRELESS COMMUNICATIONS, 2023, 30 (03) : 189 - 195
  • [48] Comprehensive analysis of embeddings and pre-training in NLP
    Tripathy, Jatin Karthik
    Sethuraman, Sibi Chakkaravarthy
    Cruz, Meenalosini Vimal
    Namburu, Anupama
    Mangalraj, P.
    Kumar, R. Nandha
    Ilango, S. Sudhakar
    Vijayakumar, Vaidehi
    COMPUTER SCIENCE REVIEW, 2021, 42
  • [49] Semi-Supervised Semantic Role Labeling with Cross-View Training
    Cai, Rui
    Lapata, Mirella
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1018 - 1027
  • [50] Heterogeneous Discriminant Analysis for Cross-View Action Recognition
    Sui, Wanchen
    Wu, Xinxiao
    Feng, Yang
    Liang, Wei
    Jia, Yunde
    NEURAL INFORMATION PROCESSING, ICONIP 2015, PT IV, 2015, 9492 : 566 - 573