CvFormer: Cross-view transFormers with pre-training for fMRI analysis of human brain

被引:0
|
作者
Meng, Xiangzhu [1 ,3 ]
Wei, Wei [4 ]
Liu, Qiang [1 ,2 ]
Wang, Yu [3 ]
Li, Min [3 ]
Wang, Liang [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Automat, Ctr Res Intelligent Percept & Comp, 95 Zhongguancun East Rd, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, 1 Yanqihu East Rd, Beijing 101408, Peoples R China
[3] Jing Dong Retail, Dept User Growth & Operat, 18,Kechuang 11th St, Beijing 100176, Peoples R China
[4] Zhengzhou Univ, Sch Management, 100 Sci Ave, Zhengzhou 450001, Henan, Peoples R China
基金
中国国家自然科学基金;
关键词
Functional MRI; Human brain; Cross-view modeling; Transformers; Self-supervised learning; CONVOLUTIONAL NEURAL-NETWORKS;
D O I
10.1016/j.patrec.2024.09.010
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, functional magnetic resonance imaging (fMRI) has been widely utilized to diagnose neurological disease, by exploiting the region of interest (RoI) nodes as well as their connectivities in human brain. However, most of existing works only rely on either RoIs or connectivities, neglecting the potential for complementary information between them. To address this issue, we study how to discover the rich cross-view information in fMRI data of human brain. This paper presents a novel method for cross-view analysis of fMRI data of the human brain, called Cross-view transFormers (CvFormer). CvFormer employs RoI and connectivity encoder modules to generate two separate views of the human brain, represented as RoI and sub-connectivity tokens. Then, basic transformer modules can be used to process the RoI and sub-connectivity tokens, and cross-view modules integrate the complement information across two views. Furthermore, CvFormer uses a global token for each branch as a query to exchange information with other branches in cross-view modules, which only requires linear time for both computational and memory complexity instead of quadratic time. To enhance the robustness of the proposed CvFormer, we propose a two-stage strategy to train its parameters. To be specific, RoI and connectivity views can be firstly utilized as self-supervised information to pre-train the CvFormer by combining it with contrastive learning and then fused to finetune the CvFormer using label information. Experiment results on two public ABIDE and ADNI datasets can show clear improvements by the proposed CvFormer, which can validate its effectiveness and superiority.
引用
收藏
页码:85 / 90
页数:6
相关论文
共 50 条
  • [1] Pre-Training Transformers for Fingerprinting to Improve Stress Prediction in fMRI
    Rosenman, Gony
    Malkiel, Itzik
    Greental, Ayam
    Hendler, Talma
    Wolf, Lior
    MEDICAL IMAGING WITH DEEP LEARNING, VOL 227, 2023, 227 : 212 - 234
  • [2] Cross-View Language Modeling: Towards Unified Cross-Lingual Cross-Modal Pre-training
    Zeng, Yan
    Zhou, Wangchunshu
    Luo, Ao
    Cheng, Ziming
    Zhang, Xinsong
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 5731 - 5746
  • [3] Spatial-Temporal Cross-View Contrastive Pre-Training for Check-in Sequence Representation Learning
    Gong, Letian
    Wan, Huaiyu
    Guo, Shengnan
    Li, Xiucheng
    Lin, Yan
    Zheng, Erwen
    Wang, Tianyi
    Zhou, Zeyu
    Lin, Youfang
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (12) : 9308 - 9321
  • [4] CroCo: Self-Supervised Pre-training for 3D Vision Tasks by Cross-View Completion
    Weinzaepfel, Philippe
    Leroy, Vincent
    Lucas, Thomas
    Bregier, Romain
    Cabon, Yohann
    Arora, Vaibhav
    Antsfeld, Leonid
    Chidlovskii, Boris
    Csurka, Gabriela
    Revaud, Jerome
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [5] CroCo v2: Improved Cross-view Completion Pre-training for Stereo Matching and Optical Flow
    Weinzaepfel, Philippe
    Lucas, Thomas
    Leroy, Vincent
    Cabon, Yohann
    Arora, Vaibhav
    Bregier, Romain
    Csurka, Gabriela
    Antsfeld, Leonid
    Chidlovskii, Boris
    Revaud, Jerome
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 17923 - 17934
  • [6] Unsupervised Pre-Training for Detection Transformers
    Dai, Zhigang
    Cai, Bolun
    Lin, Yugeng
    Chen, Junying
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (11) : 12772 - 12782
  • [7] Multi-view Analysis of Unregistered Medical Images Using Cross-View Transformers
    van Tulder, Gijs
    Tong, Yao
    Marchiori, Elena
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT III, 2021, 12903 : 104 - 113
  • [8] Evaluation of FractalDB Pre-training with Vision Transformers
    Nakashima K.
    Kataoka H.
    Satoh Y.
    Seimitsu Kogaku Kaishi/Journal of the Japan Society for Precision Engineering, 2023, 89 (01): : 99 - 104
  • [9] Pre-training of Graph Augmented Transformers for Medication Recommendation
    Shang, Junyuan
    Ma, Tengfei
    Xiao, Cao
    Sun, Jimeng
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 5953 - 5959
  • [10] Lifting the Curse of Multilinguality by Pre-training Modular Transformers
    Pfeiffer, Jonas
    Goyal, Naman
    Lin, Xi Victoria
    Li, Xian
    Cross, James
    Riedel, Sebastian
    Artetxe, Mikel
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 3479 - 3495