GMIM: Self-supervised pre-training for 3D medical image segmentation with adaptive and hierarchical masked image modeling

被引:0
|
作者
Qi L. [1 ]
Jiang Z. [1 ,2 ]
Shi W. [1 ,2 ]
Qu F. [1 ]
Feng G. [1 ]
机构
[1] Department of Computer Science and Technology, Changchun University of Science and Technology, Jilin, Changchun
[2] Zhongshan Institute of Changchun University of Science and Technology, Guangzhou, Zhongshan
关键词
Brain tumor segmentation; Masked image modeling; Self-supervised learning;
D O I
10.1016/j.compbiomed.2024.108547
中图分类号
学科分类号
摘要
Self-supervised pre-training and fully supervised fine-tuning paradigms have received much attention to solve the data annotation problem in deep learning fields. Compared with traditional pre-training on large natural image datasets, medical self-supervised learning methods learn rich representations derived from unlabeled data itself thus avoiding the distribution shift between different image domains. However, nowadays state-of-the-art medical pre-training methods were specifically designed for downstream tasks making them less flexible and difficult to apply to new tasks. In this paper, we propose grid mask image modeling, a flexible and general self-supervised method to pre-train medical vision transformers for 3D medical image segmentation. Our goal is to guide networks to learn the correlations between organs and tissues by reconstructing original images based on partial observations. The relationships are consistent within the human body and invariant to disease type or imaging modality. To achieve this, we design a Siamese framework consisting of an online branch and a target branch. An adaptive and hierarchical masking strategy is employed in the online branch to (1) learn the boundaries or small contextual mutation regions within images; (2) to learn high-level semantic representations from deeper layers of the multiscale encoder. In addition, the target branch provides representations for contrastive learning to further reduce representation redundancy. We evaluate our method through segmentation performance on two public datasets. The experimental results demonstrate our method outperforms other self-supervised methods. Codes are available at https://github.com/mobiletomb/Gmim. © 2024 Elsevier Ltd
引用
下载
收藏
相关论文
共 50 条
  • [1] Self-Supervised Pre-Training of Swin Transformers for 3D Medical Image Analysis
    Tang, Yucheng
    Yang, Dong
    Li, Wenqi
    Roth, Holger R.
    Landman, Bennett
    Xu, Daguang
    Nath, Vishwesh
    Hatamizadeh, Ali
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 20698 - 20708
  • [2] Geometric Visual Similarity Learning in 3D Medical Image Self-supervised Pre-training
    He, Yuting
    Yang, Guanyu
    Ge, Rongjun
    Chen, Yang
    Coatrieux, Jean-Louis
    Wang, Boyu
    Li, Shuo
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 9538 - 9547
  • [3] Intra-modality masked image modeling: A self-supervised pre-training method for brain tumor segmentation
    Qi, Liangce
    Shi, Weili
    Miao, Yu
    Li, Yonghui
    Feng, Guanyuan
    Jiang, Zhengang
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 95
  • [4] Correlational Image Modeling for Self-Supervised Visual Pre-Training
    Li, Wei
    Xie, Jiahao
    Loy, Chen Change
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 15105 - 15115
  • [5] SELF PRE-TRAINING WITH MASKED AUTOENCODERS FOR MEDICAL IMAGE CLASSIFICATION AND SEGMENTATION
    Zhou, Lei
    Liu, Huidong
    Bae, Joseph
    He, Junjun
    Samaras, Dimitris
    Prasanna, Prateek
    2023 IEEE 20TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING, ISBI, 2023,
  • [6] Evaluating Task-Specific Augmentations in Self-Supervised Pre-Training for 3D Medical Image Analysis
    Claessens, C. H. B.
    Hamm, J. J. M.
    Viviers, C. G. A.
    Nederend, J.
    Grunhagen, D. J.
    Tanis, P. J.
    de With, P. H. N.
    van der Sommen, F.
    MEDICAL IMAGING 2024: IMAGE PROCESSING, 2024, 12926
  • [7] Self-supervised Pre-training with Masked Shape Prediction for 3D Scene Understanding
    Jiang, Li
    Yang, Zetong
    Shi, Shaoshuai
    Golyanik, Vladislav
    Dai, Dengxin
    Schiele, Bernt
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 1168 - 1178
  • [8] Uni4Eye: Unified 2D and 3D Self-supervised Pre-training via Masked Image Modeling Transformer for Ophthalmic Image Classification
    Cai, Zhiyuan
    Lin, Li
    He, Huaqing
    Tang, Xiaoying
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT VIII, 2022, 13438 : 88 - 98
  • [9] DiT: Self-supervised Pre-training for Document Image Transformer
    Li, Junlong
    Xu, Yiheng
    Lv, Tengchao
    Cui, Lei
    Zhang, Cha
    Wei, Furu
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 3530 - 3539
  • [10] Self-supervised Pre-training for Nuclei Segmentation
    Haq, Mohammad Minhazul
    Huang, Junzhou
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT II, 2022, 13432 : 303 - 313