GMIM: Self-supervised pre-training for 3D medical image segmentation with adaptive and hierarchical masked image modeling

被引:0
|
作者
Qi L. [1 ]
Jiang Z. [1 ,2 ]
Shi W. [1 ,2 ]
Qu F. [1 ]
Feng G. [1 ]
机构
[1] Department of Computer Science and Technology, Changchun University of Science and Technology, Jilin, Changchun
[2] Zhongshan Institute of Changchun University of Science and Technology, Guangzhou, Zhongshan
关键词
Brain tumor segmentation; Masked image modeling; Self-supervised learning;
D O I
10.1016/j.compbiomed.2024.108547
中图分类号
学科分类号
摘要
Self-supervised pre-training and fully supervised fine-tuning paradigms have received much attention to solve the data annotation problem in deep learning fields. Compared with traditional pre-training on large natural image datasets, medical self-supervised learning methods learn rich representations derived from unlabeled data itself thus avoiding the distribution shift between different image domains. However, nowadays state-of-the-art medical pre-training methods were specifically designed for downstream tasks making them less flexible and difficult to apply to new tasks. In this paper, we propose grid mask image modeling, a flexible and general self-supervised method to pre-train medical vision transformers for 3D medical image segmentation. Our goal is to guide networks to learn the correlations between organs and tissues by reconstructing original images based on partial observations. The relationships are consistent within the human body and invariant to disease type or imaging modality. To achieve this, we design a Siamese framework consisting of an online branch and a target branch. An adaptive and hierarchical masking strategy is employed in the online branch to (1) learn the boundaries or small contextual mutation regions within images; (2) to learn high-level semantic representations from deeper layers of the multiscale encoder. In addition, the target branch provides representations for contrastive learning to further reduce representation redundancy. We evaluate our method through segmentation performance on two public datasets. The experimental results demonstrate our method outperforms other self-supervised methods. Codes are available at https://github.com/mobiletomb/Gmim. © 2024 Elsevier Ltd
引用
下载
收藏
相关论文
共 50 条
  • [41] Forecast-MAE: Self-supervised Pre-training for Motion Forecasting with Masked Autoencoders
    Cheng, Jie
    Mei, Xiaodong
    Liu, Ming
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 8645 - 8655
  • [42] Self-Supervised Pre-Training Boosts Semantic Scene Segmentation on LiDAR data
    Caros, Mariona
    Just, Ariadna
    Segui, Santi
    Vitria, Jordi
    2023 18TH INTERNATIONAL CONFERENCE ON MACHINE VISION AND APPLICATIONS, MVA, 2023,
  • [43] SPot-the-Difference Self-supervised Pre-training for Anomaly Detection and Segmentation
    Zou, Yang
    Jeong, Jongheon
    Pemula, Latha
    Zhang, Dongqing
    Dabeer, Onkar
    COMPUTER VISION - ECCV 2022, PT XXX, 2022, 13690 : 392 - 408
  • [44] Rotated and Masked Image Modeling: A Superior Self-Supervised Method for Classification
    Yan, Daisong
    Gong, Xun
    Zhang, Zhemin
    IEEE SIGNAL PROCESSING LETTERS, 2023, 30 : 1477 - 1481
  • [45] CroCo: Self-Supervised Pre-training for 3D Vision Tasks by Cross-View Completion
    Weinzaepfel, Philippe
    Leroy, Vincent
    Lucas, Thomas
    Bregier, Romain
    Cabon, Yohann
    Arora, Vaibhav
    Antsfeld, Leonid
    Chidlovskii, Boris
    Csurka, Gabriela
    Revaud, Jerome
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [46] Self-Supervised Pre-Training for 3-D Roof Reconstruction on LiDAR Data
    Yang, Hongxin
    Huang, Shangfeng
    Wang, Ruisheng
    Wang, Xin
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [47] Self-supervised pre-training with contrastive and masked autoencoder methods for dealing with small datasets in deep learning for medical imaging
    Daniel Wolf
    Tristan Payer
    Catharina Silvia Lisson
    Christoph Gerhard Lisson
    Meinrad Beer
    Michael Götz
    Timo Ropinski
    Scientific Reports, 13
  • [48] Self-supervised pre-training with contrastive and masked autoencoder methods for dealing with small datasets in deep learning for medical imaging
    Wolf, Daniel
    Payer, Tristan
    Lisson, Catharina Silvia
    Lisson, Christoph Gerhard
    Beer, Meinrad
    Gotz, Michael
    Ropinski, Timo
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [49] Adaptive-Masking Policy with Deep Reinforcement Learning for Self-Supervised Medical Image Segmentation
    Xu, Gang
    Wang, Shengxin
    Lukasiewicz, Thomas
    Xu, Zhenghua
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 2285 - 2290
  • [50] Self-supervised pseudo multi-class pre-training for unsupervised anomaly detection and segmentation in medical images
    Tian, Yu
    Liu, Fengbei
    Pang, Guansong
    Chen, Yuanhong
    Liu, Yuyuan
    Verjans, Johan W.
    Singh, Rajvinder
    Carneiro, Gustavo
    MEDICAL IMAGE ANALYSIS, 2023, 90