Advancing 3D medical image analysis with variable dimension transform based supervised 3D pre-training

被引:2
|
作者
Zhang, Shu [1 ]
Li, Zihao [1 ]
Zhou, Hong-Yu [2 ]
Ma, Jiechao [1 ]
Yu, Yizhou [1 ,2 ]
机构
[1] Deepwise Artificial Intelligence Lab, 8 Haidian Ave, Beijing, Peoples R China
[2] Univ Hong Kong, Dept Comp Sci, Pokfulam, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
3D medical image; Transfer learning; Variable dimension transform; Supervised pre-training; CT; NETWORK;
D O I
10.1016/j.neucom.2023.01.012
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The difficulties in both data acquisition and annotation substantially restrict the sample sizes of training datasets for 3D medical imaging applications. Therefore, it is non-trivial to build well-performing 3D con-volutional neural networks from scratch. Previous efforts on 3D pre-training have frequently relied on self-supervised approaches, which use either predictive or contrastive learning on unlabeled data to build invariant 3D representations. However, because of the unavailability of large-scale supervision informa-tion, obtaining semantically invariant and discriminative representations from these learning frame-works remains problematic. In this paper, we revisit an innovative yet simple fully-supervised 3D network pre-training framework to take advantage of semantic supervision from large-scale 2D natural image datasets. With a redesigned 3D network architecture, reformulated natural images are used to address the problem of data scarcity and develop powerful 3D representations. Comprehensive experi-ments on five benchmark datasets demonstrate that the proposed pre-trained models can effectively accelerate convergence while also improving accuracy for a variety of 3D medical imaging tasks such as classification, segmentation, and detection. In addition, as compared to training from scratch, it can save up to 60% of annotation efforts. On the NIH DeepLesion dataset, it also achieves state-of-the-art detection performance, outperforming earlier self-supervised and fully-supervised pre-training approaches, as well as methods that do training from scratch. To facilitate further development of 3D medical models, our code and pre-trained model weights are publicly available at https://github.com/u rmagicsmine/CSPR. (c) 2023 Elsevier B.V. All rights reserved.
引用
收藏
页码:11 / 22
页数:12
相关论文
共 50 条
  • [1] Self-Supervised Pre-Training of Swin Transformers for 3D Medical Image Analysis
    Tang, Yucheng
    Yang, Dong
    Li, Wenqi
    Roth, Holger R.
    Landman, Bennett
    Xu, Daguang
    Nath, Vishwesh
    Hatamizadeh, Ali
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 20698 - 20708
  • [2] Geometric Visual Similarity Learning in 3D Medical Image Self-supervised Pre-training
    He, Yuting
    Yang, Guanyu
    Ge, Rongjun
    Chen, Yang
    Coatrieux, Jean-Louis
    Wang, Boyu
    Li, Shuo
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 9538 - 9547
  • [3] Evaluating Task-Specific Augmentations in Self-Supervised Pre-Training for 3D Medical Image Analysis
    Claessens, C. H. B.
    Hamm, J. J. M.
    Viviers, C. G. A.
    Nederend, J.
    Grunhagen, D. J.
    Tanis, P. J.
    de With, P. H. N.
    van der Sommen, F.
    MEDICAL IMAGING 2024: IMAGE PROCESSING, 2024, 12926
  • [4] ReFs: A hybrid pre-training paradigm for 3D medical image segmentation
    Xie, Yutong
    Zhang, Jianpeng
    Liu, Lingqiao
    Wang, Hu
    Ye, Yiwen
    Verjans, Johan
    Xia, Yong
    MEDICAL IMAGE ANALYSIS, 2024, 91
  • [5] GMIM: Self-supervised pre-training for 3D medical image segmentation with adaptive and hierarchical masked image modeling
    Qi L.
    Jiang Z.
    Shi W.
    Qu F.
    Feng G.
    Computers in Biology and Medicine, 2024, 176
  • [6] Sculpting Holistic 3D Representation in Contrastive Language-Image-3D Pre-Training
    Gao, Yipeng
    Wang, Zeyu
    Zheng, Wei-Shi
    Xie, Cihang
    Zhou, Yuyin
    Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2024, : 22998 - 23008
  • [7] A Closer Look at Invariances in Self-supervised Pre-training for 3D Vision
    Li, Lanxiao
    Heizmann, Michael
    COMPUTER VISION - ECCV 2022, PT XXX, 2022, 13690 : 656 - 673
  • [8] Self-supervised Pre-training with Masked Shape Prediction for 3D Scene Understanding
    Jiang, Li
    Yang, Zetong
    Shi, Shaoshuai
    Golyanik, Vladislav
    Dai, Dengxin
    Schiele, Bernt
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 1168 - 1178
  • [9] Point Cloud Pre-training with Natural 3D Structures
    Yamada, Ryosuke
    Kataoka, Hirokatsu
    Chiba, Naoya
    Domae, Yukiyasu
    Ogata, Tetsuya
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 21251 - 21261
  • [10] Unsupervised Pre-Training for 3D Leaf Instance Segmentation
    Roggiolani, Gianmarco
    Magistri, Federico
    Guadagnino, Tiziano
    Behley, Jens
    Stachniss, Cyrill
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (11) : 7448 - 7455