Advancing 3D medical image analysis with variable dimension transform based supervised 3D pre-training

被引:2
|
作者
Zhang, Shu [1 ]
Li, Zihao [1 ]
Zhou, Hong-Yu [2 ]
Ma, Jiechao [1 ]
Yu, Yizhou [1 ,2 ]
机构
[1] Deepwise Artificial Intelligence Lab, 8 Haidian Ave, Beijing, Peoples R China
[2] Univ Hong Kong, Dept Comp Sci, Pokfulam, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
3D medical image; Transfer learning; Variable dimension transform; Supervised pre-training; CT; NETWORK;
D O I
10.1016/j.neucom.2023.01.012
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The difficulties in both data acquisition and annotation substantially restrict the sample sizes of training datasets for 3D medical imaging applications. Therefore, it is non-trivial to build well-performing 3D con-volutional neural networks from scratch. Previous efforts on 3D pre-training have frequently relied on self-supervised approaches, which use either predictive or contrastive learning on unlabeled data to build invariant 3D representations. However, because of the unavailability of large-scale supervision informa-tion, obtaining semantically invariant and discriminative representations from these learning frame-works remains problematic. In this paper, we revisit an innovative yet simple fully-supervised 3D network pre-training framework to take advantage of semantic supervision from large-scale 2D natural image datasets. With a redesigned 3D network architecture, reformulated natural images are used to address the problem of data scarcity and develop powerful 3D representations. Comprehensive experi-ments on five benchmark datasets demonstrate that the proposed pre-trained models can effectively accelerate convergence while also improving accuracy for a variety of 3D medical imaging tasks such as classification, segmentation, and detection. In addition, as compared to training from scratch, it can save up to 60% of annotation efforts. On the NIH DeepLesion dataset, it also achieves state-of-the-art detection performance, outperforming earlier self-supervised and fully-supervised pre-training approaches, as well as methods that do training from scratch. To facilitate further development of 3D medical models, our code and pre-trained model weights are publicly available at https://github.com/u rmagicsmine/CSPR. (c) 2023 Elsevier B.V. All rights reserved.
引用
收藏
页码:11 / 22
页数:12
相关论文
共 50 条
  • [31] 3D medical image segmentation based on semi-supervised learning using deep co-training
    Yang, Jingdong
    Li, Haoqiu
    Wang, Han
    Han, Man
    APPLIED SOFT COMPUTING, 2024, 159
  • [32] Advancing RNA FISH image analysis with 3D deep learning
    Rojewski, Alex
    Schweiger, Max
    Kruithoff, Rory
    Shepherd, Douglas P.
    Presse, Steve
    BIOPHYSICAL JOURNAL, 2024, 123 (03) : 552A - 552A
  • [33] Uni4Eye: Unified 2D and 3D Self-supervised Pre-training via Masked Image Modeling Transformer for Ophthalmic Image Classification
    Cai, Zhiyuan
    Lin, Li
    He, Huaqing
    Tang, Xiaoying
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT VIII, 2022, 13438 : 88 - 98
  • [34] 3d Image Compression with Variable Fragments
    Sokolova, Ekaterina A.
    Nyrkov, Anatoliy P.
    Ivanovskii, Aleksei N.
    PROCEEDINGS OF THE 2021 IEEE CONFERENCE OF RUSSIAN YOUNG RESEARCHERS IN ELECTRICAL AND ELECTRONIC ENGINEERING (ELCONRUS), 2021, : 699 - 702
  • [35] Multi-Camera Unified Pre-Training via 3D Scene Reconstruction
    Min, Chen
    Xiao, Liang
    Zhao, Dawei
    Nie, Yiming
    Dai, Bin
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (04) : 3243 - 3250
  • [36] FocalMix: Semi-Supervised Learning for 3D Medical Image Detection
    Wang, Dong
    Zhang, Yuan
    Zhang, Kexin
    Wang, Liwei
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 3950 - 3959
  • [37] Light mixed-supervised segmentation for 3D medical image data
    Yang, Hongxu
    Tan, Tao
    Tegzes, Pal
    Dong, Xiaomeng
    Tamada, Rajesh
    Ferenczi, Lehel
    Avinash, Gopal
    MEDICAL PHYSICS, 2024, 51 (01) : 167 - 178
  • [38] Joint-MAE: 2D-3D Joint Masked Autoencoders for 3D Point Cloud Pre-training
    Guo, Ziyu
    Zhang, Renrui
    Qiu, Longtian
    Li, Xianzhi
    Heng, Pheng-Ann
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 791 - 799
  • [39] Lang3DSG: Language-based contrastive pre-training for 3D Scene Graph prediction
    Koch, Sebastian
    Hermosilla, Pedro
    Vaskevicius, Narunas
    Colosi, Mirco
    Ropinski, Timo
    2024 INTERNATIONAL CONFERENCE IN 3D VISION, 3DV 2024, 2024, : 1037 - 1047
  • [40] 3D Shape Recovery by Aggregating 3D Wavelet Transform-Based Image Focus Volumes Through 3D Weighted Least Squares
    Ali, Usman
    Mahmood, Muhammad Tariq
    JOURNAL OF MATHEMATICAL IMAGING AND VISION, 2020, 62 (01) : 54 - 72