Self-supervised multi-modal fusion network for multi-modal thyroid ultrasound image diagnosis

被引:21
|
作者
Xiang, Zhuo [1 ]
Zhuo, Qiuluan [2 ]
Zhao, Cheng [1 ]
Deng, Xiaofei [2 ]
Zhu, Ting [2 ]
Wang, Tianfu [1 ]
Jiang, Wei [2 ]
Lei, Baiying [1 ]
机构
[1] Shenzhen Univ, Natl Reg Key Technol Engn Lab Med Ultrasound, Guangdong Key Lab Biomed Measurements & Ultrasound, Sch Biomed Engn,Hlth Sci Ctr, Shenzhen, Peoples R China
[2] Huazhong Univ Sci & Technol, Union Shenzhen Hosp, Dept Ultrasound, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
Thyroid ultrasound; Thyroid diagnosis; Multi-modal; Self-supervision; SHEAR-WAVE ELASTOGRAPHY; CONVOLUTIONAL NEURAL-NETWORK; INTRAOBSERVER REPRODUCIBILITY; NODULES; PERFORMANCE;
D O I
10.1016/j.compbiomed.2022.106164
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Ultrasound is a typical non-invasive diagnostic method often used to detect thyroid cancer lesions. However, due to the limitations of the information provided by ultrasound images, shear wave elastography (SWE) and color doppler ultrasound (CDUS) are also used clinically to assist in diagnosis, which makes the diagnosis time-consuming, labor-intensive, and highly subjective process. Therefore, automatic diagnosis of benign and ma-lignant thyroid nodules is beneficial for the clinical diagnosis of the thyroid. To this end, based on three mo-dalities of gray-scale ultrasound images(US), SWE, and CDUS, we propose a deep learning-based multi-modal feature fusion network for the automatic diagnosis of thyroid disease based on the ultrasound images. First, three ResNet18s initialized by self-supervised learning are used as branches to extract the image information of each modality, respectively. Then, a multi-modal multi-head attention branch is used to remove the common infor-mation of three modalities, and the knowledge of each modal is combined for thyroid diagnosis. At the same time, to better integrate the features between modalities, a multi-modal feature guidance module is also pro-posed to guide the feature extraction of each branch and reduce the difference between each-modal feature. We verify the multi-modal thyroid ultrasound image diagnosis method on the self-collected dataset, and the results prove that this method could provide fast and accurate assistance for sonographers in diagnosing thyroid nodules.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Self-Supervised Multi-Modal Hybrid Fusion Network for Brain Tumor Segmentation
    Fang, Feiyi
    Yao, Yazhou
    Zhou, Tao
    Xie, Guosen
    Lu, Jianfeng
    [J]. IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (11) : 5310 - 5320
  • [2] SS-SSAN: a self-supervised subspace attentional network for multi-modal medical image fusion
    Ying Zhang
    Rencan Nie
    Jinde Cao
    Chaozhen Ma
    Chengchao Wang
    [J]. Artificial Intelligence Review, 2023, 56 : 421 - 443
  • [3] SS-SSAN: a self-supervised subspace attentional network for multi-modal medical image fusion
    Zhang, Ying
    Nie, Rencan
    Cao, Jinde
    Ma, Chaozhen
    Wang, Chengchao
    [J]. ARTIFICIAL INTELLIGENCE REVIEW, 2023, 56 (SUPPL 1) : 421 - 443
  • [4] Self-Supervised Distilled Learning for Multi-modal Misinformation Identification
    Mu, Michael
    Das Bhattacharjee, Sreyasee
    Yuan, Junsong
    [J]. 2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 2818 - 2827
  • [5] Self-supervised Multi-Modal Video Forgery Attack Detection
    Zhao, Chenhui
    Li, Xiang
    Younes, Rabih
    [J]. 2023 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC, 2023,
  • [6] Self-supervised opinion summarization with multi-modal knowledge graph
    Lingyun Jin
    Jingqiang Chen
    [J]. Journal of Intelligent Information Systems, 2024, 62 : 191 - 208
  • [7] Self-supervised opinion summarization with multi-modal knowledge graph
    Jin, Lingyun
    Chen, Jingqiang
    [J]. JOURNAL OF INTELLIGENT INFORMATION SYSTEMS, 2024, 62 (01) : 191 - 208
  • [8] Multi-modal Fusion
    Liu, Huaping
    Hussain, Amir
    Wang, Shuliang
    [J]. INFORMATION SCIENCES, 2018, 432 : 462 - 462
  • [9] Self-supervised 3D Patient Modeling with Multi-modal Attentive Fusion
    Zheng, Meng
    Planche, Benjamin
    Gong, Xuan
    Yang, Fan
    Chen, Terrence
    Wu, Ziyan
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT VII, 2022, 13437 : 115 - 125
  • [10] Multi-modal emotion recognition using tensor decomposition fusion and self-supervised multi-tasking
    Wang, Rui
    Zhu, Jiawei
    Wang, Shoujin
    Wang, Tao
    Huang, Jingze
    Zhu, Xianxun
    [J]. INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2024, 13 (04)