Attention-based Fusion Network for Breast Cancer Segmentation and Classification Using Multi-modal Ultrasound Images

被引:1
|
作者
Cho, Yoonjae [1 ,2 ,3 ]
Misra, Sampa [1 ,2 ,3 ]
Managuli, Ravi [4 ]
Barr, Richard G. [5 ]
Lee, Jeongmin [6 ,7 ]
Kim, Chulhong [1 ,2 ,3 ,8 ]
机构
[1] Pohang Univ Sci & Technol, Med Device Innovat Ctr, Mech Engn, Convergence IT Engn,Dept Elect Engn, Pohang 37673, South Korea
[2] Pohang Univ Sci & Technol, Grad Sch Artificial Intelligence, Pohang 37673, South Korea
[3] Pohang Univ Sci & Technol, Med Device Innovat Ctr, Pohang, South Korea
[4] Univ Washington, Dept Bioengn, Seattle, WA USA
[5] Southwoods Imaging, Youngstown, OH USA
[6] Sungkyunkwan Univ, Sch Med, Dept Radiol, Seoul, South Korea
[7] Sungkyunkwan Univ, Ctr Imaging Sci, Samsung Med Ctr, Sch Med, Seoul, South Korea
[8] Opticho Inc, Pohang, South Korea
来源
ULTRASOUND IN MEDICINE AND BIOLOGY | 2025年 / 51卷 / 03期
基金
新加坡国家研究基金会;
关键词
Breast cancer; Breast ultrasound images; Multi-modality; Classification; Segmentation; Transfer learning; BENIGN;
D O I
10.1016/j.ultrasmedbio.2024.11.020
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Objective: Breast cancer is one of the most commonly occurring cancers in women. Thus, early detection and treatment of cancer lead to a better outcome for the patient. Ultrasound (US) imaging plays a crucial role in the early detection of breast cancer, providing a cost-effective, convenient, and safe diagnostic approach. To date, much research has been conducted to facilitate reliable and effective early diagnosis of breast cancer through US image analysis. Recently, with the introduction of machine learning technologies such as deep learning (DL), automated lesion segmentation and classification, the identification of malignant masses in US breasts has progressed, and computer-aided diagnosis (CAD) technology is being applied in clinics effectively. Herein, we propose a novel deep learning-based "segmentation + classification" model based on B- and SE-mode images. Methods: For the segmentation task, we propose a Multi-Modal Fusion U-Net (MMF-U-Net), which segments lesions by mixing B- and SE-mode information through fusion blocks. After segmenting, the lesion area from the B- and SE-mode images is cropped using a predicted segmentation mask. The encoder part of the pre-trained MMF-U-Net model is then used on the cropped B- and SE-mode breast US images to classify benign and malignant lesions. Results: The experimental results using the proposed method showed good segmentation and classification scores. The dice score, intersection over union (IoU), precision, and recall are 78.23%, 68.60%, 82.21%, and 80.58%, respectively, using the proposed MMF-U-Net on real-world clinical data. The classification accuracy is 98.46%. Conclusion: Our results show that the proposed method will effectively segment the breast lesion area and can reliably classify the benign from malignant lesions.
引用
收藏
页码:568 / 577
页数:10
相关论文
共 50 条
  • [1] Attention-Based Multi-Modal Fusion Network for Semantic Scene Completion
    Li, Siqi
    Zou, Changqing
    Li, Yipeng
    Zhao, Xibin
    Gao, Yue
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11402 - 11409
  • [2] Attention-based multi-modal fusion sarcasm detection
    Liu, Jing
    Tian, Shengwei
    Yu, Long
    Long, Jun
    Zhou, Tiejun
    Wang, Bo
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 44 (02) : 2097 - 2108
  • [3] An Attention-Based Residual U-Net for Tumour Segmentation Using Multi-Modal MRI Brain Images
    Naqvi, Najme Zehra
    Seeja, K. R.
    IEEE ACCESS, 2025, 13 : 10240 - 10251
  • [4] Cross-modal attention network for retinal disease classification based on multi-modal images
    Liu, Zirong
    Hu, Yan
    Qiu, Zhongxi
    Niu, Yanyan
    Zhou, Dan
    Li, Xiaoling
    Shen, Junyong
    Jiang, Hongyang
    Li, Heng
    Liu, Jiang
    BIOMEDICAL OPTICS EXPRESS, 2024, 15 (06): : 3699 - 3714
  • [5] A Tri-Attention fusion guided multi-modal segmentation network
    Zhou, Tongxue
    Ruan, Su
    Vera, Pierre
    Canu, Stephane
    PATTERN RECOGNITION, 2022, 124
  • [6] ARF-Net: a multi-modal aesthetic attention-based fusion
    Iffath, Fariha
    Gavrilova, Marina
    VISUAL COMPUTER, 2024, 40 (07): : 4941 - 4953
  • [7] EISNet: A Multi-Modal Fusion Network for Semantic Segmentation With Events and Images
    Xie, Bochen
    Deng, Yongjian
    Shao, Zhanpeng
    Li, Youfu
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 8639 - 8650
  • [8] UNIVERSAL MULTI-MODAL DEEP NETWORK FOR CLASSIFICATION AND SEGMENTATION OF MEDICAL IMAGES
    Harouni, Ahmed
    Karargyris, Alexandros
    Negahdar, Mohammadreza
    Beymer, David
    Syeda-Mahmood, Tanveer
    2018 IEEE 15TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2018), 2018, : 872 - 876
  • [9] An auxiliary attention-based network for joint classification and localization of breast tumor on ultrasound images
    Fan, Zong
    Gong, Ping
    Zhang, Xiaohui
    Wang, Zhimin
    Hao, Yao
    Song, Pengfei
    Chen, Shigao
    Li, Hua
    MEDICAL IMAGING 2023, 2023, 12464
  • [10] Breast Cancer Segmentation From Ultrasound Images Using Multiscale Cascaded Convolution With Residual Attention-Based Double Decoder Network
    Umer, Muhammad Junaid
    Sharif, Muhammad Irfan
    Kim, Jungeun
    IEEE ACCESS, 2024, 12 : 107888 - 107902