Joint image and feature adaptative attention-aware networks for cross-modality semantic segmentation

被引:3
|
作者
Zhong, Qihuang [1 ,2 ,3 ]
Zeng, Fanzhou [1 ]
Liao, Fei [1 ]
Liu, Juhua [2 ,3 ]
Du, Bo [3 ,4 ,5 ]
Shang, Jedi S. [6 ]
机构
[1] Wuhan Univ, Renmin Hosp, Dept Gastroenterol, Wuhan, Peoples R China
[2] Wuhan Univ, Sch Printing & Packaging, Wuhan, Peoples R China
[3] Wuhan Univ, Inst Artificial Intelligence, Natl Engn Res Ctr Multimedia Software, Wuhan, Peoples R China
[4] Wuhan Univ, Sch Comp Sci, Wuhan, Peoples R China
[5] Wuhan Univ, Hubei Key Lab Multimedia & Network Commun Engn, Wuhan, Peoples R China
[6] Thinvent Technol Co LTD, Nanchang, Jiangxi, Peoples R China
来源
NEURAL COMPUTING & APPLICATIONS | 2023年 / 35卷 / 05期
基金
中国国家自然科学基金;
关键词
Domain adaptation; Attention; Cross-modality; Semantic segmentation; AUTOMATED SEGMENTATION; PATCH;
D O I
10.1007/s00521-021-06064-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning-based methods are widely used for the task of semantic segmentation in recent years. However, due to the difficulty and labor cost of collecting pixel-level annotations, it is hard to acquire sufficient training images for a certain imaging modality, which greatly hinders the performance of these methods. The intuitive solution to this issue is to train a pre-trained model on label-rich imaging modality (source domain) and then apply the pre-trained model to the label-poor imaging modality (target domain). Unsurprisingly, since the severe domain shift between different modalities, the pre-trained model would perform poorly on the target imaging modality. To this end, we propose a novel unsupervised domain adaptation framework, called Joint Image and Feature Adaptive Attention-aware Networks (JIFAAN), to alleviate the domain shift for cross-modality semantic segmentation. The proposed framework mainly consists of two procedures. The first procedure is image adaptation, which transforms the source domain images into target-like images using the adversarial learning with cycle-consistency constraint. For further bridging the gap between transformed images and target domain images, the second procedure employs feature adaptation to extract the domain-invariant features and thus aligns the distribution in feature space. In particular, we introduce an attention module in the feature adaptation to focus on noteworthy regions and generate attention-aware results. Lastly, we combine two procedures in an end-to-end manner. Experiments on two cross-modality semantic segmentation datasets demonstrate the effectiveness of our proposed framework. Specifically, JIFAAN surpasses the cutting-edge domain adaptation methods and achieves the state-of-the-art performance.
引用
收藏
页码:3665 / 3676
页数:12
相关论文
共 50 条
  • [1] Joint image and feature adaptative attention-aware networks for cross-modality semantic segmentation
    Qihuang Zhong
    Fanzhou Zeng
    Fei Liao
    Juhua Liu
    Bo Du
    Jedi S. Shang
    Neural Computing and Applications, 2023, 35 : 3665 - 3676
  • [2] Attention-aware ensemble learning for face-periocular cross-modality matching
    Ng, Tiong-Sik
    Teoh, Andrew Beng Jin
    APPLIED SOFT COMPUTING, 2025, 175
  • [3] Partial Unbalanced Feature Transport for Cross-Modality Cardiac Image Segmentation
    Dong, Shunjie
    Pan, Zixuan
    Fu, Yu
    Xu, Dongwei
    Shi, Kuangyu
    Yang, Qianqian
    Shi, Yiyu
    Zhuo, Cheng
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2023, 42 (06) : 1758 - 1773
  • [4] CMANet: Cross-Modality Attention Network for Indoor-Scene Semantic Segmentation
    Zhu, Longze
    Kang, Zhizhong
    Zhou, Mei
    Yang, Xi
    Wang, Zhen
    Cao, Zhen
    Ye, Chenming
    SENSORS, 2022, 22 (21)
  • [5] Semantic Consistent Unsupervised Domain Adaptation for Cross-Modality Medical Image Segmentation
    Zeng, Guodong
    Lerch, Till D.
    Schmaranzer, Florian
    Zheng, Guoyan
    Burger, Juergen
    Gerber, Kate
    Tannast, Moritz
    Siebenrock, Klaus
    Gerber, Nicolas
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT III, 2021, 12903 : 201 - 210
  • [6] RGB-D Domain adaptive semantic segmentation with cross-modality feature recalibration
    Fan, Qizhe
    Shen, Xiaoqin
    Ying, Shihui
    Wang, Juan
    Du, Shaoyi
    INFORMATION FUSION, 2025, 120
  • [7] CCANet: Cross-Modality Comprehensive Feature Aggregation Network for Indoor Scene Semantic Segmentation
    Zhang, Zihao
    Yang, Yale
    Hou, Huifang
    Meng, Fanman
    Zhang, Fan
    Xie, Kangzhan
    Zhuang, Chunsheng
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2025, 17 (02) : 366 - 378
  • [8] Attention-Aware and Semantic-Aware Network for RGB-D Indoor Semantic Segmentation
    Duan L.-J.
    Sun Q.-C.
    Qiao Y.-H.
    Chen J.-C.
    Cui G.-Q.
    Jisuanji Xuebao/Chinese Journal of Computers, 2021, 44 (02): : 275 - 291
  • [9] Synergistic Image and Feature Adaptation: Towards Cross-Modality Domain Adaptation for Medical Image Segmentation
    Chen, Cheng
    Dou, Qi
    Chen, Hao
    Qin, Jing
    Heng, Pheng-Ann
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 865 - 872
  • [10] A Novel Semantic CT Segmentation Algorithm Using Boosted Attention-Aware Convolutional Neural Networks
    Kearney, V.
    Chan, J.
    Wang, T.
    Perry, A.
    Yom, S.
    Solberg, T.
    MEDICAL PHYSICS, 2019, 46 (06) : E370 - E370