Joint image and feature adaptative attention-aware networks for cross-modality semantic segmentation

被引:3
|
作者
Zhong, Qihuang [1 ,2 ,3 ]
Zeng, Fanzhou [1 ]
Liao, Fei [1 ]
Liu, Juhua [2 ,3 ]
Du, Bo [3 ,4 ,5 ]
Shang, Jedi S. [6 ]
机构
[1] Wuhan Univ, Renmin Hosp, Dept Gastroenterol, Wuhan, Peoples R China
[2] Wuhan Univ, Sch Printing & Packaging, Wuhan, Peoples R China
[3] Wuhan Univ, Inst Artificial Intelligence, Natl Engn Res Ctr Multimedia Software, Wuhan, Peoples R China
[4] Wuhan Univ, Sch Comp Sci, Wuhan, Peoples R China
[5] Wuhan Univ, Hubei Key Lab Multimedia & Network Commun Engn, Wuhan, Peoples R China
[6] Thinvent Technol Co LTD, Nanchang, Jiangxi, Peoples R China
来源
NEURAL COMPUTING & APPLICATIONS | 2023年 / 35卷 / 05期
基金
中国国家自然科学基金;
关键词
Domain adaptation; Attention; Cross-modality; Semantic segmentation; AUTOMATED SEGMENTATION; PATCH;
D O I
10.1007/s00521-021-06064-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning-based methods are widely used for the task of semantic segmentation in recent years. However, due to the difficulty and labor cost of collecting pixel-level annotations, it is hard to acquire sufficient training images for a certain imaging modality, which greatly hinders the performance of these methods. The intuitive solution to this issue is to train a pre-trained model on label-rich imaging modality (source domain) and then apply the pre-trained model to the label-poor imaging modality (target domain). Unsurprisingly, since the severe domain shift between different modalities, the pre-trained model would perform poorly on the target imaging modality. To this end, we propose a novel unsupervised domain adaptation framework, called Joint Image and Feature Adaptive Attention-aware Networks (JIFAAN), to alleviate the domain shift for cross-modality semantic segmentation. The proposed framework mainly consists of two procedures. The first procedure is image adaptation, which transforms the source domain images into target-like images using the adversarial learning with cycle-consistency constraint. For further bridging the gap between transformed images and target domain images, the second procedure employs feature adaptation to extract the domain-invariant features and thus aligns the distribution in feature space. In particular, we introduce an attention module in the feature adaptation to focus on noteworthy regions and generate attention-aware results. Lastly, we combine two procedures in an end-to-end manner. Experiments on two cross-modality semantic segmentation datasets demonstrate the effectiveness of our proposed framework. Specifically, JIFAAN surpasses the cutting-edge domain adaptation methods and achieves the state-of-the-art performance.
引用
收藏
页码:3665 / 3676
页数:12
相关论文
共 50 条
  • [31] PSIGAN: Joint Probabilistic Segmentation and Image Distribution Matching for Unpaired Cross-Modality Adaptation-Based MRI Segmentation
    Jiang, Jue
    Hu, Yu-Chi
    Tyagi, Neelam
    Rimner, Andreas
    Lee, Nancy
    Deasy, Joseph O.
    Berry, Sean
    Veeraraghavan, Harini
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2020, 39 (12) : 4071 - 4084
  • [32] Cross-modality semantic guidance for multi-label image classification
    Huang, Jun
    Wang, Dian
    Hong, Xudong
    Qu, Xiwen
    Xue, Wei
    INTELLIGENT DATA ANALYSIS, 2024, 28 (03) : 633 - 646
  • [33] Rethinking Semantic Image Compression: Scalable Representation With Cross-Modality Transfer
    Zhang, Pingping
    Wang, Shiqi
    Wang, Meng
    Li, Jiguo
    Wang, Xu
    Kwong, Sam
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (08) : 4441 - 4445
  • [34] Liver Segmentation via Learning Cross-Modality Content-Aware Representation
    Lin, Xingxiao
    Ji, Zexuan
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT XIII, 2024, 14437 : 198 - 208
  • [35] Cross-Modality Attention with Semantic Graph Embedding for Multi-Label Classification
    You, Renchun
    Guo, Zhiyao
    Cui, Lei
    Long, Xiang
    Bao, Yingze
    Wen, Shilei
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 12709 - 12716
  • [36] Cross-Modality Features Fusion for Synthetic Aperture Radar Image Segmentation
    Gao, Fei
    Huang, Heqing
    Yue, Zhenyu
    Li, Dongyu
    Ge, Shuzhi Sam
    Lee, Tong Heng
    Zhou, Huiyu
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [37] Diverse data augmentation for learning image segmentation with cross-modality annotations
    Chen, Xu
    Lian, Chunfeng
    Wang, Li
    Deng, Hannah
    Kuang, Tianshu
    Fung, Steve H.
    Gateno, Jaime
    Shen, Dinggang
    Xia, James J.
    Yap, Pew-Thian
    MEDICAL IMAGE ANALYSIS, 2021, 71
  • [38] Data Efficient Unsupervised Domain Adaptation For Cross-modality Image Segmentation
    Ouyang, Cheng
    Kamnitsas, Konstantinos
    Biffi, Carlo
    Duan, Jinming
    Rueckert, Daniel
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT II, 2019, 11765 : 669 - 677
  • [39] Deep Symmetric Adaptation Network for Cross-Modality Medical Image Segmentation
    Han, Xiaoting
    Qi, Lei
    Yu, Qian
    Zhou, Ziqi
    Zheng, Yefeng
    Shi, Yinghuan
    Gao, Yang
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2022, 41 (01) : 121 - 132
  • [40] Unpaired Cross-Modality Educed Distillation (CMEDL) for Medical Image Segmentation
    Jiang, Jue
    Rimner, Andreas
    Deasy, Joseph O.
    Veeraraghavan, Harini
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2022, 41 (05) : 1057 - 1068