Co-Enhancement of Multi-Modality Image Fusion and Object Detection via Feature Adaptation

被引:0
|
作者
Dong, Aimei [1 ,2 ,3 ]
Wang, Long [2 ]
Liu, Jian [2 ]
Xu, Jingyuan [2 ]
Zhao, Guixin [1 ,2 ,3 ]
Zhai, Yi [1 ,2 ,3 ]
Lv, Guohua [1 ,2 ,3 ]
Cheng, Jinyong [1 ,2 ,3 ]
机构
[1] Qilu Univ Technol, Shandong Acad Sci, Shandong Comp Sci Ctr, Natl Supercomp Ctr Jinan,Minist Educ,Key Lab Comp, Jinan 250316, Peoples R China
[2] Qilu Univ Technol, Shandong Acad Sci, Fac Comp Sci & Technol, Jinan 250316, Peoples R China
[3] Shandong Fundamental Res Ctr Comp Sci, Shandong Prov Key Lab Comp Networks, Jinan 250100, Peoples R China
关键词
Image fusion; Task analysis; Semantics; Feature extraction; Object detection; Visualization; Visual perception; object detection; feature adaptation; mutual promotion; MULTISCALE TRANSFORM; NETWORK; NEST;
D O I
10.1109/TCSVT.2024.3433555
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The integration of multi-modality images significantly enhances the clarity of critical details for object detection. Valuable semantic data from object detection enriches the fusion process of these images. However, the potential reciprocal relationship that could enhance their mutual performance remains largely unexplored and underutilized, despite some semantic-driven fusion methodologies catering to specific application needs. To address these limitations, this study proposes a mutually reinforcing, dual-task-driven fusion architecture. Specifically, our design integrates a feature-adaptive interlinking module into both image fusion and object detection components, effectively managing the inherent feature discrepancies. The core idea is to channel distinct features from both tasks into a unified feature space after feature transformation. We then design a feature-adaptive selection module to generate features rich in target semantic information and compatible with the fusion network. Finally, effective combination and mutual enhancement of the two tasks are achieved through an alternating training process. A diverse range of swift evaluations is performed across various datasets to corroborate the potential efficiency of our framework, actualizing visible advancements in both fusion effectiveness and detection accuracy.
引用
收藏
页码:12624 / 12637
页数:14
相关论文
共 50 条
  • [31] A novel dictionary learning approach for multi-modality medical image fusion
    Zhu, Zhiqin
    Chai, Yi
    Yin, Hongpeng
    Li, Yanxia
    Liu, Zhaodong
    NEUROCOMPUTING, 2016, 214 : 471 - 482
  • [32] Enhancement and Fusion of Multi-Scale Feature Maps for Small Object Detection
    Xue, Zhijun
    Chen, Wenjie
    Li, Jing
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 7212 - 7217
  • [33] Multi-Modality Medical Image Fusion using Discrete Wavelet Transform
    Bhavana, V
    Krishnappa, H. K.
    PROCEEDINGS OF THE 4TH INTERNATIONAL CONFERENCE ON ECO-FRIENDLY COMPUTING AND COMMUNICATION SYSTEMS, 2015, 70 : 625 - 631
  • [34] Evidence Fusion with Contextual Discounting for Multi-modality Medical Image Segmentation
    Huang, Ling
    Denoeux, Thierry
    Vera, Pierre
    Ruan, Su
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT V, 2022, 13435 : 401 - 411
  • [35] Research on image affection tagging based on multi-modality information fusion
    Tang Z.
    Liu X.
    Yang H.
    Lu C.
    Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2020, 26 (01): : 134 - 144
  • [36] Multi-modality image fusion combining sparse representation with guidance filtering
    Hu, Qiu
    Hu, Shaohai
    Zhang, Fengzhen
    SOFT COMPUTING, 2021, 25 (06) : 4393 - 4407
  • [37] Underwater image enhancement via color correction and multi-feature image fusion
    Ke, Ke
    Zhang, Biyun
    Zhang, Chunmin
    Yao, Baoli
    Guo, Shiping
    Tang, Feng
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2024, 35 (09)
  • [38] A novel multi-modality image fusion method based on image decomposition and sparse representation
    Zhu, Zhiqin
    Yin, Hongpeng
    Chai, Yi
    Li, Yanxia
    Qi, Guanqiu
    INFORMATION SCIENCES, 2018, 432 : 516 - 529
  • [39] Navigating an Automated Driving Vehicle via the Early Fusion of Multi-Modality
    Haris, Malik
    Glowacz, Adam
    SENSORS, 2022, 22 (04)
  • [40] EFGNet: Encoder steered multi-modality feature guidance network for RGB-D salient object detection
    Xia, Chenxing
    Duan, Songsong
    Fang, Xianjin
    Gao, Xiuju
    Sun, Yanguang
    Ge, Bin
    Zhang, Hanling
    Li, Kuan-Ching
    DIGITAL SIGNAL PROCESSING, 2022, 131