DFECF-DET: All-Weather Detector Based on Differential Feature Enhancement and Cross-Modal Fusion With Visible and Infrared Sensors

被引:4
|
作者
Wang, Chuanyun [1 ]
Sun, Dongdong [1 ]
Yang, Jianqi [1 ]
Li, Zhaokui [2 ]
Gao, Qian [1 ]
机构
[1] Shenyang Aerosp Univ, Coll Artificial Intelligence, Shenyang 110136, Peoples R China
[2] Shenyang Aerosp Univ, Sch Comp Sci, Shenyang 110136, Peoples R China
基金
中国国家自然科学基金;
关键词
All-weather detection; cross-modal fusion (CF); differential feature enhancement (DFE); lite-transformer; multisource sensors; IMAGE FUSION;
D O I
10.1109/JSEN.2023.3324451
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In complex environments such as night, fog, and battlefield camouflage, a single camera sensor is not sufficient to reflect scene information, and a multisource sensor can raise environmental awareness. A visible light camera with an infrared sensor is an efficient combination. However, there are huge differences in the inputs from different sensors, and how to fuse the information from two sensors and apply it to a specific task is a problem that needs to be solved. Hence, a multisource input detection algorithm with the combination of infrared sensors and visible cameras is proposed in this article. Its purpose is to solve the problem of low detection accuracy of a single sensor in complex and changing environments. First, a differential feature enhancement (DFE) module to enhance features that are constantly degraded during network transmission is designed in this article. Second, a cross-modal fusion (CF) module to fuse features from multiple sources is designed. Finally, two modules are embedded in a two-stream network. Experiments on the publicly available FLIR and LLVIP datasets show that the algorithm in this article improves mAP75 by 8.3/4.3 compared to a single-source detector. In some special environments, the module in this article uses 0.1 MB of storage at the cost of a 1.7 mAP boost! Extensive ablation experiments demonstrate that the module proposed in this article is lightweight, efficient, and plug-and-play.
引用
收藏
页码:29200 / 29210
页数:11
相关论文
共 7 条
  • [1] Infrared and visible image fusion based on cross-modal extraction strategy
    Liu, Xiaowen
    Li, Jing
    Yang, Xin
    Huo, Hongtao
    INFRARED PHYSICS & TECHNOLOGY, 2022, 124
  • [2] BCMFIFuse: A Bilateral Cross-Modal Feature Interaction-Based Network for Infrared and Visible Image Fusion
    Gao, Xueyan
    Liu, Shiguang
    REMOTE SENSING, 2024, 16 (17)
  • [3] CMFA_Net: A cross-modal feature aggregation network for infrared-visible image fusion
    Ding, Zhaisheng
    Li, Haiyan
    Zhou, Dongming
    Li, Hongsong
    Liu, Yanyu
    Hou, Ruichao
    INFRARED PHYSICS & TECHNOLOGY, 2021, 118
  • [4] MCFusion: infrared and visible image fusion based multiscale receptive field and cross-modal enhanced attention mechanism
    Jiang, Min
    Wang, Zhiyuan
    Kong, Jun
    Zhuang, Danfeng
    JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (01)
  • [5] Progressive learning in cross-modal cross-scale fusion transformer for visible-infrared video-based person reidentification
    Mukhtar, Hamza
    Mukhtar, Umar Raza
    Knowledge-Based Systems, 2024, 304
  • [6] CEFusion: An Infrared and Visible Image Fusion Network Based on Cross-Modal Multi-Granularity Information Interaction and Edge Guidance
    Yang, Bin
    Hu, Yuxuan
    Liu, Xiaowen
    Li, Jing
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, : 17794 - 17809
  • [7] Cross-Modal Feature Fusion and Interaction Strategy for CNN-Transformer-Based Object Detection in Visual and Infrared Remote Sensing Imagery
    Nie, Jinyan
    Sun, He
    Sun, Xu
    Ni, Li
    Gao, Lianru
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5