Cross-modality complementary information fusion for multispectral pedestrian detection

被引:12
|
作者
Yan, Chaoqi [1 ]
Zhang, Hong [1 ]
Li, Xuliang [1 ]
Yang, Yifan [2 ]
Yuan, Ding [1 ]
机构
[1] Beihang Univ, Image Proc Ctr, 37 Xueyuan Rd, Beijing 100191, Peoples R China
[2] Beihang Univ, Inst Artificial Intelligence, 37 Xueyuan Rd, Beijing 100191, Peoples R China
来源
NEURAL COMPUTING & APPLICATIONS | 2023年 / 35卷 / 14期
基金
中国国家自然科学基金;
关键词
Multispectral pedestrian detection; Cross-modality; Information fusion; Illumination-aware; Feature alignment; DEEP NEURAL-NETWORKS;
D O I
10.1007/s00521-023-08239-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multispectral pedestrian detection has received increasing attention in recent years as color and thermal modalities can provide complementary visual information, especially under insufficient illumination conditions. However, there is still a persistent crucial problem that how to design the cross-modality fusion mechanism to fully exploit the complementary characteristics between different modalities. In this paper, we propose a novel cross-modality complementary information fusion network (denoted as CCIFNet) to comprehensively capture the long-range interactions with precise positional information and meanwhile preserve the inter-spatial relationship between different modalities in the feature extraction stage. Further, we design an adaptive illumination-aware weight generation module to adaptively weight the final detection confidence of color and thermal modalities by taking various illumination conditions into consideration. Specifically, we comprehensively compare three different fusion strategies about this module to synthetically explore the best way for generating the final illumination-aware fusion weights. Finally, we present a simple but effective feature alignment module to alleviate the position shift problem caused by the weakly aligned color-thermal image pairs. Extensive experiments and ablation studies on KAIST, CVC-14, FLIR and LLVIP multispectral object detection datasets show that the proposed CCIFNet can achieve state-of-the-art performance under different illumination evaluation settings, while keeping a competitive speed-accuracy trade-off for real-time applications.
引用
收藏
页码:10361 / 10386
页数:26
相关论文
共 50 条
  • [31] Cross-Modal Attentive Recalibration and Dynamic Fusion for Multispectral Pedestrian Detection
    Bao, Wei
    Hu, Jingjing
    Huang, Meiyu
    Xiang, Xueshuang
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT I, 2024, 14425 : 499 - 510
  • [32] Cross-Modality Object Detection Based on DETR
    Huang, Xinyi
    Ma, Guochun
    IEEE ACCESS, 2025, 13 : 51220 - 51230
  • [33] Pedestrian Recognition Using Cross-Modality Learning in Convolutional Neural Networks
    Pop, Danut Ovidiu
    Rogozan, Alexandrina
    Nashashibi, Fawzi
    Bensrhair, Abdelaziz
    IEEE INTELLIGENT TRANSPORTATION SYSTEMS MAGAZINE, 2021, 13 (01) : 210 - 224
  • [34] Pedestrian Recognition through Different Cross-Modality Deep Learning Methods
    Pop, Danut Ovidiu
    Rogozan, Alexandrina
    Nashashibi, Fawzi
    Bensrhair, Abdelaziz
    2017 IEEE INTERNATIONAL CONFERENCE ON VEHICULAR ELECTRONICS AND SAFETY (ICVES), 2017, : 133 - 138
  • [35] Cross-Modality Proposal-Guided Feature Mining for Unregistered RGB-Thermal Pedestrian Detection
    Tian, Chao
    Zhou, Zikun
    Huang, Yuqing
    Li, Gaojun
    He, Zhenyu
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 6449 - 6461
  • [36] 2-dimensional fusion of cerebral cross-modality images employing a mutual information algorithm
    Kollmann, C
    Greiffenberg, B
    Schlachetzki, F
    Bogdahn, U
    Bergmann, H
    PHYSICA MEDICA-EUROPEAN JOURNAL OF MEDICAL PHYSICS, 2001, 17 (04): : 267 - 270
  • [37] STIMULUS INFORMATION AND SEQUENTIAL DEPENDENCIES IN CROSS-MODALITY MATCHING
    WARD, LM
    BULLETIN OF THE PSYCHONOMIC SOCIETY, 1977, 10 (04) : 275 - 275
  • [38] FDENet: Fusion Depth Semantics and Edge-Attention Information for Multispectral Pedestrian Detection
    Liu, Xiaowei
    Xu, Xinying
    Xie, Jun
    Li, Pengyue
    Wei, Jiamin
    Sang, Yiyu
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (06) : 5441 - 5448
  • [39] Facial Expression Recognition Through Cross-Modality Attention Fusion
    Ni, Rongrong
    Yang, Biao
    Zhou, Xu
    Cangelosi, Angelo
    Liu, Xiaofeng
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2023, 15 (01) : 175 - 185
  • [40] CROSS-MODALITY MATCHING
    AUERBACH, C
    QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 1973, 25 (NOV): : 492 - 495