Contrastive fusion representation learning for foreground object detection

被引:2
|
作者
Wang, Pei [1 ]
Wu, Junsheng [2 ]
Fang, Aiqing [1 ]
Zhu, Zhixiang
Wang, Chenwu [1 ,3 ]
Mu, Pengyuan [4 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci & Engn, Xian 710072, Peoples R China
[2] Northwestern Polytech Univ, Sch Software, Xian 710072, Peoples R China
[3] Xian Univ Posts & Telecommun, Sch Modern Post, Xian 710061, Peoples R China
[4] Shannxi Informat Engn Res Inst, Data Operat Dept, Xian 710065, Peoples R China
关键词
Foreground moving object detection; Feature fusion; Mutual information theory; Deep learning; VISIBLE IMAGE FUSION; GENERATIVE ADVERSARIAL NETWORK; BACKGROUND SUBTRACTION; MULTISCALE TRANSFORM; FRAMEWORK; MODEL; NEST;
D O I
10.1016/j.engappai.2023.106453
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper investigates the problem of an effective and robust fusion representation method for foreground moving object detection. Many fusion representation learning approaches pay attention to the similarity measurement between the fusion results and source images in texture details and pixel intensity, ignoring the harmful information, e.g., noise, blur, and extreme illumination. Therefore, the aggregated features of infrared and visible images will introduce much harmful information, affecting the model performance of downstream visual tasks. This paper tackles these problems by proposing a contrastive fusion representation learning method for the foreground moving object detection task, which consists of two major modules: the upstream fusion representation module (FRM) and the downstream foreground moving object detection module (FODM). Unlike the traditional fusion optimization mechanism, the former aims to extract valuable features and reject harmful features via the maximum mutual information theory. The latter is a siamese convolutional neural network to detect foreground moving objects by aggregating the time-sequence images generated by FRM. Experimental results and comparisons with the state-of-the-art on three public datasets (i.e., TNO, MF, and cross-modal FOD dataset of infrared and visible images), validate the effectiveness, robustness, and overall superiority of the proposed contrast fusion representation learning method. Concisely, our contrastive fusion representation learning has gained 53.9%, 43.2%, 46.4%, 52.3%, 2.2%, 87.1%, 3.5% on EI, SF, DF, AG, MI, and Nabf metrics compared with the best competitors.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Fusion representation learning for foreground moving object detection
    Wang, Pei
    Wu, Junsheng
    Fang, Aiqing
    Zhu, Zhixiang
    Wang, Chenwu
    Ren, Shan
    [J]. DIGITAL SIGNAL PROCESSING, 2023, 138
  • [2] SCDet: decoupling discriminative representation for dark object detection via supervised contrastive learning
    Tongxu Lin
    Guoheng Huang
    Xiaochen Yuan
    Guo Zhong
    Xiaocong Huang
    Chi-Man Pun
    [J]. The Visual Computer, 2024, 40 : 3357 - 3369
  • [3] SCDet: decoupling discriminative representation for dark object detection via supervised contrastive learning
    Lin, Tongxu
    Huang, Guoheng
    Yuan, Xiaochen
    Zhong, Guo
    Huang, Xiaocong
    Pun, Chi-Man
    [J]. VISUAL COMPUTER, 2024, 40 (05): : 3357 - 3369
  • [4] DetCo: Unsupervised Contrastive Learning for Object Detection
    Xie, Enze
    Ding, Jian
    Wang, Wenhai
    Zhan, Xiaohang
    Xu, Hang
    Sun, Peize
    Li, Zhenguo
    Luo, Ping
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 8372 - 8381
  • [5] Stopped Object Detection by Learning Foreground Model in Videos
    Maddalena, Lucia
    Petrosino, Alfredo
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2013, 24 (05) : 723 - 735
  • [6] Object-aware Contrastive Learning for Debiased Scene Representation
    Mo, Sangwoo
    Kang, Hyunwoo
    Sohn, Kihyuk
    Li, Chun-Liang
    Shin, Jinwoo
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [7] FOREGROUND DETECTION ON DEPTH MAPS USING SKELETAL REPRESENTATION OF OBJECT SILHOUETTES
    Beloborodov, D.
    Mestetskiy, L.
    [J]. INTERNATIONAL WORKSHOP PHOTOGRAMMETRIC AND COMPUTER VISION TECHNIQUES FOR VIDEO SURVEILLANCE, BIOMETRICS AND BIOMEDICINE, 2017, 42-2 (W4): : 7 - 11
  • [8] Object Discovery via Contrastive Learning for Weakly Supervised Object Detection
    Seo, Jinhwan
    Bae, Wonho
    Sutherland, Danica J.
    Noh, Junhyug
    Kim, Daijin
    [J]. COMPUTER VISION, ECCV 2022, PT XXXI, 2022, 13691 : 312 - 329
  • [9] Joint learning of foreground, background and edge for salient object detection
    Wu, Qin
    Zhu, Pengcheng
    Chai, Zhilei
    Guo, Guodong
    [J]. COMPUTER VISION AND IMAGE UNDERSTANDING, 2024, 240
  • [10] Improving the Adversarial Robustness of Object Detection with Contrastive Learning
    Zeng, Weiwei
    Gao, Song
    Zhou, Wei
    Dong, Yunyun
    Wang, Ruxin
    [J]. PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT IX, 2024, 14433 : 29 - 40