Multi-Modal Object Tracking and Image Fusion With Unsupervised Deep Learning

被引:11
|
作者
LaHaye, Nicholas [1 ,2 ]
Ott, Jordan [1 ]
Garay, Michael J. [3 ]
El-Askary, Hesham Mohamed [4 ,5 ,6 ]
Linstead, Erik [5 ,7 ]
机构
[1] Chapman Univ, Computat & Data Sci Dept, Orange, CA 92866 USA
[2] CALTECH, Jet Prop Lab, Proc Algorithms & Calibrat Engn Grp, 4800 Oak Grove Dr, Pasadena, CA 91109 USA
[3] CALTECH, Jet Prop Lab, 4800 Oak Grove Dr, Pasadena, CA 91109 USA
[4] Chapman Univ, Ctr Excellence Earth Syst Modeling & Observat, Orange, CA 92866 USA
[5] Chapman Univ, Schmid Coll Sci & Technol, Orange, CA 92866 USA
[6] Alexandria Univ, Fac Sci, Dept Environm Sci, Alexandria 21522, Egypt
[7] Chapman Univ, Machine Learning & Assist Technol Lab, Orange, CA 92866 USA
关键词
Bigdata applications; clustering; computer vision; deep belief networks (DBNs); deep learning; CLASSIFICATION; MISR;
D O I
10.1109/JSTARS.2019.2920234
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The number of different modalities for remote sensors continues to grow, bringing with it an increase in the volume and complexity of the data being collected. Although these datasets individually provide valuable information, in aggregate they provide additional opportunities to discover meaningful patterns on a large scale. However, the ability to combine and analyze disparate datasets is challenged by the potentially vast parameter space that results from aggregation. Each dataset in itself requires instrument-specific and dataset-specific knowledge. If the intention is to use multiple, diverse datasets, one needs an understanding of how to translate and combine these parameters in an efficient and effective manner. While there are established techniques for combining datasets from specific domains or platforms, there is no generic, automated method that can address the problem in general. Here, we discuss the application of deep learning to track objects across different image-like data-modalities, given data in a similar spatio-temporal range, and automatically co-register these images. Using deep belief networks combined with unsupervised learning methods, we are able to recognize and separate different objects within image-like data in a structured manner, thus making progress toward the ultimate goal of a generic tracking and fusion pipeline requiring minimal human intervention.
引用
收藏
页码:3056 / 3066
页数:11
相关论文
共 50 条
  • [1] Unsupervised RGB-T object tracking with attentional multi-modal feature fusion
    Shenglan Li
    Rui Yao
    Yong Zhou
    Hancheng Zhu
    Bing Liu
    Jiaqi Zhao
    Zhiwen Shao
    Multimedia Tools and Applications, 2023, 82 : 23595 - 23613
  • [2] Unsupervised RGB-T object tracking with attentional multi-modal feature fusion
    Li, Shenglan
    Yao, Rui
    Zhou, Yong
    Zhu, Hancheng
    Liu, Bing
    Zhao, Jiaqi
    Shao, Zhiwen
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (15) : 23595 - 23613
  • [3] Deep Object Tracking with Multi-modal Data
    Zhang, Xuezhi
    Yuan, Yuan
    Lu, Xiaoqiang
    2016 INTERNATIONAL CONFERENCE ON COMPUTER, INFORMATION AND TELECOMMUNICATION SYSTEMS (CITS), 2016, : 161 - 165
  • [4] A Quantitative Validation of Multi-Modal Image Fusion and Segmentation for Object Detection and Tracking
    LaHaye, Nicholas
    Garay, Michael J.
    Bue, Brian D.
    El-Askary, Hesham
    Linstead, Erik
    REMOTE SENSING, 2021, 13 (12)
  • [5] Unsupervised Multi-modal Learning
    Iqbal, Mohammed Shameer
    ADVANCES IN ARTIFICIAL INTELLIGENCE (AI 2015), 2015, 9091 : 343 - 346
  • [6] Multi-Modal Sensor Fusion and Object Tracking for Autonomous Racing
    Karle, Phillip
    Fent, Felix
    Huch, Sebastian
    Sauerbeck, Florian
    Lienkamp, Markus
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (07): : 3871 - 3883
  • [7] Guided Image Deblurring by Deep Multi-Modal Image Fusion
    Liu, Yuqi
    Sheng, Zehua
    Shen, Hui-Liang
    IEEE ACCESS, 2022, 10 : 130708 - 130718
  • [8] An Unsupervised Deep Learning Method for Diffeomorphic Mono-and Multi-modal Image Registration
    Theljani, Anis
    Chen, Ke
    MEDICAL IMAGE UNDERSTANDING AND ANALYSIS, MIUA 2019, 2020, 1065 : 317 - 326
  • [9] CIRF: Coupled Image Reconstruction and Fusion Strategy for Deep Learning Based Multi-Modal Image Fusion
    Zheng, Junze
    Xiao, Junyan
    Wang, Yaowei
    Zhang, Xuming
    SENSORS, 2024, 24 (11)
  • [10] Deep learning supported breast cancer classification with multi-modal image fusion
    Hamdy, Eman
    Zaghloul, Mohamed Saad
    Badawy, Osama
    2021 22ND INTERNATIONAL ARAB CONFERENCE ON INFORMATION TECHNOLOGY (ACIT), 2021, : 319 - 325