Multi-modal remote perception learning for object sensory data

被引:0
|
作者
Almujally, Nouf Abdullah [1 ]
Rafique, Adnan Ahmed [2 ]
Al Mudawi, Naif [3 ]
Alazeb, Abdulwahab [3 ]
Alonazi, Mohammed [4 ]
Algarni, Asaad [5 ]
Jalal, Ahmad [6 ,7 ]
Liu, Hui [8 ]
机构
[1] Princess Nourah Bint Abdulrahman Univ, Coll Comp & Informat Sci, Dept Informat Syst, Riyadh, Saudi Arabia
[2] Univ Poonch Rawalakot, Dept Comp Sci & IT, Rawalakot, Pakistan
[3] Najran Univ, Coll Comp Sci & Informat Syst, Dept Comp Sci, Najran, Saudi Arabia
[4] Prince Sattam Bin Abdulaziz Univ, Coll Comp Engn & Sci, Dept Comp Engn, Al Kharj, Saudi Arabia
[5] Northern Border Univ, Fac Comp & Informat Technol, Dept Comp Sci, Rafha, Saudi Arabia
[6] Air Univ, Fac Comp Sci, Islamabad, Pakistan
[7] Korea Univ, Coll Informat, Dept Comp Sci & Engn, Seoul, South Korea
[8] Univ Bremen, Cognit Syst Lab, Bremen, Germany
来源
关键词
multi-modal; sensory data; objects recognition; visionary sensor; simulation environment multi-modal; simulation environment; RECOGNITION;
D O I
10.3389/fnbot.2024.1427786
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Introduction When it comes to interpreting visual input, intelligent systems make use of contextual scene learning, which significantly improves both resilience and context awareness. The management of enormous amounts of data is a driving force behind the growing interest in computational frameworks, particularly in the context of autonomous cars.Method The purpose of this study is to introduce a novel approach known as Deep Fused Networks (DFN), which improves contextual scene comprehension by merging multi-object detection and semantic analysis.Results To enhance accuracy and comprehension in complex situations, DFN makes use of a combination of deep learning and fusion techniques. With a minimum gain of 6.4% in accuracy for the SUN-RGB-D dataset and 3.6% for the NYU-Dv2 dataset.Discussion Findings demonstrate considerable enhancements in object detection and semantic analysis when compared to the methodologies that are currently being utilized.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Imagery in multi-modal object learning
    Jüttner, M
    Rentschler, I
    BEHAVIORAL AND BRAIN SCIENCES, 2002, 25 (02) : 197 - +
  • [2] Deep Object Tracking with Multi-modal Data
    Zhang, Xuezhi
    Yuan, Yuan
    Lu, Xiaoqiang
    2016 INTERNATIONAL CONFERENCE ON COMPUTER, INFORMATION AND TELECOMMUNICATION SYSTEMS (CITS), 2016, : 161 - 165
  • [3] Multi-modal Perception
    Kondo, T.
    Denshi Joho Tsushin Gakkai Shi/Journal of the Institute of Electronics, Information and Communications Engineers, 78 (12):
  • [4] Multi-modal perception
    BT Technol J, 1 (35-46):
  • [5] Multi-modal perception
    Hollier, MP
    Rimell, AN
    Hands, DS
    Voelcker, RM
    BT TECHNOLOGY JOURNAL, 1999, 17 (01) : 35 - 46
  • [6] MULTI-MODAL REMOTE SENSING DATA FUSION FRAMEWORK
    Ghaffar, M. A. A.
    Vu, T. T.
    Maul, T. H.
    FOSS4G-EUROPE 2017 - ACADEMIC TRACK, 2017, 42-4 (W2): : 85 - 89
  • [7] Multi-modal of object trajectories
    Partsinevelos, P.
    JOURNAL OF SPATIAL SCIENCE, 2008, 53 (01) : 17 - 30
  • [8] Learning to Hash on Partial Multi-Modal Data
    Wang, Qifan
    Si, Luo
    Shen, Bin
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 3904 - 3910
  • [9] Small Object Detection Technology Using Multi-Modal Data Based on Deep Learning
    Park, Chi-Won
    Seo, Yuri
    Sun, Teh-Jen
    Lee, Ga-Won
    Huh, Eui-Nam
    2023 INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING, ICOIN, 2023, : 420 - 422
  • [10] Learning to Generate Object Segment Proposals with Multi-modal Cues
    Zhang, Haoyang
    He, Xuming
    Porikli, Fatih
    COMPUTER VISION - ACCV 2016, PT I, 2017, 10111 : 121 - 136