Multi-Modal Information Fusion for Localization of Emergency Vehicles

被引:0
|
作者
Joshi, Aruna Kumar [1 ,2 ]
Kulkarni, Shrinivasrao B. [2 ,3 ]
机构
[1] SKSVMA Coll Engn & Technol, Dept Comp Sci & Engn, Gadag 582116, Karnataka, India
[2] Visvesvaraya Technol Univ, Belagavi, Karnataka, India
[3] SDM Coll Engn & Technol, Dept Comp Sci & Engn, Dharwad 580002, Karnataka, India
关键词
Intelligent transportation; traffic density; deep learning; emergency vehicle; emergency siren; traffic flow; fusion methods; multi-modal method;
D O I
10.1142/S0219467825500500
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
In urban and city environments, road transportation contributes significantly to the generation of substantial traffic. However, this surge in vehicles leads to complex issues, including hindered emergency vehicle movement due to high density and congestion. Scarcity of human personnel amplifies these challenges. As traffic conditions worsen, the need for automated solutions to manage emergency situations becomes more evident. Intelligent traffic monitoring can identify and prioritize emergency vehicles, potentially saving lives. However, categorizing emergency vehicles through visual analysis faces difficulties such as clutter, occlusions, and traffic variations. Visual-based techniques for vehicle detection rely on clear rear views, but this is problematic in dense traffic. In contrast, audio-based methods are resilient to the Doppler Effect from moving vehicles, but handling diverse background noises remains unexplored. Using acoustics for emergency vehicle localization presents challenges related to sensor range and real-world noise. Addressing these issues, this study introduces a novel solution: combining visual and audio data for enhanced detection and localization of emergency vehicles in road networks. Leveraging this multi-modal approach aims to bolster accuracy and robustness in emergency vehicle management. The proposed methodology consists of several key steps. The presence of an emergency vehicle is initially detected through the preprocessing of visual images, involving the removal of clutter and occlusions via an adaptive background model. Subsequently, a cell-wise classification strategy utilizing a customized Visual Geometry Group Network (VGGNet) deep learning model is employed to determine the presence of emergency vehicles within individual cells. To further reinforce the accuracy of emergency vehicle presence detection, the outcomes from the audio data analysis are integrated. This involves the extraction of spectral features from audio streams, followed by classification utilizing a support vector machine (SVM) model. The fusion of information derived from both visual and audio sources is utilized in the construction of a more comprehensive and refined traffic state map. This augmented map facilitates the effective management of emergency vehicle transit. In empirical evaluations, the proposed solution demonstrates its capability to mitigate challenges like visual clutter, occlusions, and variations in traffic density common issues encountered in traditional visual analysis methods. Notably, the proposed approach achieves an impressive accuracy rate of approximately 98.15% in the localization of emergency vehicles.
引用
收藏
页数:26
相关论文
共 50 条
  • [1] MULTI-MODAL RECURRENT FUSION FOR INDOOR LOCALIZATION
    Yu, Jianyuan
    Wang, Pu
    Koike-Akino, Toshiaki
    Orlik, Philip, V
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 5083 - 5087
  • [2] Cooperative Multi-Modal Localization in Connected and Autonomous Vehicles
    Piperigkos, Nikos
    Lalos, Aris S.
    Berberidis, Kostas
    Anagnostopoulos, Christos
    [J]. 2020 IEEE 3RD CONNECTED AND AUTOMATED VEHICLES SYMPOSIUM (CAVS), 2020,
  • [3] CollabLoc: Privacy-Preserving Multi-Modal Localization via Collaborative Information Fusion
    Sadhu, Vidyasagar
    Pompili, Dario
    Zonouz, Saman
    Sritapan, Vincent
    [J]. 2017 26TH INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND NETWORKS (ICCCN 2017), 2017,
  • [4] Multi-modal Fusion
    Liu, Huaping
    Hussain, Amir
    Wang, Shuliang
    [J]. INFORMATION SCIENCES, 2018, 432 : 462 - 462
  • [5] Fusion of auxiliary information for multi-modal biometrics authentication
    Toh, KA
    Yau, WY
    Lim, E
    Chen, L
    Ng, CH
    [J]. BIOMETRIC AUTHENTICATION, PROCEEDINGS, 2004, 3072 : 678 - 685
  • [6] MULTI-MODAL INFORMATION FUSION FOR CLASSIFICATION OF KIDNEY ABNORMALITIES
    Varsha, S.
    Nasser, Sahar Almahfouz
    Bala, Gouranga
    Kurian, Nikhil Cherian
    Sethi, Amit
    [J]. 2022 IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING CHALLENGES (IEEE ISBI 2022), 2022,
  • [7] Heterogeneous Feature Fusion Approach for Multi-Modal Indoor Localization
    Zhou, Junyi
    Huang, Kaixuan
    Tang, Siyu
    Zhang, Shunqing
    [J]. 2024 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC 2024, 2024,
  • [8] Robust indoor localization based on multi-modal information fusion and multi-scale sequential feature extraction
    Wang, Qinghu
    Jia, Jie
    Chen, Jian
    Deng, Yansha
    Wang, Xingwei
    Aghvami, Abdol Hamid
    [J]. FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 155 : 164 - 178
  • [9] MOZARD: Multi-Modal Localization for Autonomous Vehicles in Urban Outdoor Environments
    Schaupp, Lukas
    Pfreundschuh, Patrick
    Buerki, Mathias
    Cadena, Cesar
    Siegwart, Roland
    Nieto, Juan
    [J]. 2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 4828 - 4833
  • [10] Multi-modal Information Extraction and Fusion with Convolutional Neural Networks
    Kumar, Dinesh
    Sharma, Dharmendra
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,