Concept Based Hybrid Fusion of Multimodal Event Signals

被引:0
|
作者
Wang, Yuhui [1 ]
von der Weth, Christian [2 ]
Zhang, Yehong [3 ]
Low, Kian Hsiang [3 ]
Singh, Vivek K. [4 ]
Kankanhalli, Mohan [3 ]
机构
[1] Natl Univ Singapore, NUS Grad Sch Integrat Sci & Engn, Singapore, Singapore
[2] Natl Univ Singapore, Interact & Digital Media Inst, SeSaMe Ctr, Singapore, Singapore
[3] Natl Univ Singapore, Dept Comp Sci, Singapore, Singapore
[4] Rutgers State Univ, Sch Commun & Informat, New Brunswick, NJ USA
关键词
multimodal fusion; situation understanding; multisensor data analysis; events; IMAGE FUSION;
D O I
10.1109/ISM.2016.64
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent years have seen a significant increase in the number of sensors and resulting event related sensor data, allowing for a better monitoring and understanding of real-world events and situations. Event-related data come from not only physical sensors (e.g., CCTV cameras, webcams) but also from social or microblogging platforms (e.g., Twitter). Given the wide-spread availability of sensors, we observe that sensors of different modalities often independently observe the same events. We argue that fusing multimodal data about an event can be helpful for more accurate detection, localization and detailed description of events of interest. However, multimodal data often include noisy observations, varying information densities and heterogeneous representations, which makes the fusion a challenging task. In this paper, we propose a hybrid fusion approach that takes the spatial and semantic characteristics of sensor signals about events into account. For this, we first adopt the concept of an image-based representation that expresses the situation of particular visual concepts (e.g. "crowdedness", "people marching") called Cmage for both physical and social sensor data. Based on this Cmage representation, we model sparse sensor information using a Gaussian process, fuse multimodal event signals with a Bayesian approach, and incorporate spatial relations between the sensor and social observations. We demonstrate the effectiveness of our approach as a proof-of-concept over real-world data. Our early results show that the proposed approach can reliably reduce the sensor-related noise, locate the event place, improve event detection reliability, and add semantic context so that the fused data provides a better picture of the observed events.
引用
收藏
页码:14 / 19
页数:6
相关论文
共 50 条
  • [31] Multimodal Fusion using Dynamic Hybrid Models
    Amer, Mohamed R.
    Siddiquie, Behjat
    Khan, Saad
    Divakaran, Ajay
    Sawhney, Harpreet
    2014 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2014, : 556 - 563
  • [32] Concept-based evidential reasoning for multimodal fusion in human-computer interaction
    Reddy, B. Srinath
    Basir, Otman A.
    APPLIED SOFT COMPUTING, 2010, 10 (02) : 567 - 577
  • [33] Knowledge Based Multimodal Result Fusion for Distributed and Heterogeneous Multimedia Environments: Concept and Ideas
    Stegmaier, Florian
    Buerger, Tobias
    Doeller, Mario
    Kosch, Harald
    ADAPTIVE MULTIMEDIA RETRIEVAL: CONTEXT, EXPLORATION, AND FUSION, 2012, 6817 : 61 - +
  • [34] HYBRID FUSION BASED APPROACH FOR MULTIMODAL EMOTION RECOGNITION WITH INSUFFICIENT LABELED DATA
    Kumar, Puneet
    Khokher, Vedanti
    Gupta, Yukti
    Raman, Balasubramanian
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 314 - 318
  • [35] A Hybrid Multimodal Data Fusion-Based Method for Identifying Gambling Websites
    Wang, Chenyang
    Zhang, Min
    Shi, Fan
    Xue, Pengfei
    Li, Yang
    ELECTRONICS, 2022, 11 (16)
  • [36] Face-Iris Multimodal Biometric System based on Hybrid Level Fusion
    Ammour, Basma
    Bouden, Toufik
    Boubchir, Larbi
    2018 41ST INTERNATIONAL CONFERENCE ON TELECOMMUNICATIONS AND SIGNAL PROCESSING (TSP), 2018, : 298 - 302
  • [37] A Smart Fusion Framework for Multimodal Object, Activity and Event Detection
    Chetty, Girija
    Yamin, Mohammad
    PROCEEDINGS OF THE 10TH INDIACOM - 2016 3RD INTERNATIONAL CONFERENCE ON COMPUTING FOR SUSTAINABLE GLOBAL DEVELOPMENT, 2016, : 1417 - 1422
  • [38] Multimodal Feature Fusion for Robust Event Detection in Web Videos
    Natarajan, Pradeep
    Wu, Shuang
    Vitaladevuni, Shiv
    Zhuang, Xiaodan
    Tsakalidis, Stavros
    Park, Unsang
    Prasad, Rohit
    Natarajan, Premkumar
    2012 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2012, : 1298 - 1305
  • [39] Model-based sensor fusion of multimodal cardiorespiratory signals using an unscented Kalman filter
    Linschmann, Onno
    Leonhardt, Steffen
    Antink, Christoph Hoog
    AT-AUTOMATISIERUNGSTECHNIK, 2020, 68 (11) : 933 - 940
  • [40] MF-Net: a multimodal fusion network for emotion recognition based on multiple physiological signals
    Zhu, Lei
    Ding, Yu
    Huang, Aiai
    Tan, Xufei
    Zhang, Jianhai
    SIGNAL IMAGE AND VIDEO PROCESSING, 2025, 19 (01)