Visual-Auditory saliency detection using event-driven visual sensors

被引:0
|
作者
Akolkar, Himanshu [1 ]
Valeiras, David Reverter [2 ]
Benosman, Ryad [2 ]
Bartolozzi, Chiara [1 ]
机构
[1] Ist Italiano Tecnol, ICub Facil, I-16163 Genoa, Italy
[2] Univ Paris 06, Vis Inst, F-75012 Paris, France
关键词
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper presents a novel architecture for audiovisual saliency detection using event-based visual sensors and traditional microphones installed on the head of a humanoid robot. In the context of collision detection, salient sensory events must be detected at the same time in vision and in the auditory domain. Real collisions in the visual space can be distinguished from fake ones (e.g. due to movements of two objects that occlude each other) because they generate a sound at the time of collision. This temporal coincidence is extremely difficult to detect with frame-based sensors, that intrinsically add a fixed delay in the sensory acquisition or can miss the collision. The high temporal resolution of event-driven vision sensors together with a real time clustering and tracking algorithm allow for the detection of potential collisions with very low latency. Auditory events corresponding to collisions are detected using simple spectral analysis of auditory signals. The visual event can be therefore temporally integrated with coherently occurring auditory events to detect fast-transitions and disentangle real collisions from visual or auditory events that do not correspond to any. The proposed audio-visual collision detection is used in the context of human robot interaction, to detect people clapping in front of the robot and orient its gaze toward the perceived collision.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] A Biomimetic Visual Detection Model: Event-Driven LGMDs Implemented With Fractional Spiking Neuron Circuits
    Deng, Yabin
    Ruan, Haojie
    He, Shan
    Yang, Tao
    Guo, Donghui
    IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2024, 71 (10) : 2978 - 2990
  • [32] DETECTION OF VISUAL-AUDITORY REACTION RATES IN INDIVIDUALS WITH TYPE 2 DIABETES MELLITUS
    Bilaloglu, Melisa
    Ergun, Ahmet
    Altay, Feride P.
    Iyidir, Ozlem T.
    Kiziltan, Erhan
    DIABETES MELLITUS, 2023, 26 (03): : 236 - 242
  • [33] Response to visual vs visual-auditory presentation in a go-to-college program
    Elliott, FR
    PSYCHOLOGICAL REVIEW, 1937, 28 : 703 - 707
  • [34] Event-driven grammars: relating abstract and concrete levels of visual languages
    Esther Guerra
    Juan de Lara
    Software & Systems Modeling, 2007, 6 : 317 - 347
  • [35] VISUAL SALIENCY DETECTION USING VIDEO DECOMPOSITION
    Bhattacharya, Saumik
    Gupta, Sumana
    Venkatesh, K. S.
    2016 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2016, : 684 - 688
  • [36] Event-driven grammars: relating abstract and concrete levels of visual languages
    Guerra, Esther
    de Lara, Juan
    SOFTWARE AND SYSTEMS MODELING, 2007, 6 (03): : 317 - 347
  • [37] Saliency detection using midlevel visual cues
    Yu, Jin-Gang
    Tian, Jinwen
    OPTICS LETTERS, 2012, 37 (23) : 4994 - 4996
  • [38] Visual saliency detection using information divergence
    Hou, Weilong
    Gao, Xinbo
    Tao, Dacheng
    Li, Xuelong
    PATTERN RECOGNITION, 2013, 46 (10) : 2658 - 2669
  • [39] Visual Saliency Detection Using Spatiotemporal Decomposition
    Bhattacharya, Saumik
    Venkatesh, K. Subramanian
    Gupta, Sumana
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (04) : 1665 - 1675
  • [40] Event-Driven Stereo Visual Tracking Algorithm to Solve Object Occlusion
    Camunas-Mesa, Luis A.
    Serrano-Gotarredona, Teresa
    Ieng, Sio-Hoi
    Benosman, Ryad
    Linares-Barranco, Bernabe
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (09) : 4223 - 4237