Visual-Auditory saliency detection using event-driven visual sensors

被引:0
|
作者
Akolkar, Himanshu [1 ]
Valeiras, David Reverter [2 ]
Benosman, Ryad [2 ]
Bartolozzi, Chiara [1 ]
机构
[1] Ist Italiano Tecnol, ICub Facil, I-16163 Genoa, Italy
[2] Univ Paris 06, Vis Inst, F-75012 Paris, France
关键词
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper presents a novel architecture for audiovisual saliency detection using event-based visual sensors and traditional microphones installed on the head of a humanoid robot. In the context of collision detection, salient sensory events must be detected at the same time in vision and in the auditory domain. Real collisions in the visual space can be distinguished from fake ones (e.g. due to movements of two objects that occlude each other) because they generate a sound at the time of collision. This temporal coincidence is extremely difficult to detect with frame-based sensors, that intrinsically add a fixed delay in the sensory acquisition or can miss the collision. The high temporal resolution of event-driven vision sensors together with a real time clustering and tracking algorithm allow for the detection of potential collisions with very low latency. Auditory events corresponding to collisions are detected using simple spectral analysis of auditory signals. The visual event can be therefore temporally integrated with coherently occurring auditory events to detect fast-transitions and disentangle real collisions from visual or auditory events that do not correspond to any. The proposed audio-visual collision detection is used in the context of human robot interaction, to detect people clapping in front of the robot and orient its gaze toward the perceived collision.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Memory for visual, auditory and visual-auditory material
    不详
    ANNEE PSYCHOLOGIQUE, 1936, 37 : 655 - 656
  • [2] From Semantic Categories to Fixations: A Novel Weakly-supervised Visual-auditory Saliency Detection Approach
    Wang, Guotao
    Chen, Chenglizhao
    Fan, Dengping
    Hao, Aimin
    Qin, Hong
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 15114 - 15123
  • [3] VISUAL-AUDITORY DISTANCE CONSTANCY
    ENGEL, GR
    DOUGHERTY, WG
    NATURE, 1971, 234 (5327) : 308 - +
  • [4] VISUAL-AUDITORY DISTANCE CONSTANCY
    DAY, RH
    NATURE, 1972, 238 (5361) : 227 - &
  • [5] Visual-auditory learning network for construction equipment action detection
    Jung, Seunghoon
    Jeoung, Jaewon
    Lee, Dong-Eun
    Jang, Hyounseung
    Hong, Taehoon
    COMPUTER-AIDED CIVIL AND INFRASTRUCTURE ENGINEERING, 2023, 38 (14) : 1916 - 1934
  • [6] Representation of visual-auditory integration effectiveness in event-related potentials
    Babenko, VV
    Kotova, MJ
    Safina, ZM
    PERCEPTION, 2005, 34 : 217 - 217
  • [7] Visual-auditory interaction in perception of the variance
    Ueda, Sachiyo
    Mizuguchi, Ayane
    Yakushijin, Reiko
    Ishiguchi, Akira
    PERCEPTION, 2016, 45 : 344 - 345
  • [8] Exploring Visual-Auditory Redirected Walking Using Auditory Cues in Reality
    Ogawa, Kumpei
    Fujita, Kazuyuki
    Sakamoto, Shuichi
    Takashima, Kazuki
    Kitamura, Yoshifumi
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2024, 30 (08) : 5782 - 5794
  • [9] Visual-auditory substitution device for indoor navigation based on fast visual marker detection
    Scalvini, Florian
    Bordeau, Camille
    Ambard, Maxime
    Migniot, Cyrille
    Argon, Stephane
    Dubois, Julien
    2022 16TH INTERNATIONAL CONFERENCE ON SIGNAL-IMAGE TECHNOLOGY & INTERNET-BASED SYSTEMS, SITIS, 2022, : 259 - 266
  • [10] Automatic auditory change detection in humans is influenced by visual-auditory associative learning
    Laine, Matti
    Kwon, Myoung Soo
    Hamalainen, Heikki
    NEUROREPORT, 2007, 18 (16) : 1697 - 1701