Finding the correspondence of audio-visual events caused by multiple movements

被引:0
|
作者
Chen, J. [1 ]
Mukai, T. [1 ]
Takeuchi, Y. [1 ]
Matsumoto, T. [1 ]
Kudo, H. [1 ]
Yamamura, T. [1 ]
Ohnishi, N. [1 ]
机构
[1] Nagoya University, Furo-cho, Chikusha-ku, Nagoya 464-8603, Japan
关键词
Cameras - Correlation methods - Microphones - Speech recognition;
D O I
10.3169/itej.55.1450
中图分类号
学科分类号
摘要
We understand the environment by integrating information obtained by the senses of sight, hearing and touch. To integrate information across different senses, we must find the correspondence of events observed by different senses. This paper presents a general method for relating the audio-visual events of more than one movement (repetitive and non-repetitive movement) observed by one camera and one microphone. The method uses general laws without object-specific knowledge. As corresponding cues, we use Gestalt's grouping laws: simultaneity of the occurrence of the sound and the change in movement, and similarity of repetition between sound and movement. We conducted experiments in the real environment, and obtained satisfactory results showing the effectiveness of the proposed method.
引用
收藏
页码:1450 / 1459
相关论文
共 50 条
  • [1] Relating audio-visual events caused by multiple movements in the real environment
    Chen, J
    Mukai, T
    Takeuchi, Y
    Matsumoto, T
    Kudo, H
    Yamamura, T
    Ohnishi, N
    CCCT 2003, VOL 4, PROCEEDINGS: COMPUTER, COMMUNICATION AND CONTROL TECHNOLOGIES: I, 2003, : 162 - 167
  • [2] Relating audio-visual events caused by multiple movements: In the case of entire object movement
    Chen, JJ
    Mukai, T
    Takeuchi, Y
    Matsumoto, T
    Kudo, H
    Yamamura, T
    Ohnishi, N
    PROCEEDINGS OF THE FIFTH INTERNATIONAL CONFERENCE ON INFORMATION FUSION, VOL I, 2002, : 213 - 219
  • [3] Relating audio-visual events caused by multiple movements: In the case of entire object movements and sound location changes
    Chen, JJ
    Mukai, T
    Takeuchi, Y
    Matsumoto, T
    Kudo, H
    Yamamura, T
    Ohnishi, N
    6TH WORLD MULTICONFERENCE ON SYSTEMICS, CYBERNETICS AND INFORMATICS, VOL XIII, PROCEEDINGS: CONCEPTS AND APPLICATIONS OF SYSTEMICS, CYBERNETICS AND INFORMATICS III, 2002, : 5 - 12
  • [4] A biologically inspired method for finding correspondence between audio-visual events based on selective attention
    Graduate School of Information Science, Nagoya University, Furo-cho, Chikusa-ku, Nagoya, 464-8603, Japan
    不详
    不详
    Kyokai Joho Imeji Zasshi, 2008, 7 (1086-1097):
  • [5] The role of audio-visual congruence in discrimination of visual events
    Sinico, M
    PERCEPTION, 2004, 33 : 141 - 141
  • [6] Sequential Audio-Visual Correspondence With Alternating Diffusion Kernels
    Dov, David
    Talmon, Ronen
    Cohen, Israel
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2018, 66 (12) : 3100 - 3111
  • [7] Temporal structure and complexity affect audio-visual correspondence detection
    Denison, Rachel N.
    Driver, Jon
    Ruff, Christian C.
    FRONTIERS IN PSYCHOLOGY, 2013, 3
  • [8] Robust Contrastive Learning Against Audio-Visual Noisy Correspondence
    Zhao, Yihan
    Xi, Wei
    Bai, Gairui
    Liu, Xinhui
    Zhao, Jizhong
    PATTERN RECOGNITION AND COMPUTER VISION, PT V, PRCV 2024, 2025, 15035 : 526 - 540
  • [9] Dynamic audio-visual correspondence in musicians and non-musicians
    Guo, Xiyu
    Qu, Jianning
    Liu, Mengying
    Liu, Chuanjun
    Huang, Jianping
    PSYCHOLOGY OF MUSIC, 2024, 52 (02) : 175 - 186
  • [10] A Multimodal Saliency Model for Videos With High Audio-Visual Correspondence
    Min, Xiongkuo
    Zhai, Guangtao
    Zhou, Jiantao
    Zhang, Xiao-Ping
    Yang, Xiaokang
    Guan, Xinping
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 3805 - 3819