AUDIO-VISUAL PERCEPTION OF OMNIDIRECTIONAL VIDEO FOR VIRTUAL REALITY APPLICATIONS

被引:0
|
作者
Chao, Fang-Yi [1 ]
Ozcinar, Cagri [2 ]
Wang, Chen [2 ]
Zerman, Emin [2 ]
Zhang, Lu [1 ]
Hamidouche, Wassim [1 ]
Deforges, Olivier [1 ]
Smolic, Aljosa [2 ]
机构
[1] Univ Rennes, INSA Rennes, CNRS, IETR UMR 6164, F-35000 Rennes, France
[2] Trinity Coll Dublin, V SENSE, Sch Comp Sci & Stat, Dublin, Ireland
基金
爱尔兰科学基金会;
关键词
Ambisonics; omnidirectional video; virtual reality (VR); visual attention; audio-visual saliency;
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Ambisonics, which constructs a sound distribution over the full viewing sphere, improves immersive experience in omnidirectional video (ODV) by enabling observers to perceive the sound directions. Thus, human attention could be guided by audio and visual stimuli simultaneously. Numerous datasets have been proposed to investigate human visual attention by collecting eye fixations of observers navigating ODV with head-mounted displays (HMD). However, there is no such dataset analyzing the impact of audio information. In this paper, we establish a new audio-visual attention dataset for ODV with mute, mono, and ambisonics. The user behavior including visual attention corresponding to sound source locations, viewing navigation congruence between observers and fixations distributions in these three audio modalities is studied based on video and audio content. From our statistical analysis, we preliminarily found that, compared to only perceiving visual cues, perceiving visual cues with salient object sound (i.e., human voice, siren of ambulance) could draw more visual attention to the objects making sound and guide viewing behaviour when such objects are not in the current field of view. The more in-depth interactive effects between audio and visual cues in mute, mono and ambisonics still require further comprehensive study. The dataset and developed testbed in this initial work will be publicly available with the paper to foster future research on audio-visual attention for ODV.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Visual Attention in Omnidirectional Video for Virtual Reality Applications
    Ozcinar, Cagri
    Smolic, Aljosa
    [J]. 2018 TENTH INTERNATIONAL CONFERENCE ON QUALITY OF MULTIMEDIA EXPERIENCE (QOMEX), 2018, : 1 - 6
  • [2] Towards Audio-Visual Saliency Prediction for Omnidirectional Video with Spatial Audio
    Chao, Fang-Yi
    Ozcinar, Cagri
    Zhang, Lu
    Hamidouche, Wassim
    Deforges, Olivier
    Smolic, Aljosa
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP), 2020, : 355 - 358
  • [3] Modeling the Impact of Head-Body Rotations on Audio-Visual Spatial Perception for Virtual Reality Applications
    Bernal-Berdun, Edurne
    Vallejo, Mateo
    Sun, Qi
    Serrano, Ana
    Gutierrez, Diego
    [J]. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2024, 30 (05) : 2624 - 2632
  • [4] Audio-Visual Perception - The Perception of Object Material in a Virtual Environment
    Anderson, Ryan
    Arro, Joosep
    Hansen, Christian Schutt
    Serafin, Stefania
    [J]. Augmented Reality, Virtual Reality, and Computer Graphics, Pt I, 2016, 9768 : 162 - 171
  • [5] Detection of inconsistent audio-visual events in virtual reality
    Sorkin, A.
    Peled, A. ]
    Weinshall, D.
    [J]. PERCEPTION, 2006, 35 : 203 - 204
  • [6] Audio, Visual, and Audio-Visual Egocentric Distance Perception by Moving Subjects in Virtual Environments
    Rebillat, Marc
    Boutillon, Xavier
    Corteel, Etienne
    Katz, Brian F. G.
    [J]. ACM TRANSACTIONS ON APPLIED PERCEPTION, 2012, 9 (04)
  • [7] AN AUDIO-VISUAL QUALITY ASSESSMENT METHODOLOGY IN VIRTUAL REALITY ENVIRONMENT
    Zhang, Bo
    Yan, Zhaoyu
    Wang, Jing
    Luo, Yiyu
    Yang, Shu
    Fei, Zesong
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW 2018), 2018,
  • [8] Bimodal Perception of Audio-Visual Material Properties for Virtual Environments
    Bonneel, Nicolas
    Suied, Clara
    Viaud-Delmon, Isabelle
    Drettakis, George
    [J]. ACM TRANSACTIONS ON APPLIED PERCEPTION, 2010, 7 (01)
  • [9] TOWARDS GENERATING AMBISONICS USING AUDIO-VISUAL CUE FOR VIRTUAL REALITY
    Rana, Aakanksha
    Ozcinar, Cagri
    Smolic, Aljosa
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 2012 - 2016
  • [10] Audio-visual aligned saliency model for omnidirectional video with implicit neural representation learning
    Zhu, Dandan
    Shao, Xuan
    Zhang, Kaiwei
    Min, Xiongkuo
    Zhai, Guangtao
    Yang, Xiaokang
    [J]. APPLIED INTELLIGENCE, 2023, 53 (19) : 22615 - 22634