Visual Hearing Aids: Artificial Visual Speech Stimuli for Audiovisual Speech Perception in Noise

被引:0
|
作者
Choudhary, Zubin Datta [1 ]
Bruder, Gerd [1 ]
Welch, Gregory F. [1 ]
机构
[1] Univ Cent Florida, Orlando, FL 32816 USA
基金
美国国家科学基金会;
关键词
Speech perception; background noise; hearing; speechreading; visualizations; virtual humans; user study; SOCIAL PRESENCE; VIRTUAL HUMANS; FACE;
D O I
10.1145/3611659.3615682
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Speech perception is optimal in quiet environments, but noise can impair comprehension and increase errors. In these situations, lip reading can help, but it is not always possible, such as during an audio call or when wearing a face mask. One approach to improve speech perception in these situations is to use an artificial visual lip reading aid. In this paper, we present a user study (N = 17) in which we compared three levels of audio stimuli visualizations and two levels of modulating the appearance of the visualization based on the speech signal, and we compared them against two control conditions: an audio-only condition, and a real human speaking. We measured participants' speech reception thresholds (SRTs) to understand the effects of these visualizations on speech perception in noise. These thresholds indicate the decibel levels of the speech signal that are necessary for a listener to receive the speech correctly 50% of the time. Additionally, we measured the usability of the approaches and the user experience. We found that the different artificial visualizations improved participants' speech reception compared to the audio-only baseline condition, but they were significantly poorer than the real human condition. This suggests that different visualizations can improve speech perception when the speaker's face is not available. However, we also discuss limitations of current plug-and-play lip sync software and abstract representations of the speaker in the context of speech perception.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] VISUAL AIDS FOR SPEECH CORRECTION
    RISBERG, A
    [J]. AMERICAN ANNALS OF THE DEAF, 1968, 113 (02) : 178 - 194
  • [32] Visual Aids to Speech Improvement
    Pronovost, Wilbert
    [J]. JOURNAL OF SPEECH DISORDERS, 1947, 12 (04): : 387 - 391
  • [33] Audiovisual Binding for Speech Perception in Noise and in Aging
    Ganesh, Attigodu Chandrashekara
    Berthommier, Frederic
    Schwartz, Jean-Luc
    [J]. LANGUAGE LEARNING, 2018, 68 : 193 - 220
  • [34] Effect of Digital Noise Reduction of Hearing Aids on Music and Speech Perception
    Kim, Hyo Jeong
    Lee, Jae Hee
    Shim, Hyun Joon
    [J]. JOURNAL OF AUDIOLOGY AND OTOLOGY, 2020, 24 (04): : 180 - 190
  • [35] Effect of face masks on speech perception in noise of individuals with hearing aids
    Choi, Jung Ho
    Choi, Hyo Jung
    Kim, Dong Hyun
    Park, Ji Hye
    An, Yong-Hwi
    Shim, Hyun Joon
    [J]. FRONTIERS IN NEUROSCIENCE, 2022, 16
  • [36] Better speech perception in noise with an assistive multimicrophone array for hearing aids
    Luts, H
    Maj, JB
    Soede, W
    Wouters, J
    [J]. EAR AND HEARING, 2004, 25 (05): : 411 - 420
  • [37] Comparison of Gated Audiovisual Speech Identification in Elderly Hearing Aid Users and Elderly Normal-Hearing Individuals: Effects of Adding Visual Cues to Auditory Speech Stimuli
    Moradi, Shahram
    Lidestam, Bjorn
    Ronnberg, Jerker
    [J]. TRENDS IN HEARING, 2016, 20
  • [38] Laterality in visual speech perception
    Smeele, PMT
    Massaro, DW
    Cohen, MM
    Sittig, AC
    [J]. JOURNAL OF EXPERIMENTAL PSYCHOLOGY-HUMAN PERCEPTION AND PERFORMANCE, 1998, 24 (04) : 1232 - 1242
  • [39] AUDITORY, VISUAL, AND AUDIOVISUAL SPEECH-INTELLIGIBILITY FOR SENTENCE-LENGTH STIMULI - AN INVESTIGATION OF CONVERSATIONAL AND CLEAR SPEECH
    GAGNE, JP
    QUERENGESSER, C
    FOLKEARD, P
    MUNHALL, KG
    MASTERSON, VM
    [J]. VOLTA REVIEW, 1995, 97 (01) : 33 - 51
  • [40] The visual mismatch negativity elicited with visual speech stimuli
    Files, Benjamin T.
    Auer, Edward T., Jr.
    Bernstein, Lynne E.
    [J]. FRONTIERS IN HUMAN NEUROSCIENCE, 2013, 7