Semantic Cues Modulate Children's and Adults' Processing of Audio-Visual Face Mask Speech

被引:16
|
作者
Schwarz, Julia [1 ]
Li, Katrina Kechun [1 ]
Sim, Jasper Hong [1 ]
Zhang, Yixin [1 ]
Buchanan-Worster, Elizabeth [2 ]
Post, Brechtje [1 ]
Gibson, Jenny Louise [3 ]
McDougall, Kirsty [1 ]
机构
[1] Univ Cambridge, Fac Modern & Medieval Languages & Linguist, Cambridge, England
[2] Univ Cambridge, Med Res Council Cognit & Brain Sci Unit, Cambridge, England
[3] Univ Cambridge, Fac Educ, Cambridge, England
来源
FRONTIERS IN PSYCHOLOGY | 2022年 / 13卷
基金
英国科研创新办公室;
关键词
speech processing; face masks; cued shadowing; audio-visual integration; semantic prediction; language development; internet-based data collection; bottom-up vs; top-down; EVENT-RELATED POTENTIALS; PERCEPTION; SPOKEN; NOISE; INTELLIGIBILITY; PREDICTION; INFANTS; MOUTH; EYES; RECOGNITION;
D O I
10.3389/fpsyg.2022.879156
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
During the COVID-19 pandemic, questions have been raised about the impact of face masks on communication in classroom settings. However, it is unclear to what extent visual obstruction of the speaker's mouth or changes to the acoustic signal lead to speech processing difficulties, and whether these effects can be mitigated by semantic predictability, i.e., the availability of contextual information. The present study investigated the acoustic and visual effects of face masks on speech intelligibility and processing speed under varying semantic predictability. Twenty-six children (aged 8-12) and twenty-six adults performed an internet-based cued shadowing task, in which they had to repeat aloud the last word of sentences presented in audio-visual format. The results showed that children and adults made more mistakes and responded more slowly when listening to face mask speech compared to speech produced without a face mask. Adults were only significantly affected by face mask speech when both the acoustic and the visual signal were degraded. While acoustic mask effects were similar for children, removal of visual speech cues through the face mask affected children to a lesser degree. However, high semantic predictability reduced audio-visual mask effects, leading to full compensation of the acoustically degraded mask speech in the adult group. Even though children did not fully compensate for face mask speech with high semantic predictability, overall, they still profited from semantic cues in all conditions. Therefore, in classroom settings, strategies that increase contextual information such as building on students' prior knowledge, using keywords, and providing visual aids, are likely to help overcome any adverse face mask effects.
引用
收藏
页数:17
相关论文
共 50 条
  • [21] DNN driven Speaker Independent Audio-Visual Mask Estimation for Speech Separation
    Gogate, Mandar
    Adeel, Ahsan
    Marxer, Ricard
    Barker, Jon
    Hussain, Amir
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 2723 - 2727
  • [22] Early visual deprivation affects the development of face recognition and of audio-visual speech perception
    Putzar, Lisa
    Hoetting, Kirsten
    Roeder, Brigitte
    RESTORATIVE NEUROLOGY AND NEUROSCIENCE, 2010, 28 (02) : 251 - 257
  • [23] Audio-Visual and Meaningful Semantic Context Enhancements in Older and Younger Adults
    Smayda, Kirsten E.
    Van Engen, Kristin J.
    Maddox, W. Todd
    Chandrasekaran, Bharath
    PLOS ONE, 2016, 11 (03):
  • [24] The impact of auditory, visual, and audio-visual sensory cues on multiple object tracking in children
    Atkins, Polly L.
    Hodgson, Timothy
    Dickinson, Patrick
    Hicks, Kieran
    Focker, Julia
    PERCEPTION, 2023, 52 (05) : 346 - 346
  • [25] Neural processing of audio-visual integration in speech perception: An MEG study
    Hiroe, Nobuo
    Shinozaki, Jun
    Yoshioka, Taku
    Sato, Masa-aki
    Sekiyama, Kaoru
    NEUROSCIENCE RESEARCH, 2010, 68 : E332 - E332
  • [26] Neural entrainment to rhythmically presented auditory, visual, and audio-visual speech in children
    Power, Alan James
    Mead, Natasha
    Barnes, Lisa
    Goswami, Usha
    FRONTIERS IN PSYCHOLOGY, 2012, 3
  • [27] Atypical audio-visual neural synchrony and speech processing in early autism
    Wang, Xiaoyue
    Bouton, Sophie
    Kojovic, Nada
    Giraud, Anne-Lise
    Schaer, Marie
    JOURNAL OF NEURODEVELOPMENTAL DISORDERS, 2025, 17 (01)
  • [28] Audio-visual speech synchrony measure for talking-face identity verification
    Bredin, Herve
    Chollet, Gerard
    2007 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL II, PTS 1-3, 2007, : 233 - +
  • [29] KAN-AV dataset for audio-visual face and speech analysis in the wild
    Kefalas, Triantafyllos
    Fotiadou, Eftychia
    Georgopoulos, Markos
    Panagakis, Yannis
    Ma, Pingchuan
    Petridis, Stavros
    Stafylakis, Themos
    Pantic, Maja
    IMAGE AND VISION COMPUTING, 2023, 140
  • [30] Speech2Talking-Face: Inferring and Driving a Face with Synchronized Audio-Visual Representation
    Sun, Yasheng
    Zhou, Hang
    Liu, Ziwei
    Koike, Hideki
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 1018 - 1024