The effect of prior visual information on recognition of speech and sounds

被引:130
|
作者
Noppeney, Uta [1 ,2 ]
Josephs, Oliver [2 ]
Hocking, Julia [2 ]
Price, Cathy J. [2 ]
Friston, Karl J. [2 ]
机构
[1] Max Planck Inst Biol Cybernet, D-72076 Tubingen, Germany
[2] Inst Neurol, Wellcome Dept Imaging Neurosci, London WC1N 3BG, England
基金
英国惠康基金;
关键词
cross-modal priming; dynamic causal modeling; effective connectivity; multisensory integration; predictive coding; semantics;
D O I
10.1093/cercor/bhm091
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
To identify and categorize complex stimuli such as familiar objects or speech, the human brain integrates information that is abstracted at multiple levels from its sensory inputs. Using cross-modal priming for spoken words and sounds, this functional magnetic resonance imaging study identified 3 distinct classes of visuoauditory incongruency effects: visuoauditory incongruency effects were selective for 1) spoken words in the left superior temporal sulcus (STS), 2) environmental sounds in the left angular gyrus (AG), and 3) both words and sounds in the lateral and medial prefrontal cortices (IFS/mPFC). From a cognitive perspective, these incongruency effects suggest that prior visual information influences the neural processes underlying speech and sound recognition at multiple levels, with the STS being involved in phonological, AG in semantic, and mPFC/IFS in higher conceptual processing. In terms of neural mechanisms, effective connectivity analyses (dynamic causal modeling) suggest that these incongruency effects may emerge via greater bottom-up effects from early auditory regions to intermediate multisensory integration areas (i.e., STS and AG). This is consistent with a predictive coding perspective on hierarchical Bayesian inference in the cortex where the domain of the prediction error (phonological vs. semantic) determines its regional expression (middle temporal gyrus/STS vs. AG/intraparietal sulcus).
引用
收藏
页码:598 / 609
页数:12
相关论文
共 50 条
  • [1] Mechanisms of enhancing visual-speech recognition by prior auditory information
    Blank, Helen
    von Kriegstein, Katharina
    [J]. NEUROIMAGE, 2013, 65 : 109 - 118
  • [2] Visual speech information for face recognition
    Lawrence D. Rosenblum
    Deborah A. Yakel
    Naser Baseer
    Anjani Panchal
    Brynn C. Nodarse
    Ryan P. Niehus
    [J]. Perception & Psychophysics, 2002, 64 : 220 - 229
  • [3] Visual speech information for face recognition
    Rosenblum, LD
    Yakel, DA
    Baseer, N
    Panchal, A
    Nodarse, BC
    Niehus, RP
    [J]. PERCEPTION & PSYCHOPHYSICS, 2002, 64 (02): : 220 - 229
  • [4] EFFECT OF INTERAURAL SWITCHING ON RECOGNITION OF SPEECH SOUNDS
    RUPF, JA
    HUGHES, GW
    HOUSE, AS
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 1972, 51 (04): : 1304 - &
  • [5] Articulatory uncertainty as a result of visual recognition of modified sounds in Russian speech
    Myasoedova, M. A.
    Myasoedova, Z. P.
    Farkhadov, M. P.
    [J]. 2017 11TH IEEE INTERNATIONAL CONFERENCE ON APPLICATION OF INFORMATION AND COMMUNICATION TECHNOLOGIES (AICT 2017), 2017, : 169 - 172
  • [6] VISUAL CLASSIFICATION OF SPEECH SOUNDS
    KACZMAREK, BLJ
    [J]. CLINICAL LINGUISTICS & PHONETICS, 1990, 4 (03) : 247 - 252
  • [7] SIMULATION OF A RECOGNITION SYSTEM FOR CONNECTED SPEECH SOUNDS USING LINGUISTIC INFORMATION
    TAKEYA, SI
    KAWAGUCHI, E
    [J]. ELECTRONICS & COMMUNICATIONS IN JAPAN, 1973, 56 (09): : 38 - 46
  • [8] MODELING NONVERBAL SOUNDS FOR SPEECH RECOGNITION
    WARD, W
    [J]. SPEECH AND NATURAL LANGUAGE, 1989, : 47 - 50
  • [9] THE INTERVALGRAM AS A VISUAL PRESENTATION OF SPEECH SOUNDS
    CHANG, SH
    PIHL, GE
    WIREN, J
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 1951, 23 (05): : 632 - 632
  • [10] THE INTERVALGRAM AS A VISUAL REPRESENTATION OF SPEECH SOUNDS
    CHANG, SH
    PIHL, GE
    WIREN, J
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 1951, 23 (06): : 674 - 679