The early maximum likelihood estimation model of audiovisual integration in speech perception

被引:9
|
作者
Andersen, Tobias S. [1 ]
机构
[1] Tech Univ Denmark, Sect Cognit Syst, Dept Appl Math & Comp Sci, DK-2800 Lyngby, Denmark
来源
关键词
FUZZY-LOGICAL MODEL; CROSSMODAL INTEGRATION; GOOD FIT; INFORMATION; FUSION;
D O I
10.1121/1.4916691
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Speech perception is facilitated by seeing the articulatory mouth movements of the talker. This is due to perceptual audiovisual integration, which also causes the McGurk - MacDonald illusion, and for which a comprehensive computational account is still lacking. Decades of research have largely focused on the fuzzy logical model of perception (FLMP), which provides excellent fits to experimental observations but also has been criticized for being too flexible, post hoc and difficult to interpret. The current study introduces the early maximum likelihood estimation (MLE) model of audiovisual integration to speech perception along with three model variations. In early MLE, integration is based on a continuous internal representation before categorization, which can make the model more parsimonious by imposing constraints that reflect experimental designs. The study also shows that cross-validation can evaluate models of audiovisual integration based on typical data sets taking both goodness-of-fit and model flexibility into account. All models were tested on a published data set previously used for testing the FLMP. Cross-validation favored the early MLE while more conventional error measures favored more complex models. This difference between conventional error measures and cross-validation was found to be indicative of over-fitting in more complex models such as the FLMP. (C) 2015 Acoustical Society of America.
引用
收藏
页码:2884 / 2891
页数:8
相关论文
共 50 条
  • [1] Integration of audiovisual spatial signals is not consistent with maximum likelihood estimation
    Meijer, David
    Veselic, Sebastijan
    Calafiore, Carmelo
    Noppeney, Uta
    [J]. CORTEX, 2019, 119 : 74 - 88
  • [2] Automatic audiovisual integration in speech perception
    Gentilucci, M
    Cattaneo, L
    [J]. EXPERIMENTAL BRAIN RESEARCH, 2005, 167 (01) : 66 - 75
  • [3] Automatic audiovisual integration in speech perception
    Maurizio Gentilucci
    Luigi Cattaneo
    [J]. Experimental Brain Research, 2005, 167 : 66 - 75
  • [4] Perception based method for the investigation of audiovisual integration of speech
    Huhn, Zsofia
    Szirtes, Gabor
    Lorincz, Andras
    Csepe, Valeria
    [J]. NEUROSCIENCE LETTERS, 2009, 465 (03) : 204 - 209
  • [5] Modeling the Development of Audiovisual Cue Integration in Speech Perception
    Getz, Laura M.
    Nordeen, Elke R.
    Vrabic, Sarah C.
    Toscano, Joseph C.
    [J]. BRAIN SCIENCES, 2017, 7 (03):
  • [6] Schizotypal traits are not related to multisensory integration or audiovisual speech perception
    Muller, Anne-Marie
    Dalal, Tyler C.
    Stevenson, Ryan A.
    [J]. CONSCIOUSNESS AND COGNITION, 2020, 86
  • [7] Audiovisual speech perception
    Sams, M.
    [J]. PERCEPTION, 1997, 26 : 78 - 78
  • [8] MAXIMUM SMOOTHED LIKELIHOOD ESTIMATION AND SMOOTHED MAXIMUM LIKELIHOOD ESTIMATION IN THE CURRENT STATUS MODEL
    Groeneboom, Piet
    Jongbloed, Geurt
    Witte, Birgit I.
    [J]. ANNALS OF STATISTICS, 2010, 38 (01): : 352 - 387
  • [9] MAXIMUM LIKELIHOOD ESTIMATION WITH WEIBULL MODEL
    ROCKETTE, H
    ANTLE, C
    KLIMKO, LA
    [J]. JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1974, 69 (345) : 246 - 249
  • [10] The Contribution of Proprioceptive and Cutaneous Cues in Weight Perception: Early Evidence for Maximum-Likelihood Integration
    Giachritsis, Christos
    Wright, Rachel
    Wing, Alan
    [J]. HAPTICS: GENERATING AND PERCEIVING TANGIBLE SENSATIONS, PT I, PROCEEDINGS, 2010, 6191 : 11 - 16