A neural network model of the articulatory-acoustic forward mapping trained on recordings of articulatory parameters

被引:44
|
作者
Kello, CT [1 ]
Plaut, DC
机构
[1] George Mason Univ, Dept Psychol, Fairfax, VA 22030 USA
[2] Carnegie Mellon Univ, Dept Psychol, Ctr Neural Basis Cognit, Pittsburgh, PA 15213 USA
来源
关键词
D O I
10.1121/1.1715112
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Three neural network models were trained on the forward mapping from articulatory positions to acoustic outputs for a single speaker of the Edinburgh multi-channel articulatory speech database. The model parameters (i.e., connection weights) were learned via the backpropagation of error signals generated by the difference between acoustic outputs of the models, and their acoustic targets. Efficacy of the trained models was assessed by subjecting the models' acoustic outputs to speech intelligibility tests. The results of these tests showed that enough phonetic information was captured by the models to support rates of word identification as high as 84%, approaching an identification rate of 92% for the actual target stimuli. These forward models could serve as one component of a data-driven articulatory synthesizer. The models also provide the first step toward building a model of spoken word acquisition and phonological development trained on real speech. (C) 2004 Acoustical Society of America.
引用
收藏
页码:2354 / 2364
页数:11
相关论文
共 50 条
  • [1] A neural network model of the articulatory-acoustic forward mapping trained on recordings of articulatory parameters
    Kello, Christopher T.
    Plaut, David C.
    Journal of the Acoustical Society of America, 2004, 116 (4 I): : 2354 - 2364
  • [2] Acoustic to articulatory mapping with deep neural network
    Zhiyong Wu
    Kai Zhao
    Xixin Wu
    Xinyu Lan
    Helen Meng
    Multimedia Tools and Applications, 2015, 74 : 9889 - 9907
  • [3] Acoustic to articulatory mapping with deep neural network
    Wu, Zhiyong
    Zhao, Kai
    Wu, Xixin
    Lan, Xinyu
    Meng, Helen
    MULTIMEDIA TOOLS AND APPLICATIONS, 2015, 74 (22) : 9889 - 9907
  • [4] An Analysis of Articulatory-Acoustic Data based on Articulatory Strokes
    Kato, Tsuneo
    Lee, Sungbok
    Narayanan, Shrikanth
    2009 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS 1- 8, PROCEEDINGS, 2009, : 4493 - 4496
  • [5] Recognition of articulatory-acoustic deviations in pathological speech
    Bilibajkic, Ruzica B.
    2016 24TH TELECOMMUNICATIONS FORUM (TELFOR), 2016, : 395 - 402
  • [6] MAPPING ARABIC ACOUSTIC PARAMETERS TO THEIR ARTICULATORY FEATURES USING NEURAL NETWORKS
    Alotaibi, Yousef Ajami
    Seddiq, Yasser Mohammad
    2015 IEEE SIGNAL PROCESSING AND SIGNAL PROCESSING EDUCATION WORKSHOP (SP/SPE), 2015, : 409 - 414
  • [7] Articulatory-acoustic kinematics: The production of American English /s/
    Iskarous, Khalil
    Shadle, Christine H.
    Proctor, Michael I.
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2011, 129 (02): : 944 - 954
  • [8] Learning Joint Articulatory-Acoustic Representations with Normalizing Flows
    Saha, Pramit
    Fels, Sidney
    INTERSPEECH 2020, 2020, : 3196 - 3200
  • [10] Relevance-Weighted-Reconstruction of Articulatory Features in Deep-Neural-Network-Based Acoustic-to-Articulatory Mapping
    Canevari, Claudia
    Badino, Leonardo
    Fadiga, Luciano
    Metta, Giorgio
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 1296 - 1300