Continuous Articulatory-to-Acoustic Mapping using Phone-based Trajectory HMM for a Silent Speech Interface

被引:0
|
作者
Hueber, Thomas [1 ]
Bailly, Gerard [1 ]
Denby, Bruce
机构
[1] UJF, U Stendhal, INP, GIPSA Lab,CNRS,UMR 5216, Grenoble, France
关键词
silent speech interface; handicap; HMM-based speech synthesis; audiovisual speech processing;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The article presents an HMM-based mapping approach for converting ultrasound and video images of the vocal tract into an audible speech signal, for a silent speech interface application. The proposed technique is based on the joint modeling of articulatory and spectral features, for each phonetic class, using Hidden Markov Models (HMM) and multivariate Gaussian distributions with full covariance matrices. The articulatory-to-acoustic mapping is achieved in 2 steps: 1) finding the most likely HMM state sequence from the articulatory observations; 2) inferring the spectral trajectories from both the decoded state sequence and the articulatory observations. The proposed technique is compared to our previous approach, in which only the decoded state sequence was used for the inference of the spectral trajectories, independently from the articulatory observations. Both objective and perceptual evaluations show that this new approach leads to a better estimation of the spectral trajectories.
引用
收藏
页码:722 / 725
页数:4
相关论文
共 50 条
  • [1] Autoencoder-Based Articulatory-to-Acoustic Mapping for Ultrasound Silent Speech Interfaces
    Gosztolya, Gabor
    Pinter, Adam
    Toth, Laszlo
    Grosz, Tamas
    Marko, Alexandra
    Csapo, Tamas Gabor
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [2] A stochastic articulatory-to-acoustic mapping as a basis for speech recognition
    Hogden, J
    Valdez, P
    IMTC/2001: PROCEEDINGS OF THE 18TH IEEE INSTRUMENTATION AND MEASUREMENT TECHNOLOGY CONFERENCE, VOLS 1-3: REDISCOVERING MEASUREMENT IN THE AGE OF INFORMATICS, 2001, : 1105 - 1110
  • [3] Ultrasound-based Articulatory-to-Acoustic Mapping with WaveGlow Speech Synthesis
    Csapo, Tamas Gabor
    Zainko, Csaba
    Toth, Laszlo
    Gosztolya, Gabor
    Marko, Alexandra
    INTERSPEECH 2020, 2020, : 2727 - 2731
  • [4] Statistical Mapping between Articulatory and Acoustic Data for an Ultrasound-based Silent Speech Interface
    Hueber, Thomas
    Benaroya, Elie-Laurent
    Denby, Bruce
    Chollet, Gerard
    12TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2011 (INTERSPEECH 2011), VOLS 1-5, 2011, : 600 - +
  • [5] Cross-speaker Acoustic-to-Articulatory Inversion using Phone-based Trajectory HAM for Pronunciation Training
    Hueber, Thomas
    Ben-Youssef, Atef
    Bailly, Gerard
    Badin, Pierre
    Elisei, Frederic
    13TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2012 (INTERSPEECH 2012), VOLS 1-3, 2012, : 782 - 785
  • [6] Phone-based speech synthesis with neural network and articulatory control
    Lo, WK
    Ching, PC
    ICSLP 96 - FOURTH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, PROCEEDINGS, VOLS 1-4, 1996, : 2227 - 2230
  • [7] DEEP-LEVEL ACOUSTIC-TO-ARTICULATORY MAPPING FOR DBN-HMM BASED PHONE RECOGNITION
    Badino, Leonardo
    Canevari, Claudia
    Fadiga, Luciano
    Metta, Giorgio
    2012 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2012), 2012, : 370 - 375
  • [8] Articulatory-to-Acoustic Conversion of Mandarin Emotional Speech Based on PSO-LSSVM
    Ren, Guofeng
    Fu, Jianmei
    Shao, Guicheng
    Xun, Yanqin
    COMPLEXITY, 2021, 2021
  • [9] Speech modelling based on acoustic-to-articulatory mapping
    Schoentgen, J
    NONLINEAR SPEECH MODELING AND APPLICATIONS, 2005, 3445 : 114 - 135
  • [10] A HMM Based Speech Synthesis Method Using Articulatory Feature
    Li, Yong
    Yin, Qing
    PROCEEDINGS OF 2019 IEEE 3RD INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC 2019), 2019, : 185 - 189