Continuous Articulatory-to-Acoustic Mapping using Phone-based Trajectory HMM for a Silent Speech Interface

被引:0
|
作者
Hueber, Thomas [1 ]
Bailly, Gerard [1 ]
Denby, Bruce
机构
[1] UJF, U Stendhal, INP, GIPSA Lab,CNRS,UMR 5216, Grenoble, France
关键词
silent speech interface; handicap; HMM-based speech synthesis; audiovisual speech processing;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The article presents an HMM-based mapping approach for converting ultrasound and video images of the vocal tract into an audible speech signal, for a silent speech interface application. The proposed technique is based on the joint modeling of articulatory and spectral features, for each phonetic class, using Hidden Markov Models (HMM) and multivariate Gaussian distributions with full covariance matrices. The articulatory-to-acoustic mapping is achieved in 2 steps: 1) finding the most likely HMM state sequence from the articulatory observations; 2) inferring the spectral trajectories from both the decoded state sequence and the articulatory observations. The proposed technique is compared to our previous approach, in which only the decoded state sequence was used for the inference of the spectral trajectories, independently from the articulatory observations. Both objective and perceptual evaluations show that this new approach leads to a better estimation of the spectral trajectories.
引用
收藏
页码:722 / 725
页数:4
相关论文
共 50 条
  • [41] Articulatory Control of HMM-Based Parametric Speech Synthesis Using Feature-Space-Switched Multiple Regression
    Ling, Zhen-Hua
    Richmond, Korin
    Yamagishi, Junichi
    IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2013, 21 (01): : 205 - 217
  • [42] Enhancing HMM Based Malayalam Continuous Speech Recognizer Using Artificial Neural Networks
    Mohamed, Anuj
    Nair, K. N. Ramachandran
    COMPUTATIONAL INTELLIGENCE IN DATA MINING, VOL 2, 2015, 32 : 259 - 267
  • [43] RETRACTION: Articulatory-to-Acoustic Conversion Using BiLSTM-CNN Word-Attention-Based Method (Retraction of Vol 2020, art no 4356981, 2020)
    Ren, G.
    Shao, G.
    Fu, J.
    COMPLEXITY, 2024, 2024
  • [44] A hybrid HMM-based speech recognizer using kernel-based discriminants as acoustic models
    Andelic, Edin
    Schaffoener, Martin
    Katz, Marcel
    Krueger, Sven E.
    Wendemuth, Andreas
    18TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL 2, PROCEEDINGS, 2006, : 1158 - +
  • [45] Ultrasound-Based Silent Speech Interface Using Convolutional and Recurrent Neural Networks
    Moliner Juanpere, Eloi
    Csapo, Tamas Gabor
    ACTA ACUSTICA UNITED WITH ACUSTICA, 2019, 105 (04) : 587 - 590
  • [46] Speech recognition using non-linear trajectories in a formant-based articulatory layer of a multiple-level segmental HMM
    Hu, Hongwei
    Russell, Martin J.
    INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 2422 - 2425
  • [47] State-based Gaussian selection in large vocabulary continuous speech recognition using HMM's
    Gales, MJF
    Knill, KM
    Young, SJ
    IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, 1999, 7 (02): : 152 - 161
  • [48] Ultrasound-Based Silent Speech Interface using Sequential Convolutional Auto-encoder
    Xu, Kele
    Wu, Yuxiang
    Gao, Zhifeng
    PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 2194 - 2195
  • [49] Using Noisy Speech to Study the Robustness of a Continuous F0 Modelling Method in HMM-based Speech Synthesis
    Ogbureke, Kalu U.
    Cabral, Joao P.
    Carson-Berndsen, Julie
    PROCEEDINGS OF THE 6TH INTERNATIONAL CONFERENCE ON SPEECH PROSODY, VOLS I AND II, 2012, : 67 - 70
  • [50] Analysis of the Activity and Travel Patterns of the Elderly Using Mobile Phone-Based Hourly Locational Trajectory Data: Case Study of Gangnam, Korea
    Lee, Kwang-Sub
    Eom, Jin Ki
    Lee, Jun
    Ko, Sangpil
    SUSTAINABILITY, 2021, 13 (06)