Study of articulators’ contribution and compensation during speech by articulatory speech recognition

被引:0
|
作者
Jianguo Wei
Yan Ji
Jingshu Zhang
Qiang Fang
Wenhuan Lu
Kiyoshi Honda
Xugang Lu
机构
[1] Tianjin University,School of Computer Software
[2] Tianjin University,School of Computer Science and Technology
[3] Chinese Academy of Social Sciences,undefined
[4] NICT,undefined
来源
关键词
DNN; Articulatory recognition; Articulators’ contribution; Crucial level; Compensation;
D O I
暂无
中图分类号
学科分类号
摘要
In this paper, the contributions of dynamic articulatory information were evaluated by using an articulatory speech recognition system. The Electromagnetic Articulographic dataset is relatively small and hard to be recorded compared with popular speech corpora used for modern speech study. We used articulatory data to study the contribution of each observation channel of vocal tracts in speech recognition by DNN framework. We also analyzed the recognition results of each phoneme according to speech production rules. The contribution rate of each articulator can be considered as the crucial level of each phoneme in speech production. Furthermore, the results indicate that the contribution of each observation point is not relevant to a specific method. The tendency of a contribution of each sensor is identical to the rules of Japanese phonology. In this work, we also evaluated the compensation effect between different channels. We discovered that crucial points are hard to be compensated for compared with non-crucial points. The proposed method can help us identify the crucial points of each phoneme during speech. The results of this paper can contribute to the study of speech production and articulatory-based speech recognition.
引用
收藏
页码:18849 / 18864
页数:15
相关论文
共 50 条
  • [41] A FIBERSCOPIC STUDY OF ARTICULATORY ADJUSTMENTS IN ESOPHAGEAL SPEECH
    HIROSE, H
    SAWASHIMA, M
    [J]. FOLIA PHONIATRICA, 1983, 35 (3-4): : 135 - 136
  • [42] TELEVISED VISUAL CONTRIBUTION TO SPEECH RECOGNITION
    BROADBENT, D
    [J]. IEEE TRANSACTIONS ON EDUCATION, 1970, E 13 (02) : 79 - +
  • [43] DISCUSSION OF STUDIES OF ARTICULATORY ACTIVITY AND AIRFLOW DURING SPEECH
    BRUBAKER, RS
    KLATT, DH
    LADEFOGE.P
    STEVENS, KN
    KOIKE, Y
    [J]. ANNALS OF THE NEW YORK ACADEMY OF SCIENCES, 1968, 155 (A1) : 54 - &
  • [44] Speech-in-speech recognition: A training study
    Van Engen, Kristin J.
    [J]. LANGUAGE AND COGNITIVE PROCESSES, 2012, 27 (7-8): : 1089 - 1107
  • [45] Contextual variability during speech-in-speech recognition
    Brouwer, Susanne
    Bradlow, Ann R.
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2014, 136 (01): : EL26 - EL32
  • [46] Contextual variability during speech-in-speech recognition
    20142917942722
    [J]. Brouwer, S. (s.m.brouwer@uu.nl), 1600, Acoustical Society of America (136):
  • [47] Joint compensation of noise and channel in speech recognition
    Zhao, Rui
    Wang, Zuoying
    [J]. Shengxue Xuebao/Acta Acustica, 2006, 31 (05): : 466 - 470
  • [48] Channel compensation for robust telephone speech recognition
    Han, JQ
    Han, MS
    Gao, W
    [J]. IEEE TENCON'97 - IEEE REGIONAL 10 ANNUAL CONFERENCE, PROCEEDINGS, VOLS 1 AND 2: SPEECH AND IMAGE TECHNOLOGIES FOR COMPUTING AND TELECOMMUNICATIONS, 1997, : 169 - 172
  • [49] Compensation of Nonlinear Distortions in Speech for Automatic Recognition
    Malek, Jiri
    Silovsky, Jan
    Cerva, Petr
    Koldovsky, Zbynek
    Nouza, Jan
    Zdansky, Jindrich
    [J]. 2015 38TH INTERNATIONAL CONFERENCE ON TELECOMMUNICATIONS AND SIGNAL PROCESSING (TSP), 2015,
  • [50] Analysis and compensation of speech under stress and noise for environmental robustness in speech recognition
    Hansen, JHL
    [J]. SPEECH COMMUNICATION, 1996, 20 (1-2) : 151 - 173