ARTICULATORY FEATURES FOR EXPRESSIVE SPEECH SYNTHESIS

被引:0
|
作者
Black, Alan W. [1 ]
Bunnell, H. Timothy [2 ]
Dou, Ying [3 ]
Muthukumar, Prasanna Kumar [1 ]
Metze, Florian [1 ]
Perry, Daniel [4 ]
Polzehl, Tim [5 ]
Prahallad, Kishore [6 ]
Steidl, Stefan [7 ]
Vaughn, Callie [8 ]
机构
[1] Carnegie Mellon Univ, Language Technol Inst, Pittsburgh, PA 15213 USA
[2] Nemours Biomed Res, Wilmington, DE USA
[3] Johns Hopkins Univ, Baltimore, MD 21218 USA
[4] Univ Calif Los Angeles, Los Angeles, CA 90024 USA
[5] Tech Univ Berlin, Deutsche Telekom Lab, Berlin, Germany
[6] Int Inst Informat Technol, Hyderabad, Andhra Pradesh, India
[7] Int Comp Sci Inst, Berkeley, CA USA
[8] Oberlin Coll, Oberlin, OH 44074 USA
关键词
speech synthesis; articulatory features; emotional speech; meta-data extraction; evaluation;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This paper describes some of the results from the project entitled "New Parameterization for Emotional Speech Synthesis" held at the Summer 2011 JHU CLSP workshop. We describe experiments on how to use articulatory features as a meaningful intermediate representation for speech synthesis. This parameterization not only allows us to reproduce natural sounding speech but also allows us to generate stylistically varying speech. We show methods for deriving articulatory features from speech, predicting articulatory features from text and reconstructing natural sounding speech from the predicted articulatory features. The methods were tested on clean speech databases in English and German, as well as databases of emotionally and personality varying speech. The resulting speech was evaluated both objectively, using techniques normally used for emotion identification, and subjectively, using crowd-sourcing.
引用
收藏
页码:4005 / 4008
页数:4
相关论文
共 50 条
  • [1] On the Contribution of Articulatory Features to Speech Synthesis
    Matura, Martin
    Juzova, Marketa
    Matousek, Jindrich
    [J]. SPEECH AND COMPUTER (SPECOM 2018), 2018, 11096 : 398 - 407
  • [2] Effect of articulatory and acoustic features on the intelligibility of speech in noise: An articulatory synthesis study
    Thuanvan Ngo
    Akagi, Masato
    Birkholz, Peter
    [J]. SPEECH COMMUNICATION, 2020, 117 : 13 - 20
  • [3] ARTICULATORY TIMING IN SPEECH PRODUCTION . ARTICULATORY DISTINCTIVE FEATURES
    COKER, CH
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 1970, 47 (1P1): : 94 - +
  • [4] Articulatory features for speech-driven head motion synthesis
    Ben-Youssef, Atef
    Shimodaira, Hiroshi
    Braude, David A.
    [J]. 14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 2757 - 2761
  • [5] Articulatory Features for "Meeting" Speech Recognition
    Metze, Florian
    [J]. INTERSPEECH 2006 AND 9TH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, VOLS 1-5, 2006, : 581 - 584
  • [6] Measurements of articulatory variation in expressive speech for a set of Swedish vowels
    Nordstrand, M
    Svanfeldt, G
    Granström, B
    House, D
    [J]. SPEECH COMMUNICATION, 2004, 44 (1-4) : 187 - 196
  • [7] Articulatory Features for ASR of Pathological Speech
    Yilmaz, Emre
    Mitra, Vikramjit
    Bartels, Chris
    Franco, Horacio
    [J]. 19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 2958 - 2962
  • [8] Integrating Articulatory Features Into HMM-Based Parametric Speech Synthesis
    Ling, Zhen-Hua
    Richmond, Korin
    Yamagishi, Junichi
    Wang, Ren-Hua
    [J]. IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2009, 17 (06): : 1171 - 1185
  • [9] Towards Expressive Speech Synthesis: Analysis and Modeling of Expressive Speech
    Raptis, Spyros
    Karabetsos, Sotiris
    Chalamandaris, Aimilios
    Tsiakoulis, Pirros
    [J]. 2014 5th IEEE Conference on Cognitive Infocommunications (CogInfoCom), 2014, : 461 - 465
  • [10] Speech recognition using cepstral articulatory features
    Najnin, Shamima
    Banerjee, Bonny
    [J]. SPEECH COMMUNICATION, 2019, 107 : 26 - 37