Automatic Modelling of Depressed Speech: Relevant Features and Relevance of Gender

被引:0
|
作者
Hoenig, Florian [1 ]
Batliner, Anton [1 ,2 ]
Noeth, Elmar [1 ,3 ]
Schnieder, Sebastian [4 ]
Krajewski, Jarek [4 ]
机构
[1] Friedrich Alexander Univ Erlangen Nurnberg, Pattern Recognit Lab, Erlangen, Germany
[2] Tech Univ Munich, Inst Human Machine Commun, Munich, Germany
[3] King Abdulaziz Univ, Elect & Comp Engn Dept, Jeddah, Saudi Arabia
[4] Univ Wuppertal, Expt Ind Psychol, Wuppertal, Germany
关键词
depression; acoustic features; brute forcing; interpretation; paralinguistics; CLASSIFICATION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Depression is an affective disorder characterised by psychomotor retardation; in speech, this shows up in reduction of pitch (variation, range), loudness, and tempo, and in voice qualities different from those of typical modal speech. A similar reduction can be observed in sleepy speech (relaxation). In this paper, we employ a small group of acoustic features modelling prosody and spectrum that have been proven successful in the modelling of sleepy speech, enriched with voice quality features, for the modelling of depressed speech within a regression approach. This knowledge-based approach is complemented by and compared with brute-forcing and automatic feature selection. We further discuss gender differences and the contributions of (groups of) features both for the modelling of depression and across depression and sleepiness.
引用
收藏
页码:1248 / 1252
页数:5
相关论文
共 50 条
  • [41] Perception and modelling of vowels and vocal gender in synthetic speech
    Mannell, R.
    [J]. AUSTRALIAN JOURNAL OF PSYCHOLOGY, 2006, 58 : 9 - 9
  • [42] A STUDY ON ROBUSTNESS OF ARTICULATORY FEATURES FOR AUTOMATIC SPEECH RECOGNITION OF NEUTRAL AND WHISPERED SPEECH
    Srinivasan, Gokul
    Illa, Aravind
    Ghosh, Prasanta Kumar
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 5936 - 5940
  • [43] On the Correlation and Transferability of Features between Automatic Speech Recognition and Speech Emotion Recognition
    Fayek, Haytham M.
    Lech, Margaret
    Cavedon, Lawrence
    [J]. 17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 3618 - 3622
  • [44] Unveiling relevant acoustic features for bird species automatic classification
    Ugarte, Juan P.
    Arias-Arias, Jose
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2024, 257
  • [45] Automatic Speech-Based Classification of Gender, Age and Accent
    Phuoc Nguyen
    Tran, Dat
    Huang, Xu
    Sharma, Dharmendra
    [J]. KNOWLEDGE MANAGEMENT AND ACQUISITION FOR SMART SYSTEMS AND SERVICES, 2010, 6232 : 288 - 299
  • [46] An Automatic Diagnosis and Assessment of Dysarthric Speech using Speech Disorder Specific Prosodic Features
    Vyas, Garima
    Dutta, Malay Kishore
    Prinosil, Jiri
    Harar, Pavol
    [J]. 2016 39TH INTERNATIONAL CONFERENCE ON TELECOMMUNICATIONS AND SIGNAL PROCESSING (TSP), 2016, : 515 - 518
  • [47] Noise-Robust Algorithm of Speech Features Extraction for Automatic Speech Recognition System
    Yakhnev, A. N.
    Pisarev, A. S.
    [J]. PROCEEDINGS OF THE XIX IEEE INTERNATIONAL CONFERENCE ON SOFT COMPUTING AND MEASUREMENTS (SCM 2016), 2016, : 206 - 208
  • [48] A Further Step Towards Automatic Domain Modelling by Relevant Information Extraction
    Krupp, Lars
    Bahle, Gernot
    Gruenerbl, Agnes
    Lukowicz, Paul
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON PERVASIVE COMPUTING AND COMMUNICATIONS WORKSHOPS (PERCOM WORKSHOPS), 2020,
  • [49] Listening in the dips: Comparing relevant features for speech recognition in humans and machines
    Spille, Constantin
    Meyer, Bernd T.
    [J]. 18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 2968 - 2972
  • [50] Vocal tract length invariant features for automatic speech recognition
    Mertins, A
    Rademacher, J
    [J]. 2005 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2005, : 308 - 312