Emotion and Word Recognition for Unprocessed and Vocoded Speech Stimuli

被引:2
|
作者
Morgan, Shae D. [1 ]
Garrard, Stacy [1 ]
Hoskins, Tiffany [1 ]
机构
[1] Univ Louisville, Dept Otolaryngol Head & Neck Surg & Commun Disord, 627 S Preston St,Suite 220, Louisville, KY 40241 USA
来源
EAR AND HEARING | 2022年 / 43卷 / 02期
关键词
Cochlear implants; Emotion; Speech recognition; Vocoder; Word recognition; COCHLEAR IMPLANT USERS; FACIAL EXPRESSIONS; AFFECTIVE PROSODY; VOCAL EXPRESSION; IDENTIFICATION; PERFORMANCE; PERCEPTION; CUES; COMMUNICATION; COMPREHENSION;
D O I
10.1097/AUD.0000000000001100
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Objectives: Individuals with cochlear implants (CIs) show reduced word and auditory emotion recognition abilities relative to their peers with normal hearing. Modern CI processing strategies are designed to preserve acoustic cues requisite for word recognition rather than those cues required for accessing other signal information (e.g., talker gender or emotional state). While word recognition is undoubtedly important for communication, the inaccessibility of this additional signal information in speech may lead to negative social experiences and outcomes for individuals with hearing loss. This study aimed to evaluate whether the emphasis on word recognition preservation in CI processing has unintended consequences on the perception of other talker information, such as emotional state. Design: Twenty-four young adult listeners with normal hearing listened to sentences and either reported a target word in each sentence (word recognition task) or selected the emotion of the talker (emotion recognition task) from a list of options (Angry, Calm, Happy, and Sad). Sentences were blocked by task type (emotion recognition versus word recognition) and processing condition (unprocessed versus 8-channel noise vocoder) and presented randomly within the block at three signal-to-noise ratios (SNRs) in a background of speech-shaped noise. Confusion matrices showed the number of errors in emotion recognition by listeners. Results: Listeners demonstrated better emotion recognition performance than word recognition performance at the same SNR. Unprocessed speech resulted in higher recognition rates than vocoded stimuli. Recognition performance (for both words and emotions) decreased with worsening SNR. Vocoding speech resulted in a greater negative impact on emotion recognition than it did for word recognition. Conclusions: These data confirm prior work that suggests that in background noise, emotional prosodic information in speech is easier to recognize than word information, even after simulated CI processing. However, emotion recognition may be more negatively impacted by background noise and CI processing than word recognition. Future work could explore CI processing strategies that better encode prosodic information and investigate this effect in individuals with CIs as opposed to vocoded simulation. This study emphasized the need for clinicians to consider not only word recognition but also other aspects of speech that are critical to successful social communication.
引用
收藏
页码:398 / 407
页数:10
相关论文
共 50 条
  • [1] Speech Emotion Recognition Using Speech Feature and Word Embedding
    Atmaja, Bagus Tris
    Shirai, Kiyoaki
    Akagi, Masato
    [J]. 2019 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2019, : 519 - 523
  • [2] Toddlers' recognition of noise-vocoded speech
    Newman, Rochelle
    Chatterjee, Monita
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2013, 133 (01): : 483 - 494
  • [3] COMBINING SPEECH RECOGNITION AND ACOUSTIC WORD EMOTION MODELS FOR ROBUST TEXT-INDEPENDENT EMOTION RECOGNITION
    Schuller, Bjoern
    Vlasenko, Bogdan
    Arsic, Dejan
    Rigoll, Gerhard
    Wendemuth, Andreas
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOLS 1-4, 2008, : 1333 - +
  • [4] Word segmentation from noise-band vocoded speech
    Grieco-Calub, Tina M.
    Simeon, Katherine M.
    Snyder, Hillary E.
    Lew-Williams, Casey
    [J]. LANGUAGE COGNITION AND NEUROSCIENCE, 2017, 32 (10) : 1344 - 1356
  • [5] Effects of sleep loss on emotion recognition: a dissociation between face and word stimuli
    Lisa Maccari
    Diana Martella
    Andrea Marotta
    Mara Sebastiani
    Nerisa Banaj
    Luis J. Fuentes
    Maria Casagrande
    [J]. Experimental Brain Research, 2014, 232 : 3147 - 3157
  • [6] Effects of sleep loss on emotion recognition: a dissociation between face and word stimuli
    Maccari, Lisa
    Martella, Diana
    Marotta, Andrea
    Sebastiani, Mara
    Banaj, Nerisa
    Fuentes, Luis J.
    Casagrande, Maria
    [J]. EXPERIMENTAL BRAIN RESEARCH, 2014, 232 (10) : 3147 - 3157
  • [7] Dimensional speech emotion recognition from speech features and word embeddings by using multitask learning
    Atmaja, Bagus Tris
    Akagi, Masato
    [J]. APSIPA TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING, 2020, 9
  • [8] Speech Emotion Recognition
    Lalitha, S.
    Madhavan, Abhishek
    Bhushan, Bharath
    Saketh, Srinivas
    [J]. 2014 INTERNATIONAL CONFERENCE ON ADVANCES IN ELECTRONICS, COMPUTERS AND COMMUNICATIONS (ICAECC), 2014,
  • [9] Cognitive disruption by noise-vocoded speech stimuli: Effects of spectral variation
    Senan, Toros Ufuk
    Jelfs, Sam
    Kohlrausch, Armin
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2018, 143 (03): : 1407 - 1416
  • [10] Dichotic Speech Recognition Using CVC Word and Nonsense CVC Syllable Stimuli
    Findlen, Ursula M.
    Roup, Christina M.
    [J]. JOURNAL OF THE AMERICAN ACADEMY OF AUDIOLOGY, 2011, 22 (01) : 13 - 22