Monaural speech intelligibility and detection in maskers with varying amounts of spectro-temporal speech features

被引:28
|
作者
Schubotz, Wiebke
Brand, Thomas
Kollmeier, Birger
Ewert, Stephan D. [1 ]
机构
[1] Carl von Ossietzky Univ Oldenburg, Med Phys, D-26111 Oldenburg, Germany
来源
关键词
COMODULATION MASKING RELEASE; HEARING-IMPAIRED LISTENERS; INFORMATIONAL MASKING; FLUCTUATING NOISE; FREQUENCY-SELECTIVITY; SIMULTANEOUS TALKERS; RECEPTION THRESHOLD; PERCEPTION; INDEX; SEPARATION;
D O I
10.1121/1.4955079
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Speech intelligibility is strongly affected by the presence of maskers. Depending on the spectro-temporal structure of the masker and its similarity to the target speech, different masking aspects can occur which are typically referred to as energetic, amplitude modulation, and informational masking. In this study speech intelligibility and speech detection was measured in maskers that vary systematically in the time-frequency domain from steady-state noise to a single interfering talker. Male and female target speech was used in combination with maskers based on speech for the same or different gender. Observed data were compared to predictions of the speech intelligibility index, extended speech intelligibility index, multi-resolution speech-based envelope-power-spectrum model, and the short-time objective intelligibility measure. The different models served as analysis tool to help distinguish between the different masking aspects. Comparison shows that overall masking can to a large extent be explained by short-term energetic masking. However, the other masking aspects (amplitude modulation an informational masking) influence speech intelligibility as well. Additionally, it was obvious that all models showed considerable deviations from the data. Therefore, the current study provides a benchmark for further evaluation of speech prediction models. (C) 2016 Acoustical Society of America.
引用
收藏
页码:524 / 540
页数:17
相关论文
共 50 条
  • [1] Spectro-Temporal Representation of Speech for Intelligibility Assessment of Dysarthria
    Chandrashekar, H. M.
    Karjigi, Veena
    Sreedevi, N.
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (02) : 390 - 399
  • [2] Spectro-temporal modulation glimpsing for speech intelligibility prediction
    Edraki, Amin
    Chan, Wai-Yip
    Jensen, Jesper
    Fogerty, Daniel
    HEARING RESEARCH, 2022, 426
  • [3] Spectro-temporal modulation transfer functions and speech intelligibility
    Chi, TS
    Gao, YJ
    Guyton, MC
    Ru, PW
    Shamma, S
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 1999, 106 (05): : 2719 - 2732
  • [4] Development of spectro-temporal features of speech in children
    Gautam S.
    Singh L.
    Gautam, Sumanlata (suman.gautam82@gmail.com), 1600, Springer Science and Business Media, LLC (20): : 543 - 551
  • [5] A spectro-temporal modulation index (STMI) for assessment of speech intelligibility
    Elhilali, M
    Chi, T
    Shamma, SA
    SPEECH COMMUNICATION, 2003, 41 (2-3) : 331 - 348
  • [6] A Spectro-Temporal Glimpsing Index (STGI) for Speech Intelligibility Prediction
    Edraki, Amin
    Chan, Wai-Yip
    Jensen, Jesper
    Fogerty, Daniel
    INTERSPEECH 2021, 2021, : 206 - 210
  • [7] Speech Intelligibility Prediction Using Spectro-Temporal Modulation Analysis
    Edraki, Amin
    Chan, Wai-Yip
    Jensen, Jesper
    Fogerty, Daniel
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 : 210 - 225
  • [8] Improvement and Assessment of Spectro-Temporal Modulation Analysis for Speech Intelligibility Estimation
    Edraki, Amin
    Chan, Wai-Yip
    Jensen, Jesper
    Fogerty, Daniel
    INTERSPEECH 2019, 2019, : 1378 - 1382
  • [9] Hierarchical spectro-temporal features for robust speech recognition
    Domont, Xavier
    Heckmann, Martin
    Joublin, Frank
    Goerick, Christian
    2008 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, VOLS 1-12, 2008, : 4417 - 4420
  • [10] Spectro-Temporal SubNet for Real-Time Monaural Speech Denoising and Dereverberation
    Xiong, Feifei
    Chen, Weiguang
    Wang, Pengyu
    Li, Xiaofei
    Feng, Jinwei
    INTERSPEECH 2022, 2022, : 931 - 935