Musicians Show Improved Speech Segregation in Competitive, Multi-Talker Cocktail Party Scenarios

被引:27
|
作者
Bidelman, Gavin M. [1 ,2 ,3 ]
Yoo, Jessica [2 ]
机构
[1] Univ Memphis, Inst Intelligent Syst, Memphis, TN 38152 USA
[2] Univ Memphis, Sch Commun Sci & Disorders, Memphis, TN 38152 USA
[3] Univ Tennessee, Ctr Hlth Sci, Dept Anat & Neurobiol, Memphis, TN 38163 USA
来源
FRONTIERS IN PSYCHOLOGY | 2020年 / 11卷
基金
美国国家卫生研究院;
关键词
acoustic scene analysis; stream segregation; experience-dependent plasticity; musical training; speech-in-noise perception; IN-NOISE PERCEPTION; MUSICAL EXPERIENCE; BRAIN-STEM; ATTENTION; HEARING; REVERBERATION; BILINGUALISM; INTELLIGENCE; PLASTICITY; LISTENERS;
D O I
10.3389/fpsyg.2020.01927
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Studies suggest that long-term music experience enhances the brain's ability to segregate speech from noise. Musicians' "speech-in-noise (SIN) benefit" is based largely on perception from simple figure-ground tasks rather than competitive, multi-talker scenarios that offer realistic spatial cues for segregation and engage binaural processing. We aimed to investigate whether musicians show perceptual advantages in cocktail party speech segregation in a competitive, multi-talker environment. We used the coordinate response measure (CRM) paradigm to measure speech recognition and localization performance in musicians vs. non-musicians in a simulated 3D cocktail party environment conducted in an anechoic chamber. Speech was delivered through a 16-channel speaker array distributed around the horizontal soundfield surrounding the listener. Participants recalled the color, number, and perceived location of target callsign sentences. We manipulated task difficulty by varying the number of additional maskers presented at other spatial locations in the horizontal soundfield (0-1-2-3-4-6-8 multi-talkers). Musicians obtained faster and better speech recognition amidst up to around eight simultaneous talkers and showed less noise-related decline in performance with increasing interferers than their non-musician peers. Correlations revealed associations between listeners' years of musical training and CRM recognition and working memory. However, better working memory correlated with better speech streaming. Basic (QuickSIN) but not more complex (speech streaming) SIN processing was still predicted by music training after controlling for working memory. Our findings confirm a relationship between musicianship and naturalistic cocktail party speech streaming but also suggest that cognitive factors at least partially drive musicians' SIN advantage.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Multi-talker Speech Separation Based on Permutation Invariant Training and Beamforming
    Yin, Lu
    Wang, Ziteng
    Xia, Risheng
    Li, Junfeng
    Yan, Yonghong
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 851 - 855
  • [32] Text-aware Speech Separation for Multi-talker Keyword Spotting
    Li, Haoyu
    Yang, Baochen
    Xi, Yu
    Yu, Linfeng
    Tan, Tian
    Li, Hao
    Yu, Kai
    INTERSPEECH 2024, 2024, : 337 - 341
  • [33] Speech-derived haptic stimulation enhances speech recognition in a multi-talker background
    Rautu, I. Sabina
    De Tiege, Xavier
    Jousmaki, Veikko
    Bourguignon, Mathieu
    Bertels, Julie
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [34] Speech-derived haptic stimulation enhances speech recognition in a multi-talker background
    I. Sabina Răutu
    Xavier De Tiège
    Veikko Jousmäki
    Mathieu Bourguignon
    Julie Bertels
    Scientific Reports, 13
  • [35] Single-channel multi-talker speech recognition with permutation invariant training
    Qian, Yanmin
    Chang, Xuankai
    Yu, Dong
    SPEECH COMMUNICATION, 2018, 104 : 1 - 11
  • [36] Deep Neural Networks for Single-Channel Multi-Talker Speech Recognition
    Weng, Chao
    Yu, Dong
    Seltzer, Michael L.
    Droppo, Jasha
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2015, 23 (10) : 1670 - 1679
  • [37] MULTI-MICROPHONE NEURAL SPEECH SEPARATION FOR FAR-FIELD MULTI-TALKER SPEECH RECOGNITION
    Yoshioka, Takuya
    Erdogan, Hakan
    Chen, Zhuo
    Alleva, Fil
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 5739 - 5743
  • [38] The effects of selective attention and speech acoustics on neural speech-tracking in a multi-talker scene
    Rimmele, Johanna M.
    Golumbic, Elana Zion
    Schroeger, Erich
    Poeppel, David
    CORTEX, 2015, 68 : 144 - 154
  • [39] Super-human multi-talker speech recognition: A graphical modeling approach
    Hershey, John R.
    Rennie, Steven J.
    Olsen, Peder A.
    Kristjansson, Trausti T.
    COMPUTER SPEECH AND LANGUAGE, 2010, 24 (01): : 45 - 66
  • [40] SURT 2.0: Advances in Transducer-Based Multi-Talker Speech Recognition
    Raj, Desh
    Povey, Daniel
    Khudanpur, Sanjeev
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 3800 - 3813