Musicians Show Improved Speech Segregation in Competitive, Multi-Talker Cocktail Party Scenarios

被引:27
|
作者
Bidelman, Gavin M. [1 ,2 ,3 ]
Yoo, Jessica [2 ]
机构
[1] Univ Memphis, Inst Intelligent Syst, Memphis, TN 38152 USA
[2] Univ Memphis, Sch Commun Sci & Disorders, Memphis, TN 38152 USA
[3] Univ Tennessee, Ctr Hlth Sci, Dept Anat & Neurobiol, Memphis, TN 38163 USA
来源
FRONTIERS IN PSYCHOLOGY | 2020年 / 11卷
基金
美国国家卫生研究院;
关键词
acoustic scene analysis; stream segregation; experience-dependent plasticity; musical training; speech-in-noise perception; IN-NOISE PERCEPTION; MUSICAL EXPERIENCE; BRAIN-STEM; ATTENTION; HEARING; REVERBERATION; BILINGUALISM; INTELLIGENCE; PLASTICITY; LISTENERS;
D O I
10.3389/fpsyg.2020.01927
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Studies suggest that long-term music experience enhances the brain's ability to segregate speech from noise. Musicians' "speech-in-noise (SIN) benefit" is based largely on perception from simple figure-ground tasks rather than competitive, multi-talker scenarios that offer realistic spatial cues for segregation and engage binaural processing. We aimed to investigate whether musicians show perceptual advantages in cocktail party speech segregation in a competitive, multi-talker environment. We used the coordinate response measure (CRM) paradigm to measure speech recognition and localization performance in musicians vs. non-musicians in a simulated 3D cocktail party environment conducted in an anechoic chamber. Speech was delivered through a 16-channel speaker array distributed around the horizontal soundfield surrounding the listener. Participants recalled the color, number, and perceived location of target callsign sentences. We manipulated task difficulty by varying the number of additional maskers presented at other spatial locations in the horizontal soundfield (0-1-2-3-4-6-8 multi-talkers). Musicians obtained faster and better speech recognition amidst up to around eight simultaneous talkers and showed less noise-related decline in performance with increasing interferers than their non-musician peers. Correlations revealed associations between listeners' years of musical training and CRM recognition and working memory. However, better working memory correlated with better speech streaming. Basic (QuickSIN) but not more complex (speech streaming) SIN processing was still predicted by music training after controlling for working memory. Our findings confirm a relationship between musicianship and naturalistic cocktail party speech streaming but also suggest that cognitive factors at least partially drive musicians' SIN advantage.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Chinese speech identification in multi-talker babble with diotic and dichotic listening
    PENG JianXin 1
    2 Department of Architecture
    ChineseScienceBulletin, 2012, 57 (20) : 2561 - 2566
  • [22] Chinese speech identification in multi-talker babble with diotic and dichotic listening
    Peng JianXin
    Zhang HongHu
    Wang ZiYou
    CHINESE SCIENCE BULLETIN, 2012, 57 (20): : 2548 - 2553
  • [23] EFFECTS OF MULTI-TALKER COMPETING SPEECH ON THE VARIABILITY OF THE CALIFORNIA CONSONANT TEST
    SURR, RK
    SCHWARTZ, DM
    EAR AND HEARING, 1980, 1 (06): : 319 - 323
  • [24] Hierarchical Variational Loopy Belief Propagation for Multi-talker Speech Recognition
    Rennie, Steven J.
    Hershey, John R.
    Olsen, Peder A.
    2009 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION & UNDERSTANDING (ASRU 2009), 2009, : 176 - 181
  • [25] Selective cortical representation of attended speaker in multi-talker speech perception
    Nima Mesgarani
    Edward F. Chang
    Nature, 2012, 485 : 233 - 236
  • [26] Speaker Identification in Multi-Talker Overlapping Speech Using Neural Networks
    Tran, Van-Thuan
    Tsai, Wei-Ho
    IEEE ACCESS, 2020, 8 : 134868 - 134879
  • [27] Effects of face masks on speech recognition in multi-talker babble noise
    Toscano, Joseph C.
    Toscano, Cheyenne M.
    PLOS ONE, 2021, 16 (02):
  • [28] USING BINARUAL PROCESSING FOR AUTOMATIC SPEECH RECOGNITION IN MULTI-TALKER SCENES
    Spille, Constantin
    Dietz, Mathias
    Hohmann, Volker
    Meyer, Bernd T.
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 7805 - 7809
  • [29] Selective cortical representation of attended speaker in multi-talker speech perception
    Mesgarani, Nima
    Chang, Edward F.
    NATURE, 2012, 485 (7397) : 233 - U118
  • [30] Auditory spatial cuing for speech perception in a dynamic multi-talker environment
    Tomoriova, Beata
    Kopco, Norbert
    2008 6TH INTERNATIONAL SYMPOSIUM ON APPLIED MACHINE INTELLIGENCE AND INFORMATICS, 2008, : 230 - 233