Musicians Show Improved Speech Segregation in Competitive, Multi-Talker Cocktail Party Scenarios

被引:27
|
作者
Bidelman, Gavin M. [1 ,2 ,3 ]
Yoo, Jessica [2 ]
机构
[1] Univ Memphis, Inst Intelligent Syst, Memphis, TN 38152 USA
[2] Univ Memphis, Sch Commun Sci & Disorders, Memphis, TN 38152 USA
[3] Univ Tennessee, Ctr Hlth Sci, Dept Anat & Neurobiol, Memphis, TN 38163 USA
来源
FRONTIERS IN PSYCHOLOGY | 2020年 / 11卷
基金
美国国家卫生研究院;
关键词
acoustic scene analysis; stream segregation; experience-dependent plasticity; musical training; speech-in-noise perception; IN-NOISE PERCEPTION; MUSICAL EXPERIENCE; BRAIN-STEM; ATTENTION; HEARING; REVERBERATION; BILINGUALISM; INTELLIGENCE; PLASTICITY; LISTENERS;
D O I
10.3389/fpsyg.2020.01927
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Studies suggest that long-term music experience enhances the brain's ability to segregate speech from noise. Musicians' "speech-in-noise (SIN) benefit" is based largely on perception from simple figure-ground tasks rather than competitive, multi-talker scenarios that offer realistic spatial cues for segregation and engage binaural processing. We aimed to investigate whether musicians show perceptual advantages in cocktail party speech segregation in a competitive, multi-talker environment. We used the coordinate response measure (CRM) paradigm to measure speech recognition and localization performance in musicians vs. non-musicians in a simulated 3D cocktail party environment conducted in an anechoic chamber. Speech was delivered through a 16-channel speaker array distributed around the horizontal soundfield surrounding the listener. Participants recalled the color, number, and perceived location of target callsign sentences. We manipulated task difficulty by varying the number of additional maskers presented at other spatial locations in the horizontal soundfield (0-1-2-3-4-6-8 multi-talkers). Musicians obtained faster and better speech recognition amidst up to around eight simultaneous talkers and showed less noise-related decline in performance with increasing interferers than their non-musician peers. Correlations revealed associations between listeners' years of musical training and CRM recognition and working memory. However, better working memory correlated with better speech streaming. Basic (QuickSIN) but not more complex (speech streaming) SIN processing was still predicted by music training after controlling for working memory. Our findings confirm a relationship between musicianship and naturalistic cocktail party speech streaming but also suggest that cognitive factors at least partially drive musicians' SIN advantage.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Audio-Visual Multi-Talker Speech Recognition in A Cocktail Party
    Wu, Yifei
    Hi, Chenda
    Yang, Song
    Wu, Zhongqin
    Qian, Yanmin
    INTERSPEECH 2021, 2021, : 3021 - 3025
  • [2] The effects of speech processing units on auditory stream segregation and selective attention in a multi-talker (cocktail party) situation
    Toth, Brigitta
    Honbolygo, Ferenc
    Szalardy, Orsolya
    Orosz, Gabor
    Farkas, David
    Winkler, Istvan
    CORTEX, 2020, 130 : 387 - 400
  • [3] The cocktail-party problem revisited: early processing and selection of multi-talker speech
    Adelbert W. Bronkhorst
    Attention, Perception, & Psychophysics, 2015, 77 : 1465 - 1487
  • [4] The cocktail-party problem revisited: early processing and selection of multi-talker speech
    Bronkhorst, Adelbert W.
    ATTENTION PERCEPTION & PSYCHOPHYSICS, 2015, 77 (05) : 1465 - 1487
  • [5] Speech prosody supports speaker selection and auditory stream segregation in a multi-talker situation
    Kovacs, Petra
    Toth, Brigitta
    Honbolygo, Ferenc
    Szalardy, Orsolya
    Kohari, Anna
    Mady, Katalin
    Magyari, Lilla
    Winkler, Istvan
    BRAIN RESEARCH, 2023, 1805
  • [6] Detection of attention in multi-talker scenarios: A fuzzy approach
    Minguillon, Jesus
    Angel Lopez-Gordo, M.
    Pelayo, Francisco
    EXPERT SYSTEMS WITH APPLICATIONS, 2016, 64 : 261 - 268
  • [7] Recognizing Multi-talker Speech with Permutation Invariant Training
    Yu, Dong
    Chang, Xuankai
    Qian, Yanmin
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 2456 - 2460
  • [8] Auditory masking of speech in reverberant multi-talker environments
    Weller, Tobias
    Buchholz, Joerg M.
    Best, Virginia
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2016, 139 (03): : 1303 - 1313
  • [9] Modeling speech localization, talker identification, and word recognition in a multi-talker setting
    Josupeit, Angela
    Hohmann, Volker
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2017, 142 (01): : 35 - 54
  • [10] Multi-Channel Speaker Verification for Single and Multi-talker Speech
    Kataria, Saurabh
    Zhang, Shi-Xiong
    Yu, Dong
    INTERSPEECH 2021, 2021, : 4608 - 4612