Temporal Coherence and the Streaming of Complex Sounds

被引:25
|
作者
Shamma, Shihab [1 ]
Elhilali, Mounya [3 ]
Ma, Ling [1 ,2 ]
Micheyl, Christophe [4 ]
Oxenham, Andrew J. [4 ]
Pressnitzer, Daniel [5 ,6 ]
Yin, Pingbo [1 ]
Xu, Yanbo [1 ]
机构
[1] Univ Maryland, Syst Res Inst, Dept Elect & Comp Engn, College Pk, MD 20742 USA
[2] Univ Maryland, Bioengn Program, College Pk, MD 20742 USA
[3] Johns Hopkins Univ, Dept Elect & Comp Engn, Baltimore, MD 21218 USA
[4] Univ Minnesota, Dept Psychol, Minneapolis, MN 55455 USA
[5] Ecole Normale Super, Equipe Audit, Dept Etudes Cognit, F-75231 Paris, France
[6] Univ Paris 05, UMR CNRS 8158, Lab Psychol Percept, Paris, France
关键词
ATTENTION;
D O I
10.1007/978-1-4614-1590-9_59
中图分类号
R-3 [医学研究方法]; R3 [基础医学];
学科分类号
1001 ;
摘要
Humans and other animals can attend to one of multiple sounds, and follow it selectively over time. The neural underpinnings of this perceptual feat remain mysterious. Some studies have concluded that sounds are heard as separate streams when they activate well-separated populations of central auditory neurons, and that this process is largely pre-attentive. Here, we propose instead that stream formation depends primarily on temporal coherence between responses that encode various features of a sound source. Furthermore, we postulate that only when attention is directed toward a particular feature (e.g., pitch or location) do all other temporally coherent features of that source (e.g., timbre and location) become bound together as a stream that is segregated from the incoherent features of other sources. Experimental neurophysiological evidence in support of this hypothesis will be presented. The focus, however, will be on a computational realization of this idea and a discussion of the insights learned from simulations to disentangle complex sound sources such as speech and music. The model consists of a representational stage of early and cortical auditory processing that creates a multidimensional depiction of various sound attributes such as pitch, location, and spectral resolution. The following stage computes a coherence matrix that summarizes the pair-wise correlations between all channels making up the cortical representation. Finally, the perceived segregated streams are extracted by decomposing the coherence matrix into its uncorrelated components. Questions raised by the model are discussed, especially on the role of attention in streaming and the search for further neural correlates of streaming percepts.
引用
收藏
页码:535 / 543
页数:9
相关论文
共 50 条
  • [41] Analysis and classification of respiratory sounds by signal coherence method
    Baydar, KS
    Ertuzun, A
    Kahya, YP
    PROCEEDINGS OF THE 25TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY, VOLS 1-4: A NEW BEGINNING FOR HUMAN HEALTH, 2003, 25 : 2950 - 2953
  • [42] Temporal Biased Streaming Submodular Optimization
    Zhao, Junzhou
    Wang, Pinghui
    Deng, Chao
    Tao, Jing
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 2305 - 2315
  • [43] Temporal Sentence Grounding in Streaming Videos
    Gan, Tian
    Wang, Xiao
    Sun, Yan
    Wu, Jianlong
    Guo, Qingpei
    Nie, Liqiang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 4637 - 4646
  • [44] Streaming Temporal Graphs: Subgraph Matching
    Goodman, Eric L.
    Grunwald, Dirk
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 4977 - 4986
  • [45] Temporal Random Testing for Spark Streaming
    Riesco, Adrian
    Rodriguez-Hortala, Juan
    INTEGRATED FORMAL METHODS (IFM 2016), 2016, 9681 : 393 - 408
  • [46] Spatio-Temporal Memory Streaming
    Somogyi, Stephen
    Wenisch, Thomas F.
    Ailamaki, Anastasia
    Falsafi, Babak
    ISCA 2009: 36TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE, 2009, : 69 - 80
  • [47] The Encoding of Speech Sounds in the Superior Temporal Gyrus
    Yi, Han Gyol
    Leonard, Matthew K.
    Chang, Edward F.
    NEURON, 2019, 102 (06) : 1096 - 1110
  • [48] Discrimination and streaming of speech sounds based on differences in interaural and spectral cues
    David, Marion
    Lavandier, Mathieu
    Grimault, Nicolas
    Oxenham, Andrew J.
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2017, 142 (03): : 1674 - 1685
  • [49] SEPARABILITY OF TEMPORAL COHERENCE AND SPATIAL COHERENCE IN INTERFEROMETRIC EXPERIMENTS
    MARECHAL, A
    CHAVEL, P
    JOURNAL OF OPTICS-NOUVELLE REVUE D OPTIQUE, 1982, 13 (04): : 193 - 197
  • [50] Coding the temporal structure of sounds in auditory cortex
    Moore, DR
    Schnupp, JWH
    King, AJ
    NATURE NEUROSCIENCE, 2001, 4 (11) : 1055 - 1056