Convolutional neural network-based encoding and decoding of visual object recognition in space and time

被引:57
|
作者
Seeliger, K. [1 ]
Fritsche, M. [1 ]
Guclu, U. [1 ]
Schoenmakers, S. [1 ]
Schoffelen, J. -M. [1 ]
Bosch, S. E. [1 ]
van Gerven, M. A. J. [1 ]
机构
[1] Radboud Univ Nijmegen, Donders Inst Brain Cognit & Behav, Montessorilaan 3, NL-6525 HR Nijmegen, Netherlands
关键词
Visual neuroscience; Deep learning; Encoding; Decoding; Magnetoencephalography; BRAIN; MODELS; IMAGES;
D O I
10.1016/j.neuroimage.2017.07.018
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Representations learned by deep convolutional neural networks (CNNs) for object recognition are a widely investigated model of the processing hierarchy in the human visual system. Using functional magnetic resonance imaging, CNN representations of visual stimuli have previously been shown to correspond to processing stages in the ventral and dorsal streams of the visual system. Whether this correspondence between models and brain signals also holds for activity acquired a high temporal resolution has been explored less exhaustively. Here, we addressed this question by combining CNN-based encoding models with magnetoencephalography (MEG). Human participants passively viewed 1,000 images of objects while MEG signals were acquired. We modelled their high temporal resolution source-reconstructed cortical activity with CNNs, and observed a feed-forward sweep across the visual hierarchy between 75 and 200 ms after stimulus onset. This spatiotemporal cascade was captured by the network layer representations, where the increasingly abstract stimulus representation in the hierarchical network model was reflected in different parts of the visual cortex, following the visual ventral stream. We further validated the accuracy of our encoding model by decoding stimulus identity in a left-out validation set of viewed objects, achieving state-of-the-art decoding accuracy.
引用
收藏
页码:253 / 266
页数:14
相关论文
共 50 条
  • [1] A convolutional neural network for visual object recognition in marine sector
    Kumar, Aiswarya S.
    Sherly, Elizabeth
    [J]. 2017 2ND INTERNATIONAL CONFERENCE FOR CONVERGENCE IN TECHNOLOGY (I2CT), 2017, : 304 - 307
  • [2] Time-Frequency Representation and Convolutional Neural Network-Based Emotion Recognition
    Khare, Smith K.
    Bajaj, Varun
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (07) : 2901 - 2909
  • [3] Radar Based Object Recognition with Convolutional Neural Network
    Loi, Kin Chong
    Cheong, Pedro
    Choi, Wai Wa
    [J]. PROCEEDINGS OF THE 2019 IEEE ASIA-PACIFIC MICROWAVE CONFERENCE (APMC), 2019, : 87 - 89
  • [4] Underwater Object Recognition Based on Deep Encoding-Decoding Network
    WANG Xinhua
    OUYANG Jihong
    LI Dayu
    ZHANG Guang
    [J]. Journal of Ocean University of China, 2019, 18 (02) : 376 - 382
  • [5] Underwater Object Recognition Based on Deep Encoding-Decoding Network
    Xinhua Wang
    Jihong Ouyang
    Dayu Li
    Guang Zhang
    [J]. Journal of Ocean University of China, 2019, 18 : 376 - 382
  • [6] Underwater Object Recognition Based on Deep Encoding-Decoding Network
    Wang, Xinhua
    Ouyang, Jihong
    Li, Dayu
    Zhang, Guang
    [J]. JOURNAL OF OCEAN UNIVERSITY OF CHINA, 2019, 18 (02) : 376 - 382
  • [7] Simultaneous Space Object Recognition and Pose Estimation by Convolutional Neural Network
    Afshar, Roya
    Chu, Zhongyi
    Lu, Shuai
    [J]. PROCEEDINGS OF THE 2020 INTERNATIONAL CONFERENCE ON ARTIFICIAL LIFE AND ROBOTICS (ICAROB2020), 2020, : 490 - 495
  • [8] Deep Convolutional Neural Network-Based Approaches for Face Recognition
    Almabdy, Soad
    Elrefaei, Lamiaa
    [J]. APPLIED SCIENCES-BASEL, 2019, 9 (20):
  • [9] Residual Convolutional Neural Network-Based Dysarthric Speech Recognition
    Kumar, Raj
    Tripathy, Manoj
    Anand, R. S.
    Kumar, Niraj
    [J]. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2024,
  • [10] Convolutional Neural Network-Based Periocular Recognition in Surveillance Environments
    Kim, Min Cheol
    Koo, Ja Hyung
    Cho, Se Woon
    Baek, Na Rae
    Park, Kang Ryoung
    [J]. IEEE ACCESS, 2018, 6 : 57291 - 57310