A Spiking Neural Network Model of Model-Free Reinforcement Learning with High-Dimensional Sensory Input and Perceptual Ambiguity

被引:13
|
作者
Nakano, Takashi [1 ]
Otsuka, Makoto [2 ]
Yoshimoto, Junichiro [2 ]
Doya, Kenji [2 ]
机构
[1] Okinawa Inst Sci & Technol, Neurobiol Res Unit, Kunigami, Okinawa 9040495, Japan
[2] Okinawa Inst Sci & Technol, Neural Computat Unit, Kunigami, Okinawa 9040495, Japan
来源
PLOS ONE | 2015年 / 10卷 / 03期
关键词
REPRESENTATION; CATEGORIES;
D O I
10.1371/journal.pone.0115620
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
A theoretical framework of reinforcement learning plays an important role in understanding action selection in animals. Spiking neural networks provide a theoretically grounded means to test computational hypotheses on neurally plausible algorithms of reinforcement learning through numerical simulation. However, most of these models cannot handle observations which are noisy, or occurred in the past, even though these are inevitable and constraining features of learning in real environments. This class of problem is formally known as partially observable reinforcement learning (PORL) problems. It provides a generalization of reinforcement learning to partially observable domains. In addition, observations in the real world tend to be rich and high-dimensional. In this work, we use a spiking neural network model to approximate the free energy of a restricted Boltzmann machine and apply it to the solution of PORL problems with high-dimensional observations. Our spiking network model solves maze tasks with perceptually ambiguous high-dimensional observations without knowledge of the true environment. An extended model with working memory also solves history-dependent tasks. The way spiking neural networks handle PORL problems may provide a glimpse into the underlying laws of neural information processing which can only be discovered through such a top-down approach.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Spiking neural network model of free-energy-based reinforcement learning
    Takashi Nakano
    Makoto Otsuka
    BMC Neuroscience, 12 (Suppl 1)
  • [2] Synaptic plasticity model of a spiking neural network for reinforcement learning
    Lee, Kyoobin
    Kwon, Dong-Soo
    NEUROCOMPUTING, 2008, 71 (13-15) : 3037 - 3043
  • [3] Model-Free Statistical Inference on High-Dimensional Data
    Guo, Xu
    Li, Runze
    Zhang, Zhe
    Zou, Changliang
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2024,
  • [4] A spiking neural network model of primary visual cortex for perceptual learning
    Naito, Satoshi
    Yukinawa, Naoto
    Ishii, Shin
    NEUROSCIENCE RESEARCH, 2010, 68 : E210 - E211
  • [5] A model-free, reinforcement learning algorithm for perceptual decision making under uncertainty
    Esmaily, Jamal
    Moran, Rani
    Roudi, Yasser
    Bahrami, Bahador
    JOURNAL OF COMPUTATIONAL NEUROSCIENCE, 2024, 52 : S21 - S21
  • [6] Model-free feature screening for high-dimensional survival data
    Lin, Yuanyuan
    Liu, Xianhui
    Hao, Meiling
    SCIENCE CHINA-MATHEMATICS, 2018, 61 (09) : 1617 - 1636
  • [7] A model-free, reinforcement learning algorithm for perceptual decision making under uncertainty
    Esmaily, Jamal
    Moran, Rani
    Roudi, Yasser
    Bahrami, Bahador
    JOURNAL OF COMPUTATIONAL NEUROSCIENCE, 2024, 52 : S21 - S21
  • [8] Model-free feature screening for high-dimensional survival data
    Yuanyuan Lin
    Xianhui Liu
    Meiling Hao
    Science China Mathematics, 2018, 61 : 1617 - 1636
  • [9] Model-free feature screening for high-dimensional survival data
    Yuanyuan Lin
    Xianhui Liu
    Meiling Hao
    Science China(Mathematics), 2018, 61 (09) : 79 - 98
  • [10] Spiking neural network model of reinforcement learning in the honeybee implemented on the GPU
    Esin Yavuz
    Pascale Maul
    Thomas Nowotny
    BMC Neuroscience, 16 (Suppl 1)