A Model of Neuronal Specialization Using Hebbian Policy-Gradient with "Slow" Noise

被引:0
|
作者
Dauce, Emmanuel [1 ]
机构
[1] INRIA Lille Nord Europe, Villeneuve Dascq, France
关键词
SPIKING NEURAL-NETWORKS; REINFORCEMENT;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study a model of neuronal specialization using a policy gradient reinforcement approach. (1) The neurons stochastically fire according to their synaptic input plus a noise term; (2) The environment is a closed-loop system composed of a rotating eye and a visual punctual target; (3) The network is composed of a foveated retina, a primary layer and a motoneuron layer; (4) The reward depends on the distance between the subjective target position and the fovea and (5) the weight update depends on a Hebbian trace defined according to a policy gradient principle. In order to take into account the mismatch between neuronal and environmental integration times, we distort the firing probability with a "pink noise" term whose autocorrelation is of the order of 100 ms, so that the firing probability is overestimated (or underestimated) for about 100 ms periods. The rewards occuring meanwhile assess the "value" of those elementary shifts, and modify the firing probability accordingly. Every motoneuron being associated to a particular angular direction, we test at the end of the learning process the preferred output of the visual cells. We find that accordingly with the observed final behavior, the visual cells preferentially excite the motoneurons heading in the opposite angular direction.
引用
收藏
页码:218 / 228
页数:11
相关论文
共 24 条
  • [1] A model of cell specialization using a Hebbian policy-gradient approach with "slow" noise
    Emmanuel Daucé
    [J]. BMC Neuroscience, 10 (Suppl 1)
  • [2] Active structural control framework using policy-gradient reinforcement learning
    Eshkevari, Soheila Sadeghi
    Eshkevari, Soheil Sadeghi
    Sen, Debarshi
    Pakzad, Shamim N.
    [J]. ENGINEERING STRUCTURES, 2022, 274
  • [3] Exploiting locality of interactions using a policy-gradient approach in multiagent learning
    Melo, Francisco S.
    [J]. ECAI 2008, PROCEEDINGS, 2008, 178 : 157 - +
  • [4] Traffic light control using deep policy-gradient and value-function-based reinforcement learning
    Mousavi, Seyed Sajad
    Schukat, Michael
    Howley, Enda
    [J]. IET INTELLIGENT TRANSPORT SYSTEMS, 2017, 11 (07) : 417 - 423
  • [5] Slow light using semiconductor optical amplifiers: Model and noise characteristics
    Berger, Perrine
    Alouini, Mehdi
    Bourderionnet, Jerome
    Bretenaker, Fabien
    Dolfi, Daniel
    [J]. COMPTES RENDUS PHYSIQUE, 2009, 10 (10) : 991 - 999
  • [6] Smart Noise Jamming Power Adjustment Using Exploratory Deep Deterministic Policy Gradient
    Zhang, Yujie
    Huo, Weibo
    Zhang, Cui
    Pei, Jifang
    Zhang, Yin
    Huang, Yulin
    [J]. 2023 IEEE RADAR CONFERENCE, RADARCONF23, 2023,
  • [7] Smart Noise Jamming Power Adjustment Using Exploratory Deep Deterministic Policy Gradient
    Zhang, Yujie
    Huo, Weibo
    Zhang, Cui
    Pei, Jifang
    Zhang, Yin
    Huang, Yulin
    [J]. Proceedings of the IEEE Radar Conference, 2023, 2023-May
  • [8] Semi-Supervised Speech Recognition Acoustic Model Training Using Policy Gradient
    Chung, Hoon
    Lee, Sung Joo
    Jeon, Hyeong Bae
    Park, Jeon Gue
    [J]. APPLIED SCIENCES-BASEL, 2020, 10 (10):
  • [9] Model predictive lighting control for a factory building using a deep deterministic policy gradient
    Kim, Young Sub
    Shin, Han Sol
    Park, Cheol Soo
    [J]. JOURNAL OF BUILDING PERFORMANCE SIMULATION, 2022, 15 (02) : 174 - 193
  • [10] MODEL FOR JET NOISE-ANALYSIS USING PRESSURE-GRADIENT CORRELATIONS ON AN IMAGINARY CONE
    NORUM, TD
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 1974, 56 : S49 - S49