Temporal encoding in deep reinforcement learning agents

被引:0
|
作者
Dongyan Lin
Ann Zixiang Huang
Blake Aaron Richards
机构
[1] McGill University,Integrated Program in Neuroscience
[2] Mila,School of Computer Science
[3] McGill University,Department of Neurology and Neurosurgery, Montreal Neurological Institute
[4] McGill University,Learning in Machines and Brains Program
[5] Canadian Institute for Advanced Research,undefined
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Neuroscientists have observed both cells in the brain that fire at specific points in time, known as “time cells”, and cells whose activity steadily increases or decreases over time, known as “ramping cells”. It is speculated that time and ramping cells support temporal computations in the brain and carry mnemonic information. However, due to the limitations in animal experiments, it is difficult to determine how these cells really contribute to behavior. Here, we show that time cells and ramping cells naturally emerge in the recurrent neural networks of deep reinforcement learning models performing simulated interval timing and working memory tasks, which have learned to estimate expected rewards in the future. We show that these cells do indeed carry information about time and items stored in working memory, but they contribute to behavior in large part by providing a dynamic representation on which policy can be computed. Moreover, the information that they do carry depends on both the task demands and the variables provided to the models. Our results suggest that time cells and ramping cells could contribute to temporal and mnemonic calculations, but the way in which they do so may be complex and unintuitive to human observers.
引用
收藏
相关论文
共 50 条
  • [41] Agents and reinforcement learning
    Harlequin's Adaptive Systems Group
    Dr Dobb's J Software Tools Prof Program, 3 (3pp):
  • [42] Reinforcement learning agents
    Ribeiro, C
    ARTIFICIAL INTELLIGENCE REVIEW, 2002, 17 (03) : 223 - 250
  • [43] Agents and reinforcement learning
    Singh, S
    Norvig, P
    Cohn, D
    DR DOBBS JOURNAL, 1997, 22 (03): : 28 - +
  • [44] Temporally transferable crop mapping with temporal encoding and deep learning augmentations
    Pham, Vu-Dong
    Tetteh, Gideon
    Thiel, Fabian
    Erasmi, Stefan
    Schwieder, Marcel
    Frantz, David
    van der Linden, Sebastian
    INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2024, 129
  • [45] The Advance of Reinforcement Learning and Deep Reinforcement Learning
    Lyu, Le
    Shen, Yang
    Zhang, Sicheng
    2022 IEEE INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, BIG DATA AND ALGORITHMS (EEBDA), 2022, : 644 - 648
  • [46] A Search-Based Testing Approach for Deep Reinforcement Learning Agents
    Zolfagharian, Amirhossein
    Abdellatif, Manel
    Briand, Lionel C.
    Bagherzadeh, Mojtaba
    Ramesh, S.
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2023, 49 (07) : 3715 - 3735
  • [47] Deep Reinforcement Learning Approach for Flocking Control of Multi-agents
    Zhang, Han
    Cheng, Jin
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 5002 - 5007
  • [48] Scaling up Deep Reinforcement Learning for Intelligent Video Game Agents
    Debner, Anton
    2022 IEEE INTERNATIONAL CONFERENCE ON SMART COMPUTING (SMARTCOMP 2022), 2022, : 192 - 193
  • [49] Uncertainty-Aware Action Advising for Deep Reinforcement Learning Agents
    Da Silva, Felipe Lena
    Hernandez-Leal, Pablo
    Kartal, Bilal
    Taylor, Matthew E.
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 5792 - 5799
  • [50] Spatiotemporally Constrained Action Space Attacks on Deep Reinforcement Learning Agents
    Lee, Xian Yeow
    Ghadai, Sambit
    Tan, Kai Liang
    Hegde, Chinmay
    Sarkar, Soumik
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 4577 - 4584