Learning offline: memory replay in biological and artificial reinforcement learning

被引:17
|
作者
Roscow, Emma L. [1 ]
Chua, Raymond [2 ]
Costa, Rui Ponte [3 ]
Jones, Matt W. [4 ]
Lepora, Nathan [5 ,6 ]
机构
[1] Ctr Recerca Matemat, Bellaterra, Spain
[2] McGill Univ & Mila, Montreal, PQ, Canada
[3] Univ Bristol, Dept Comp Sci, Intelligent Syst Lab, Bristol Computat Neurosci Unit, Bristol, Avon, England
[4] Univ Bristol, Sch Physiol Pharmacol & Neurosci, Bristol, Avon, England
[5] Univ Bristol, Dept Engn Math, Bristol, Avon, England
[6] Univ Bristol, Bristol Robot Lab, Bristol, Avon, England
基金
英国惠康基金; 加拿大自然科学与工程研究理事会;
关键词
SHARP-WAVE RIPPLES; HIPPOCAMPAL REPLAY; PREFRONTAL CORTEX; VALUE REPRESENTATIONS; NEURAL-NETWORKS; SPATIAL MEMORY; VISUAL-CORTEX; AWAKE REPLAY; REACTIVATION; SLEEP;
D O I
10.1016/j.tins.2021.07.007
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Learning to act in an environment to maximise rewards is among the brain's key functions. This process has often been conceptualised within the framework of reinforcement learning, which has also gained prominence in machine learning and artificial intelligence (AI) as a way to optimise decision making. A common aspect of both biological and machine reinforcement learning is the reactivation of previously experienced episodes, referred to as replay. Replay is important for memory consolidation in biological neural networks and is key to stabilising learning in deep neural networks. Here, we review recent developments concerning the functional roles of replay in the fields of neuroscience and AI. Complementary progress suggests how replay might support learning processes, including generalisation and continual learning, affording opportunities to transfer knowledge across the two fields to advance the understanding of biological and artificial learning and memory.
引用
下载
收藏
页码:808 / 821
页数:14
相关论文
共 50 条
  • [21] Measuring the Impact of Memory Replay in Training Pacman Agents using Reinforcement Learning
    Fallas-Moya, Fabian
    Duncan, Jeremiah
    Samuel, Tabitha
    Sadovnik, Amir
    2021 XLVII LATIN AMERICAN COMPUTING CONFERENCE (CLEI 2021), 2021,
  • [22] Learning to Influence Human Behavior with Offline Reinforcement Learning
    Hong, Joey
    Levine, Sergey
    Dragan, Anca
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [23] A Review of Offline Reinforcement Learning Based on Representation Learning
    Wang X.-S.
    Wang R.-R.
    Cheng Y.-H.
    Zidonghua Xuebao/Acta Automatica Sinica, 2024, 50 (06): : 1104 - 1128
  • [24] Bootstrapped Transformer for Offline Reinforcement Learning
    Wang, Kerong
    Zhao, Hanye
    Luo, Xufang
    Ren, Kan
    Zhang, Weinan
    Li, Dongsheng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [25] Offline Reinforcement Learning with Differential Privacy
    Qiao, Dan
    Wang, Yu-Xiang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [26] Conservative Offline Distributional Reinforcement Learning
    Ma, Yecheng Jason
    Jayaraman, Dinesh
    Bastani, Osbert
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [27] Survival Instinct in Offline Reinforcement Learning
    Li, Anqi
    Misra, Dipendra
    Kolobov, Andrey
    Cheng, Ching-An
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [28] Offline reinforcement learning with task hierarchies
    Schwab, Devin
    Ray, Soumya
    MACHINE LEARNING, 2017, 106 (9-10) : 1569 - 1598
  • [29] Conservative network for offline reinforcement learning
    Peng, Zhiyong
    Liu, Yadong
    Chen, Haoqiang
    Zhou, Zongtan
    KNOWLEDGE-BASED SYSTEMS, 2023, 282
  • [30] Conservative Offline Distributional Reinforcement Learning
    Ma, Yecheng Jason
    Jayaraman, Dinesh
    Bastani, Osbert
    Advances in Neural Information Processing Systems, 2021, 23 : 19235 - 19247