Memory Efficient Experience Replay for Streaming Learning

被引:30
|
作者
Hayes, Tyler L. [1 ]
Cahill, Nathan D. [1 ]
Kanan, Christopher [1 ]
机构
[1] Rochester Inst Technol, Carlson Ctr Imaging Sci, Rochester, NY 14623 USA
关键词
WEIGHTED MAJORITY; NEURAL-NETWORK; ALGORITHM; ARTMAP; CLASSIFICATION;
D O I
10.1109/icra.2019.8793982
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In supervised machine learning, an agent is typically trained once and then deployed. While this works well for static settings, robots often operate in changing environments and must quickly learn new things from data streams. In this paradigm, known as streaming learning, a learner is trained online, in a single pass, from a data stream that cannot be assumed to be independent and identically distributed (iid). Streaming learning will cause conventional deep neural networks (DNNs) to fail for two reasons: 1) they need multiple passes through the entire dataset; and 2) non-iid data will cause catastrophic forgetting. An old fix to both of these issues is rehearsal. To learn a new example, rehearsal mixes it with previous examples, and then this mixture is used to update the DNN. Full rehearsal is slow and memory intensive because it stores all previously observed examples, and its effectiveness for preventing catastrophic forgetting has not been studied in modern DNNs. Here, we describe the ExStream algorithm for memory efficient rehearsal and compare it to alternatives. We find that full rehearsal can eliminate catastrophic forgetting in a variety of streaming learning settings, with ExStream performing well using far less memory and computation.
引用
收藏
页码:9769 / 9776
页数:8
相关论文
共 50 条
  • [1] Learning on Streaming Graphs with Experience Replay
    Perini, Massimo
    Ramponi, Giorgia
    Carbone, Paris
    Kalavri, Vasiliki
    [J]. 37TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, 2022, : 470 - 478
  • [2] Experience replay is associated with efficient nonlocal learning
    Liu, Yunzhe
    Mattar, Marcelo G.
    Behrens, Timothy E. J.
    Daw, Nathaniel D.
    Dolan, Raymond J.
    [J]. SCIENCE, 2021, 372 (6544) : 807 - +
  • [3] Efficient experience replay architecture for offline reinforcement learning
    Zhang, Longfei
    Feng, Yanghe
    Wang, Rongxiao
    Xu, Yue
    Xu, Naifu
    Liu, Zeyi
    Du, Hang
    [J]. ROBOTIC INTELLIGENCE AND AUTOMATION, 2023, 43 (01): : 35 - 43
  • [4] Associative Memory Based Experience Replay for Deep Reinforcement Learning
    Li, Mengyuan
    Kazemi, Arman
    Laguna, Ann Franchesca
    Hu, X. Sharon
    [J]. 2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [5] Map-based experience replay: a memory-efficient solution to catastrophic forgetting in reinforcement learning
    Hafez, Muhammad Burhan
    Immisch, Tilman
    Weber, Tom
    Wermter, Stefan
    [J]. FRONTIERS IN NEUROROBOTICS, 2023, 17
  • [6] A Dual Memory Structure for Efficient Use of Replay Memory in Deep Reinforcement Learning
    Ko, Wonshick
    Chang, Dong Eui
    [J]. 2019 19TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2019), 2019, : 1483 - 1486
  • [7] Memory Reduction through Experience Classification for Deep Reinforcement Learning with Prioritized Experience Replay
    Shen, Kai-Huan
    Tsai, Pei-Yun
    [J]. PROCEEDINGS OF THE 2019 IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS 2019), 2019, : 166 - 171
  • [8] Learning and memory - Replay that track
    Welberg, Leonie
    [J]. NATURE REVIEWS NEUROSCIENCE, 2008, 9 (10) : 739 - 739
  • [9] Streaming Linear System Identification with Reverse Experience Replay
    Jain, Prateek
    Kowshik, Suhas S.
    Nagaraj, Dheeraj
    Netrapalli, Praneeth
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [10] A divided and prioritized experience replay approach for streaming regression
    Arno, Mikkel Leite
    Godhavn, John-Morten
    Aamo, Ole Morten
    [J]. METHODSX, 2021, 8