Sample Efficient Reinforcement Learning Using Graph-Based Memory Reconstruction

被引:0
|
作者
Kang Y. [1 ,2 ]
Zhao E. [1 ,2 ]
Zang Y. [1 ,2 ]
Li L. [2 ]
Li K. [2 ]
Tao P. [3 ]
Xing J. [3 ]
机构
[1] School of Artificial Intelligence, University of Chinese, Academy of Sciences, Beijing
[2] Institute of Automation, Chinese Academy of Sciences, Beijing
[3] Department of Computer Science and Technology, Tsinghua University, Beijing
来源
基金
中国国家自然科学基金;
关键词
Experience replay (ER); graph model; memory reconstruction; reinforcement learning (RL); sample efficiency;
D O I
10.1109/TAI.2023.3268612
中图分类号
学科分类号
摘要
Reinforcement learning (RL) algorithms typically require orders of magnitude more interactions than humans to learn effective policies. Research on memory in neuroscience suggests that humans' learning efficiency benefits from associating their experiences and reconstructing potential events. Inspired by this finding, we introduce a human brainlike memory structure for agents and build a general learning framework based on this structure to improve the RL sampling efficiency. Since this framework is similar to the memory reconstruction process in psychology, we name the newly proposed RL framework as graph-based memory reconstruction (GBMR). In particular, GBMR first maintains an attribute graph on the agent's memory and then retrieves its critical nodes to build and update potential paths among these nodes. This novel pipeline drives the RL agent to learn faster with its memory-enhanced value functions and reduces interactions with the environment by reconstructing its valuable paths. Extensive experimental analyses and evaluations in the grid maze and some challenging Atari environments demonstrate GBMRs superiority over traditional RL methods. We will release the source code and trained models to facilitate further studies in this research direction. © 2023 IEEE.
引用
收藏
页码:751 / 762
页数:11
相关论文
共 50 条
  • [31] Readjoiner: a fast and memory efficient string graph-based sequence assembler
    Giorgio Gonnella
    Stefan Kurtz
    BMC Bioinformatics, 13
  • [32] Readjoiner: a fast and memory efficient string graph-based sequence assembler
    Gonnella, Giorgio
    Kurtz, Stefan
    BMC BIOINFORMATICS, 2012, 13
  • [33] Partially Occluded Face Reconstruction Using Graph-based Algorithm
    Meena, Manisha Kumari
    Meena, Hemant Kumar
    Sharma, Ramnivas
    JOURNAL OF ELECTRICAL ENGINEERING & TECHNOLOGY, 2024, 19 (06) : 3655 - 3664
  • [34] A GRAPH-BASED APPROACH TO SURFACE RECONSTRUCTION
    MENCL, R
    COMPUTER GRAPHICS FORUM, 1995, 14 (03) : C445 - C456
  • [35] Oracle-SAGE: Planning Ahead in Graph-Based Deep Reinforcement Learning
    Chester, Andrew
    Dann, Michael
    Zambetta, Fabio
    Thangarajah, John
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT IV, 2023, 13716 : 52 - 67
  • [36] A graph-based evolutionary algorithm: Genetic Network Programming (GNP) and its extension using reinforcement learning
    Mabu, Shingo
    Hirasawa, Kotaro
    Hu, Jinglu
    EVOLUTIONARY COMPUTATION, 2007, 15 (03) : 369 - 398
  • [37] Interpretable and Effective Reinforcement Learning for Attacking against Graph-based Rumor Detection
    Lyu, Yuefei
    Yang, Xiaoyu
    Liu, Jiaxin
    Xie, Sihong
    Yu, Philip
    Zhang, Xi
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [38] A Graph-Based Deep Reinforcement Learning Approach to Grasping Fully Occluded Objects
    Zuo, Guoyu
    Tong, Jiayuan
    Wang, Zihao
    Gong, Daoxiong
    COGNITIVE COMPUTATION, 2023, 15 (01) : 36 - 49
  • [39] Playing Text-Adventure Games with Graph-Based Deep Reinforcement Learning
    Ammanabrolu, Prithviraj
    Riedl, Mark O.
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 3557 - 3565
  • [40] Graph-based strategy evaluation for large-scale multiagent reinforcement learning
    Sun, Yiyun
    Liu, Meiqin
    Zhang, Senlin
    Zheng, Ronghao
    Dong, Shanling
    SCIENCE CHINA-INFORMATION SCIENCES, 2025, 68 (08)