DEEP REINFORCEMENT LEARNING WITH SPARSE DISTRIBUTED MEMORY FOR "WATER WORLD" PROBLEM SOLVING

被引:0
|
作者
Novotarskyi, M. A. [1 ]
Stirenko, S. G. [1 ]
Gordienko, Y. G. [1 ]
Kuzmych, V. A. [1 ]
机构
[1] Natl Tech Univ Ukraine, Igor Sikorsky Kyiv Polytech Inst, Dept Comp Engn, Kiev, Ukraine
关键词
Deep Reinforcement Learning; DQN-algorithm; Sparse Distributed Memory; Water World" problem;
D O I
10.15588/1607-3274-2021-1-14
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Context. Machine learning is one of the actively developing areas of data processing. Reinforcement learning is a class of machine learning methods where the problem involves mapping the sequence of environmental states to agent's actions. Significant progress in this area has been achieved using DQN-algorithms, which became one of the first classes of stable algorithms for learning using deep neural networks. The main disadvantage of this approach is the rapid growth of RAM in real-world tasks. The approach proposed in this paper can partially solve this problem. Objective. The aim is to develop a method of forming the structure and nature of access to the sparse distributed memory with increased information content to improve reinforcement learning without additional memory. Method. A method of forming the structure and modification of sparse distributed memory for storing previous transitions of the actor in the form of prototypes is proposed. The method allows increasing the informativeness of the stored data and, as a result, to improve the process of creating a model of the studied process by intensifying the learning of the deep neural network. Increasing the informativeness of the stored data is the result of this sequence of actions. First, we compare the new transition and the last saved transition. To perform this comparison, this method introduces a rate estimate for the distance between transitions. If the distance between the new transition and the last saved transition is smaller than the specified threshold, the new transition is written in place of the previous one without increasing the amount of memory. Otherwise, we create a new prototype in memory while deleting the prototype that has been stored in memory the longest. Results. The work of the proposed method was studied during the solution of the popular "Water World" test problem. The results showed a 1.5-times increase in the actor's survival time in a hostile environment. This result was achieved by increasing the informativeness of the stored data without increasing the amount of RAM. Conclusions. The proposed method of forming and modifying the structure of sparse distributed memory allowed to increase the informativeness of the stored data. As a result of this approach, improved reinforcement learning parameters on the example of the "Water World" problem by increasing the accuracy of the model of the physical process represented by a deep neural network.
引用
收藏
页码:136 / 143
页数:8
相关论文
共 50 条
  • [1] Deep Reinforcement Learning for Solving Distributed Permutation Flow Shop Scheduling Problem
    Wang, Yijun
    Qian, Bin
    Hu, Rong
    Yang, Yuanyuan
    Chen, Wenbo
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, ICIC 2023, PT I, 2023, 14086 : 333 - 342
  • [2] Solving the k-Sparse Eigenvalue Problem with Reinforcement Learning
    Zhou, Li
    Yan, Lihao
    Caprio, Mark A.
    Gao, Weiguo
    Yang, Chao
    CSIAM TRANSACTIONS ON APPLIED MATHEMATICS, 2021, 2 (04): : 697 - 723
  • [3] Solving Permutation Flowshop Problem with Deep Reinforcement Learning
    Pan, Ruyuan
    Dong, Xingye
    Han, Sheng
    2020 PROGNOSTICS AND SYSTEM HEALTH MANAGEMENT CONFERENCE (PHM-BESANCON 2020), 2020, : 349 - 353
  • [4] Deep Reinforcement Learning for Solving AGVs Routing Problem
    Lu, Chengxuan
    Long, Jinjun
    Xing, Zichao
    Wu, Weimin
    Gu, Yong
    Luo, Jiliang
    Huang, Yisheng
    VERIFICATION AND EVALUATION OF COMPUTER AND COMMUNICATION SYSTEMS, VECOS 2020, 2020, 12519 : 222 - 236
  • [5] Deep Graph Reinforcement Learning for Solving Multicut Problem
    Li, Zhenchen
    Yang, Xu
    Zhang, Yanchao
    Zeng, Shaofeng
    Yuan, Jingbin
    Liu, Jiazheng
    Liu, Zhiyong
    Han, Hua
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [6] Deep reinforcement learning for solving the single container loading problem
    Hajlaoui, Yakin
    Jaoua, Amel
    Layeb, Safa Bhar
    ENGINEERING OPTIMIZATION, 2023, 55 (04) : 668 - 684
  • [7] Solving the train dispatching problem via deep reinforcement learning
    Agasucci, Valerio
    Grani, Giorgio
    Lamorgese, Leonardo
    JOURNAL OF RAIL TRANSPORT PLANNING & MANAGEMENT, 2023, 26
  • [8] Solving the online batching problem using deep reinforcement learning
    Cals, Bram
    Zhang, Yingqian
    Dijkman, Remco
    van Dorst, Claudy
    COMPUTERS & INDUSTRIAL ENGINEERING, 2021, 156
  • [9] Deep reinforcement learning algorithm for solving material emergency dispatching problem
    Jiang, Huawei
    Guo, Tao
    Yang, Zhen
    Zhao, Like
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2022, 19 (11) : 10864 - 10881
  • [10] Solving the Traveling Salesperson Problem with Precedence Constraints by Deep Reinforcement Learning
    Loewens, Christian
    Ashraf, Inaam
    Gembus, Alexander
    Cuizon, Genesis
    Falkner, Jonas K.
    Schmidt-Thieme, Lars
    ADVANCES IN ARTIFICIAL INTELLIGENCE, KI 2022, 2022, 13404 : 160 - 172