DEEP REINFORCEMENT LEARNING WITH SPARSE DISTRIBUTED MEMORY FOR "WATER WORLD" PROBLEM SOLVING

被引:0
|
作者
Novotarskyi, M. A. [1 ]
Stirenko, S. G. [1 ]
Gordienko, Y. G. [1 ]
Kuzmych, V. A. [1 ]
机构
[1] Natl Tech Univ Ukraine, Igor Sikorsky Kyiv Polytech Inst, Dept Comp Engn, Kiev, Ukraine
关键词
Deep Reinforcement Learning; DQN-algorithm; Sparse Distributed Memory; Water World" problem;
D O I
10.15588/1607-3274-2021-1-14
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Context. Machine learning is one of the actively developing areas of data processing. Reinforcement learning is a class of machine learning methods where the problem involves mapping the sequence of environmental states to agent's actions. Significant progress in this area has been achieved using DQN-algorithms, which became one of the first classes of stable algorithms for learning using deep neural networks. The main disadvantage of this approach is the rapid growth of RAM in real-world tasks. The approach proposed in this paper can partially solve this problem. Objective. The aim is to develop a method of forming the structure and nature of access to the sparse distributed memory with increased information content to improve reinforcement learning without additional memory. Method. A method of forming the structure and modification of sparse distributed memory for storing previous transitions of the actor in the form of prototypes is proposed. The method allows increasing the informativeness of the stored data and, as a result, to improve the process of creating a model of the studied process by intensifying the learning of the deep neural network. Increasing the informativeness of the stored data is the result of this sequence of actions. First, we compare the new transition and the last saved transition. To perform this comparison, this method introduces a rate estimate for the distance between transitions. If the distance between the new transition and the last saved transition is smaller than the specified threshold, the new transition is written in place of the previous one without increasing the amount of memory. Otherwise, we create a new prototype in memory while deleting the prototype that has been stored in memory the longest. Results. The work of the proposed method was studied during the solution of the popular "Water World" test problem. The results showed a 1.5-times increase in the actor's survival time in a hostile environment. This result was achieved by increasing the informativeness of the stored data without increasing the amount of RAM. Conclusions. The proposed method of forming and modifying the structure of sparse distributed memory allowed to increase the informativeness of the stored data. As a result of this approach, improved reinforcement learning parameters on the example of the "Water World" problem by increasing the accuracy of the model of the physical process represented by a deep neural network.
引用
收藏
页码:136 / 143
页数:8
相关论文
共 50 条
  • [21] Heterogeneous Attentions for Solving Pickup and Delivery Problem via Deep Reinforcement Learning
    Li, Jingwen
    Xin, Liang
    Cao, Zhiguang
    Lim, Andrew
    Song, Wen
    Zhang, Jie
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (03) : 2306 - 2315
  • [22] Deep sparse representation via deep dictionary learning for reinforcement learning
    Tang, Jianhao
    Li, Zhenni
    Xie, Shengli
    Ding, Shuxue
    Zheng, Shaolong
    Chen, Xueni
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 2398 - 2403
  • [23] The State of Sparse Training in Deep Reinforcement Learning
    Graesser, Laura
    Evci, Utku
    Elsen, Erich
    Castro, Pablo Samuel
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [24] Deep reinforcement learning for dynamic distributed job shop scheduling problem with transfers
    Lei, Yong
    Deng, Qianwang
    Liao, Mengqi
    Gao, Shuocheng
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 251
  • [25] Reinforcement Learning for Solving the Vehicle Routing Problem
    Nazari, Mohammadreza
    Oroojlooy, Afshin
    Takac, Martin
    Snyder, Lawrence V.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [26] Hierarchical reinforcement learning as creative problem solving
    Colin, Thomas R.
    Belpaeme, Tony
    Cangelosi, Angelo
    Hemion, Nikolas
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2016, 86 : 196 - 206
  • [27] A reinforcement learning approach to cooperative problem solving
    Yoshida, T
    Hori, K
    Nakasuka, S
    INTERNATIONAL CONFERENCE ON MULTI-AGENT SYSTEMS, PROCEEDINGS, 1998, : 479 - 480
  • [28] Deep learning, reinforcement learning, and world models
    Matsuo, Yutaka
    LeCun, Yann
    Sahani, Maneesh
    Precup, Doina
    Silver, David
    Sugiyama, Masashi
    Uchibe, Eiji
    Morimoto, Jun
    NEURAL NETWORKS, 2022, 152 : 267 - 275
  • [29] Solving a Joint Pricing and Inventory Control Problem for Perishables via Deep Reinforcement Learning
    Wang, Rui
    Gan, Xianghua
    Li, Qing
    Yan, Xiao
    COMPLEXITY, 2021, 2021
  • [30] Solving the RNA design problem with reinforcement learning
    Eastman, Peter
    Shi, Jade
    Ramsundar, Bharath
    Pande, Vijay S.
    PLOS COMPUTATIONAL BIOLOGY, 2018, 14 (06)