DEEP REINFORCEMENT LEARNING WITH SPARSE DISTRIBUTED MEMORY FOR "WATER WORLD" PROBLEM SOLVING

被引:0
|
作者
Novotarskyi, M. A. [1 ]
Stirenko, S. G. [1 ]
Gordienko, Y. G. [1 ]
Kuzmych, V. A. [1 ]
机构
[1] Natl Tech Univ Ukraine, Igor Sikorsky Kyiv Polytech Inst, Dept Comp Engn, Kiev, Ukraine
关键词
Deep Reinforcement Learning; DQN-algorithm; Sparse Distributed Memory; Water World" problem;
D O I
10.15588/1607-3274-2021-1-14
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Context. Machine learning is one of the actively developing areas of data processing. Reinforcement learning is a class of machine learning methods where the problem involves mapping the sequence of environmental states to agent's actions. Significant progress in this area has been achieved using DQN-algorithms, which became one of the first classes of stable algorithms for learning using deep neural networks. The main disadvantage of this approach is the rapid growth of RAM in real-world tasks. The approach proposed in this paper can partially solve this problem. Objective. The aim is to develop a method of forming the structure and nature of access to the sparse distributed memory with increased information content to improve reinforcement learning without additional memory. Method. A method of forming the structure and modification of sparse distributed memory for storing previous transitions of the actor in the form of prototypes is proposed. The method allows increasing the informativeness of the stored data and, as a result, to improve the process of creating a model of the studied process by intensifying the learning of the deep neural network. Increasing the informativeness of the stored data is the result of this sequence of actions. First, we compare the new transition and the last saved transition. To perform this comparison, this method introduces a rate estimate for the distance between transitions. If the distance between the new transition and the last saved transition is smaller than the specified threshold, the new transition is written in place of the previous one without increasing the amount of memory. Otherwise, we create a new prototype in memory while deleting the prototype that has been stored in memory the longest. Results. The work of the proposed method was studied during the solution of the popular "Water World" test problem. The results showed a 1.5-times increase in the actor's survival time in a hostile environment. This result was achieved by increasing the informativeness of the stored data without increasing the amount of RAM. Conclusions. The proposed method of forming and modifying the structure of sparse distributed memory allowed to increase the informativeness of the stored data. As a result of this approach, improved reinforcement learning parameters on the example of the "Water World" problem by increasing the accuracy of the model of the physical process represented by a deep neural network.
引用
收藏
页码:136 / 143
页数:8
相关论文
共 50 条
  • [31] Solving the Deadlock Problem with Deep Reinforcement Learning Using Information from Multiple Vehicles
    Goto, Tsuyoshi
    Itaya, Hidenori
    Hirakawa, Tsubasa
    Yamashita, Takayoshi
    Fujiyoshi, Hironobu
    2022 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2022, : 1026 - 1032
  • [32] Deep reinforcement learning for solving the joint scheduling problem of machines and AGVs in job shop
    Sun A.-H.
    Lei Q.
    Song Y.-C.
    Yang Y.-F.
    Lei, Qi (leiqi@cqu.edu.cn), 1600, Northeast University (39): : 253 - 262
  • [33] Solving the Vehicle Routing Problem with Stochastic Travel Cost Using Deep Reinforcement Learning
    Cai, Hao
    Xu, Peng
    Tang, Xifeng
    Lin, Gan
    ELECTRONICS, 2024, 13 (16)
  • [34] A Distributed-GPU Deep Reinforcement Learning System for Solving Large Graph Optimization Problems
    Zheng, Weijian
    Wang, Dali
    Song, Fengguang
    ACM TRANSACTIONS ON PARALLEL COMPUTING, 2023, 10 (02)
  • [35] Bayesian Reinforcement Learning via Deep, Sparse Sampling
    Grover, Divya
    Basu, Debabrota
    Dimitrakakis, Christos
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 3036 - 3044
  • [36] Sparse Distributed Memory Approach for Reinforcement Learning Driven Efficient Routing in Mobile Wireless Network System
    Vidyadhar, Varshini
    Nagaraj, R.
    Sudha, G.
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2021, 12 (11) : 144 - 152
  • [37] A Distributed Framework for Deep Reinforcement Learning by Consensus
    Liu, Bo
    Zhu, Shuang
    Sun, Peng
    Huang, Qisheng
    Ding, Zhenglao
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 269 - 274
  • [38] Distributed Deep Reinforcement Learning using TensorFlow
    Rao, P. Ajay
    Kumar, Navaneesh B.
    Cadabam, Siddharth
    Praveena, T.
    2017 INTERNATIONAL CONFERENCE ON CURRENT TRENDS IN COMPUTER, ELECTRICAL, ELECTRONICS AND COMMUNICATION (CTCEEC), 2017, : 171 - 174
  • [39] Distributed deep reinforcement learning for simulation control
    Pawar, Suraj
    Maulik, Romit
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2021, 2 (02):
  • [40] Metaoptimization on a Distributed System for Deep Reinforcement Learning
    Heinrich, Greg
    Frosio, Iuri
    PROCEEDINGS OF 2019 5TH IEEE/ACM WORKSHOP ON MACHINE LEARNING IN HIGH PERFORMANCE COMPUTING ENVIRONMENTS (MLHPC 2019), 2019, : 19 - 30