Deep reinforcement learning with significant multiplications inference

被引:0
|
作者
Ivanov, Dmitry A. [1 ,2 ]
Larionov, Denis A. [2 ,3 ]
Kiselev, Mikhail V. [2 ,3 ]
Dylov, Dmitry V. [4 ,5 ]
机构
[1] Lomonosov Moscow State Univ, GSP 1,Leninskie Gory, Moscow 119991, Russia
[2] Cifrum, 3 Kholodilnyy per, Moscow 115191, Russia
[3] Chuvash State Univ, 15 Moskovsky pr, Cheboksary 428015, Chuvash, Russia
[4] Skolkovo Inst Sci & Technol, 30 1 Bolshoi blvd, Moscow 121205, Russia
[5] Artificial Intelligence Res Inst, 32 1 Kutuzovsky pr, Moscow 121170, Russia
基金
俄罗斯基础研究基金会;
关键词
D O I
10.1038/s41598-023-47245-y
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
We propose a sparse computation method for optimizing the inference of neural networks in reinforcement learning (RL) tasks. Motivated by the processing abilities of the brain, this method combines simple neural network pruning with a delta-network algorithm to account for the input data correlations. The former mimics neuroplasticity by eliminating inefficient connections; the latter makes it possible to update neuron states only when their changes exceed a certain threshold. This combination significantly reduces the number of multiplications during the neural network inference for fast neuromorphic computing. We tested the approach in popular deep RL tasks, yielding up to a 100-fold reduction in the number of required multiplications without substantial performance loss (sometimes, the performance even improved).
引用
收藏
页数:10
相关论文
共 50 条
  • [41] VIREL: A Variational Inference Framework for Reinforcement Learning
    Fellows, Matthew
    Mahajan, Anuj
    Rudner, Tim G. J.
    Whiteson, Shimon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [42] Probabilistic Inference in Reinforcement Learning Done Right
    Tarbouriech, Jean
    Lattimore, Tor
    O'Donoghue, Brendan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [43] Probabilistic inference for determining options in reinforcement learning
    Christian Daniel
    Herke van Hoof
    Jan Peters
    Gerhard Neumann
    Machine Learning, 2016, 104 : 337 - 357
  • [44] Fuzzy inference system learning by reinforcement methods
    Jouffe, L
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 1998, 28 (03): : 338 - 355
  • [45] Online reinforcement learning control by Bayesian inference
    Xia, Zhongpu
    Zhao, Dongbin
    IET CONTROL THEORY AND APPLICATIONS, 2016, 10 (12): : 1331 - 1338
  • [46] Explainability in deep reinforcement learning
    Heuillet, Alexandre
    Couthouis, Fabien
    Diaz-Rodriguez, Natalia
    KNOWLEDGE-BASED SYSTEMS, 2021, 214 (214)
  • [47] Deep Reinforcement Learning with Adjustments
    Khorasgani, Hamed
    Wang, Haiyan
    Gupta, Chetan
    Serita, Susumu
    2021 IEEE 19TH INTERNATIONAL CONFERENCE ON INDUSTRIAL INFORMATICS (INDIN), 2021,
  • [48] Deep Reinforcement Learning: An Overview
    Mousavi, Seyed Sajad
    Schukat, Michael
    Howley, Enda
    PROCEEDINGS OF SAI INTELLIGENT SYSTEMS CONFERENCE (INTELLISYS) 2016, VOL 2, 2018, 16 : 426 - 440
  • [49] Implementation of Deep Reinforcement Learning
    Li, Meng-Jhe
    Li, An-Hong
    Huang, Yu-Jung
    Chu, Shao-I
    PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND SYSTEMS (ICISS 2019), 2019, : 232 - 236
  • [50] A Survey on Deep Reinforcement Learning
    Liu Q.
    Zhai J.-W.
    Zhang Z.-Z.
    Zhong S.
    Zhou Q.
    Zhang P.
    Xu J.
    2018, Science Press (41): : 1 - 27