Deep reinforcement learning with significant multiplications inference

被引:0
|
作者
Ivanov, Dmitry A. [1 ,2 ]
Larionov, Denis A. [2 ,3 ]
Kiselev, Mikhail V. [2 ,3 ]
Dylov, Dmitry V. [4 ,5 ]
机构
[1] Lomonosov Moscow State Univ, GSP 1,Leninskie Gory, Moscow 119991, Russia
[2] Cifrum, 3 Kholodilnyy per, Moscow 115191, Russia
[3] Chuvash State Univ, 15 Moskovsky pr, Cheboksary 428015, Chuvash, Russia
[4] Skolkovo Inst Sci & Technol, 30 1 Bolshoi blvd, Moscow 121205, Russia
[5] Artificial Intelligence Res Inst, 32 1 Kutuzovsky pr, Moscow 121170, Russia
基金
俄罗斯基础研究基金会;
关键词
D O I
10.1038/s41598-023-47245-y
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
We propose a sparse computation method for optimizing the inference of neural networks in reinforcement learning (RL) tasks. Motivated by the processing abilities of the brain, this method combines simple neural network pruning with a delta-network algorithm to account for the input data correlations. The former mimics neuroplasticity by eliminating inefficient connections; the latter makes it possible to update neuron states only when their changes exceed a certain threshold. This combination significantly reduces the number of multiplications during the neural network inference for fast neuromorphic computing. We tested the approach in popular deep RL tasks, yielding up to a 100-fold reduction in the number of required multiplications without substantial performance loss (sometimes, the performance even improved).
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Online Reinforcement Learning by Bayesian Inference
    Xia, Zhongpu
    Zhao, Dongbin
    2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [22] Model-Based Deep Reinforcement Learning with Traffic Inference for Traffic Signal Control
    Wang, Hao
    Zhu, Jinan
    Gu, Bao
    APPLIED SCIENCES-BASEL, 2023, 13 (06):
  • [23] Gene Networks Inference by Reinforcement Learning
    Bonini, Rodrigo Cesar
    Martins-, David Correa, Jr.
    ADVANCES IN BIOINFORMATICS AND COMPUTATIONAL BIOLOGY, BSB 2023, 2023, 13954 : 136 - 147
  • [24] DNN Inference Acceleration for Smart Devices in Industry 5.0 by Decentralized Deep Reinforcement Learning
    Dong, Chongwu
    Shafiq, Muhammad
    Al Dabel, Maryam M.
    Sun, Yanbin
    Tian, Zhihong
    IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2024, 70 (01) : 1519 - 1530
  • [25] Communication Scheduling by Deep Reinforcement Learning for Remote Traffic State Estimation With Bayesian Inference
    Peng, Bile
    Xie, Yuhang
    Seco-Granados, Gonzalo
    Wymeersch, Henk
    Jorswieck, Eduard A.
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (04) : 4287 - 4300
  • [26] DECAF: Deep Case-based Policy Inference for knowledge transfer in Reinforcement Learning
    Glatt, Ruben
    Da Silva, Felipe Leno
    da Costa Bianchi, Reinaldo Augusto
    Reali Costa, Anna Helena
    EXPERT SYSTEMS WITH APPLICATIONS, 2020, 156
  • [27] From Reinforcement Learning to Deep Reinforcement Learning: An Overview
    Agostinelli, Forest
    Hocquet, Guillaume
    Singh, Sameer
    Baldi, Pierre
    BRAVERMAN READINGS IN MACHINE LEARNING: KEY IDEAS FROM INCEPTION TO CURRENT STATE, 2018, 11100 : 298 - 328
  • [28] AdderNet: Do We Really Need Multiplications in Deep Learning?
    Chen, Hanting
    Wang, Yunhe
    Xu, Chunjing
    Shi, Boxin
    Xu, Chao
    Tian, Qi
    Xu, Chang
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 1465 - 1474
  • [29] Reinforcement Learning Based Online Request Scheduling Framework for Workload-Adaptive Edge Deep Learning Inference
    Tan, Xinrui
    Li, Hongjia
    Xie, Xiaofei
    Guo, Lu
    Ansari, Nirwan
    Huang, Xueqing
    Wang, Liming
    Xu, Zhen
    Liu, Yang
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (12) : 13222 - 13239
  • [30] Transfer Learning in Deep Reinforcement Learning
    Islam, Tariqul
    Abid, Dm. Mehedi Hasan
    Rahman, Tanvir
    Zaman, Zahura
    Mia, Kausar
    Hossain, Ramim
    PROCEEDINGS OF SEVENTH INTERNATIONAL CONGRESS ON INFORMATION AND COMMUNICATION TECHNOLOGY, ICICT 2022, VOL 1, 2023, 447 : 145 - 153