Spiking Neural Network Discovers Energy-Efficient Hexapod Motion in Deep Reinforcement Learning

被引:4
|
作者
Naya, Katsumi [1 ]
Kutsuzawa, Kyo [1 ]
Owaki, Dai [1 ]
Hayashibe, Mitsuhiro [1 ]
机构
[1] Tohoku Univ, Grad Sch Engn, Dept Robot, Neurorobot Lab, Sendai, Miyagi 9808579, Japan
关键词
Neurons; Legged locomotion; Task analysis; Robots; Computational modeling; Torque; Mathematical models; Spiking neural network; deep reinforcement learning; energy efficiency; hexapod gait; spatio-temporal backpropagation; MODEL;
D O I
10.1109/ACCESS.2021.3126311
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In Deep Reinforcement Learning (DRL) for robotics application, it is important to find energy-efficient motions. For this purpose, a standard method is to set an action penalty in the reward to find the optimal motion considering the energy expenditure. This method is widely used for the simplicity of implementation. However, since the reward is a linear sum, if the penalty is too large, the system will fall into local minima and no moving solution can be obtained. In contrast, if the penalty is too small, the effect may not be sufficient. Therefore, it is necessary to adjust the amount of the penalty so that the agent always moves dynamically, and the energy-saving effect is sufficient. Nevertheless, since adjusting the hyperparameters is computationally expensive, we need a learning method that is robust to the penalty setting problem. We investigated on the Spiking Neural Network (SNN), which has been attracting attention for its computational efficiency and neuromorphic architecture. We conducted gait experiments using a hexapod agent while varying the energy penalty settings in the simulation environment. By applying SNN to the conventional state-of-the-art DRL algorithms, we examined whether the agent could explore for an optimal gait with a larger penalty variation and obtain an energy-efficient gait verified with Cost of Transport (CoT), a metric of energy efficiency for gait. Soft Actor-Critic (SAC)+SNN resulted in a CoT of 1.64, Twin Delayed Deep Deterministic policy gradient (TD3)+SNN resulted in a CoT of 2.21, and Deep Deterministic policy gradient (DDPG)+SNN resulted in a CoT of 2.08 (1.91 for normal SAC, 2.38 for TD3, and 2.40 for DDPG). DRL combined with SNN succeeded in learning more energy efficient gait with lower CoT.
引用
收藏
页码:150345 / 150354
页数:10
相关论文
共 50 条
  • [1] A Hybrid Spiking Neural Network Reinforcement Learning Agent for Energy-Efficient Object Manipulation
    Oikonomou, Katerina Maria
    Kansizoglou, Ioannis
    Gasteratos, Antonios
    [J]. MACHINES, 2023, 11 (02)
  • [2] Reinforcement co-Learning of Deep and Spiking Neural Networks for Energy-Efficient Mapless Navigation with Neuromorphic Hardware
    Tang, Guangzhi
    Kumar, Neelesh
    Michmizos, Konstantinos P.
    [J]. 2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 6090 - 6097
  • [3] Spiking-NeRF: Spiking Neural Network for Energy-Efficient Neural Rendering
    Li, Ziwen
    Ma, Yu
    Zhou, Jindong
    Zhou, Pingqiang
    [J]. ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2024, 20 (03)
  • [4] Energy-Efficient Ultra-Dense Network With Deep Reinforcement Learning
    Ju, Hyungyu
    Kim, Seungnyun
    Kim, Youngjoon
    Shim, Byonghyo
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (08) : 6539 - 6552
  • [5] Spiking-YOLO: Spiking Neural Network for Energy-Efficient Object Detection
    Kim, Seijoon
    Park, Seongsik
    Na, Byunggook
    Yoon, Sungroh
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11270 - 11277
  • [6] Area- and Energy-Efficient STDP Learning Algorithm for Spiking Neural Network SoC
    Kim, Giseok
    Kim, Kiryong
    Choi, Sara
    Jang, Hyo Jung
    Jung, Seong-Ook
    [J]. IEEE ACCESS, 2020, 8 : 216922 - 216932
  • [7] Energy-Efficient Reinforcement Learning for Motion Planning of AUV
    Wen, Jiayi
    Zhu, Jingwei
    Lin, Yejin
    Zhang, Guichen
    [J]. 2022 IEEE 9TH INTERNATIONAL CONFERENCE ON UNDERWATER SYSTEM TECHNOLOGY: THEORY AND APPLICATIONS, USYS, 2022,
  • [8] ACCURATE, ENERGY-EFFICIENT CLASSIFICATION WITH SPIKING RANDOM NEURAL NETWORK
    Hussain, Khaled F.
    Bassyouni, Mohamed Yousef
    Gelenbe, Erol
    [J]. PROBABILITY IN THE ENGINEERING AND INFORMATIONAL SCIENCES, 2021, 35 (01) : 51 - 61
  • [9] Towards an energy-efficient Data Center Network based on deep reinforcement learning
    Wang, Yang
    Li, Yutong
    Wang, Ting
    Liu, Gang
    [J]. Computer Networks, 2022, 210
  • [10] Towards an energy-efficient Data Center Network based on deep reinforcement learning
    Wang, Yang
    Li, Yutong
    Wang, Ting
    Liu, Gang
    [J]. COMPUTER NETWORKS, 2022, 210