Real-time Motion Generation for Imaginary Creatures Using Hierarchical Reinforcement Learning

被引:0
|
作者
Ogaki, Keisuke [1 ]
Nakamura, Masayoshi [1 ]
机构
[1] DWANGO Co Ltd, Tokyo, Japan
关键词
Reinforcement Learning; Q-Learning; Neural Network;
D O I
10.1145/3214822.3214826
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Describing the motions of imaginary original creatures is an essential part of animations and computer games. One approach to generate such motions involves finding an optimal motion for approaching a goal by using the creatures' body and motor skills. Currently, researchers are employing deep reinforcement learning (DeepRL) to find such optimal motions. Some end-to-end DeepRL approaches learn the policy function, which outputs target pose for each joint according to the environment. In our study, we employed a hierarchical approach with a separate DeepRL decision maker and simple exploration-based sequence maker, and an action token, through which these two layers can communicate. By optimizing these two functions independently, we can achieve a light, fast-learning system available on mobile devices. In addition, we propose another technique to learn the policy at a faster pace with the help of a heuristic rule. By treating the heuristic rule as an additional action token, we can naturally incorporate it via Q-learning. The experimental results show that creatures can achieve better performance with the use of both heuristics and DeepRL than by using them independently.
引用
下载
收藏
页数:2
相关论文
共 50 条
  • [31] Real-time model calibration with deep reinforcement learning
    Tian, Yuan
    Chao, Manuel Arias
    Kulkarni, Chetan
    Goebel, Kai
    Fink, Olga
    MECHANICAL SYSTEMS AND SIGNAL PROCESSING, 2022, 165
  • [32] BoB: Bandwidth Prediction for Real-Time Communications Using Heuristic and Reinforcement Learning
    Bentaleb, Abdelhak
    Akcay, Mehmet N.
    Lim, May
    Begen, Ali C.
    Zimmermann, Roger
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 6930 - 6945
  • [33] Real-Time Microgrid Energy Scheduling Using Meta-Reinforcement Learning
    Shen, Huan
    Shen, Xingfa
    Chen, Yiming
    ENERGIES, 2024, 17 (10)
  • [34] Experience Replay for Real-Time Reinforcement Learning Control
    Adam, Sander
    Busoniu, Lucian
    Babuska, Robert
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2012, 42 (02): : 201 - 212
  • [35] Real-Time Bidding by Reinforcement Learning in Display Advertising
    Cai, Han
    Ren, Kan
    Zhang, Weinan
    Malialis, Kleanthis
    Wang, Jun
    Yu, Yong
    Guo, Defeng
    WSDM'17: PROCEEDINGS OF THE TENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2017, : 661 - 670
  • [36] Real-Time Lane Configuration with Coordinated Reinforcement Learning
    Gunarathna, Udesh
    Xie, Hairuo
    Tanin, Egemen
    Karunasekara, Shanika
    Borovica-Gajic, Renata
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: APPLIED DATA SCIENCE TRACK, ECML PKDD 2020, PT IV, 2021, 12460 : 291 - 307
  • [37] Real-Time Trajectory Adaptation for Quadrupedal Locomotion using Deep Reinforcement Learning
    Gangapurwala, Siddhant
    Geisert, Mathieu
    Orsolino, Romeo
    Fallon, Maurice
    Havoutis, Ioannis
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 5973 - 5979
  • [38] Real-Time Safety Optimization of Connected Vehicle Trajectories Using Reinforcement Learning
    Ghoul, Tarek
    Sayed, Tarek
    SENSORS, 2021, 21 (11)
  • [39] Evolving population method for real-time reinforcement learning
    Kim, Man-Je
    Kim, Jun Suk
    Ahn, Chang Wook
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 229
  • [40] Reinforcement Learning Based on Real-Time Iteration NMPC
    Zanon, Mario
    Kungurtsev, Vyacheslav
    Gros, Sebastien
    IFAC PAPERSONLINE, 2020, 53 (02): : 5213 - 5218