Dynamic Action Space Handling Method for Reinforcement Learning models

被引:4
|
作者
Woo, Sangchul [1 ]
Sung, Yunsick [1 ]
机构
[1] Dongguk Univ, Dept Multimedia Engn, Seoul, South Korea
来源
关键词
Dance Tutorial System; Q-Learning; Reinforcement Learning; Virtual Tutor;
D O I
10.3745/JIPS.02.0146
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, extensive studies have been conducted to apply deep learning to reinforcement learning to solve the state-space problem. If the state-space problem was solved, reinforcement learning would become applicable in various fields. For example, users can utilize dance-tutorial systems to learn how to dance by watching and imitating a virtual instructor. The instructor can perform the optimal dance to the music, to which reinforcement learning is applied. In this study, we propose a method of reinforcement learning in which the action space is dynamically adjusted. Because actions that are not performed or are unlikely to be optimal are not learned, and the state space is not allocated, the learning time can be shortened, and the state space can be reduced. In an experiment, the proposed method shows results similar to those of traditional Q-learning even when the state space of the proposed method is reduced to approximately 0.33% of that of Q-learning. Consequently, the proposed method reduces the cost and time required for learning. Traditional Q-learning requires 6 million state spaces for learning 100,000 times. In contrast, the proposed method requires only 20,000 state spaces. A higher winning rate can be achieved in a shorter period of time by retrieving 20,000 state spaces instead of 6 million.
引用
收藏
页码:1223 / 1230
页数:8
相关论文
共 50 条
  • [1] Constrained Reinforcement Learning for Dynamic Material Handling
    Hu, Chengpeng
    Wang, Ziming
    Liu, Jialin
    Wen, Junyi
    Mao, Bifei
    Yao, Xin
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [2] Dynamic action sequences in reinforcement learning
    Moren, J
    [J]. FROM ANIMALS TO ANIMATS 5, 1998, : 366 - 371
  • [3] Adversarial Attacks on Multiagent Deep Reinforcement Learning Models in Continuous Action Space
    Zhou, Ziyuan
    Liu, Guanjun
    Guo, Weiran
    Zhou, MengChu
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024,
  • [4] Reinforcement Learning in Latent Action Sequence Space
    Kim, Heecheol
    Yamada, Masanori
    Miyoshi, Kosuke
    Iwata, Tomoharu
    Yamakawa, Hiroshi
    [J]. 2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 5497 - 5503
  • [5] Switching reinforcement learning for continuous action space
    Nagayoshi, Masato
    Murao, Hajime
    Tamaki, Hisashi
    [J]. ELECTRONICS AND COMMUNICATIONS IN JAPAN, 2012, 95 (03) : 37 - 44
  • [6] Action Space Shaping in Deep Reinforcement Learning
    Kanervisto, Anssi
    Scheller, Christian
    Hautamaki, Ville
    [J]. 2020 IEEE CONFERENCE ON GAMES (IEEE COG 2020), 2020, : 479 - 486
  • [7] Couple Particles in Action Space for Reinforcement Learning
    Notsu, Akira
    Honda, Katsuhiro
    Ichihashi, Hidetomo
    [J]. INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY, 2010, 10 (12): : 200 - 203
  • [8] Dynamic Action Repetition for Deep Reinforcement Learning
    Lakshminarayanan, Aravind S.
    Sharma, Sahil
    Ravindran, Balaraman
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2133 - 2139
  • [9] Reinforcement Learning Method for Continuous State Space Based on Dynamic Neural Network
    Sun, Wei
    Wang, Xuesong
    Cheng, Yuhu
    [J]. 2008 7TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-23, 2008, : 750 - 754
  • [10] LASER: Learning a Latent Action Space for Efficient Reinforcement Learning
    Allshire, Arthur
    Martin-Martin, Roberto
    Lin, Charles
    Manuel, Shawn
    Savarese, Silvio
    Garg, Animesh
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 6650 - 6656