Continuous Value Iteration (CVI) Reinforcement Learning and Imaginary Experience Replay (IER) for learning multi-goal, continuous action and state space controllers

被引:0
|
作者
Gerken, Andreas [1 ]
Spranger, Michael [1 ]
机构
[1] Sony Comp Sci Labs Inc, Tokyo, Japan
关键词
KERNEL;
D O I
10.1109/icra.2019.8794347
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a novel model-free Reinforcement Learning algorithm for learning behavior in continuous action, state, and goal spaces. The algorithm approximates optimal value functions using non-parametric estimators. It is able to efficiently learn to reach multiple arbitrary goals in deterministic and nondeterministic environments. To improve generalization in the goal space, we propose a novel sample augmentation technique. Using these methods, robots learn faster and overall better controllers. We benchmark the proposed algorithms using simulation and a real-world voltage controlled robot that learns to maneuver in a non-observable Cartesian task space.
引用
收藏
页码:7173 / 7179
页数:7
相关论文
共 50 条
  • [1] A reinforcement learning with switching controllers for a continuous action space
    Nagayoshi, Masato
    Murao, Hajime
    Tamaki, Hisashi
    ARTIFICIAL LIFE AND ROBOTICS, 2010, 15 (01) : 97 - 100
  • [2] Learning Multi-Goal Dialogue Strategies Using Reinforcement Learning With Reduced State-Action Spaces
    Cuayahuitl, Heriberto
    Renals, Steve
    Lemon, Oliver
    Shimodaira, Hiroshi
    INTERSPEECH 2006 AND 9TH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, VOLS 1-5, 2006, : 469 - +
  • [3] Efficient Multi-Goal Reinforcement Learning via Value Consistency Prioritization
    Xu, Jiawei
    Li, Shuxing
    Yang, Rui
    Yuan, Chun
    Han, Lei
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2023, 77 : 355 - 376
  • [4] Efficient Multi-Goal Reinforcement Learning via Value Consistency Prioritization
    Xu J.
    Li S.
    Yang R.
    Yuan C.
    Han L.
    Journal of Artificial Intelligence Research, 2023, 77 : 355 - 376
  • [5] Switching reinforcement learning for continuous action space
    Nagayoshi, Masato
    Murao, Hajime
    Tamaki, Hisashi
    ELECTRONICS AND COMMUNICATIONS IN JAPAN, 2012, 95 (03) : 37 - 44
  • [6] Quantum reinforcement learning in continuous action space
    Wu, Shaojun
    Jin, Shan
    Wen, Dingding
    Han, Donghong
    Wang, Xiaoting
    QUANTUM, 2025, 9 : 1 - 18
  • [7] Budgeted Reinforcement Learning in Continuous State Space
    Carrara, Nicolas
    Leurent, Edouard
    Laroche, Romain
    Urvoy, Tanguy
    Maillard, Odalric-Ambrym
    Pietquin, Olivier
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [8] Goal-Oriented Obstacle Avoidance with Deep Reinforcement Learning in Continuous Action Space
    Cimurs, Reinis
    Lee, Jin Han
    Suh, Il Hong
    ELECTRONICS, 2020, 9 (03)
  • [9] Reinforcement learning algorithm with CTRNN in continuous action space
    Arie, Hiroaki
    Namikawa, Jun
    Ogata, Tetsuya
    Tani, Jun
    Sugano, Shigeki
    NEURAL INFORMATION PROCESSING, PT 1, PROCEEDINGS, 2006, 4232 : 387 - 396
  • [10] Swarm Reinforcement Learning Methods for Problems with Continuous State-Action Space
    Iima, Hitoshi
    Kuroe, Yasuaki
    Emoto, Kazuo
    2011 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2011, : 2173 - 2180