Null Space Based Efficient Reinforcement Learning with Hierarchical Safety Constraints

被引:0
|
作者
Yang, Quantao [1 ]
Stork, Johannes A. [1 ]
Stoyanov, Todor [1 ]
机构
[1] Orebro Univ, Autonomous Mobile Manipulat Lab AMM, Orebro, Sweden
关键词
D O I
10.1109/ECMR50962.2021.9568848
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning is inherently unsafe for use in physical systems, as learning by trial-and-error can cause harm to the environment or the robot itself. One way to avoid unpredictable exploration is to add constraints in the action space to restrict the robot behavior. In this paper, we propose a null space based framework of integrating reinforcement learning methods in constrained continuous action spaces. We leverage a hierarchical control framework to decompose target robotic skills into higher ranked tasks (e: g:, joint limits and obstacle avoidance) and lower ranked reinforcement learning task. Safe exploration is guaranteed by only learning policies in the null space of higher prioritized constraints. Meanwhile multiple constraint phases for different operational spaces are constructed to guide the robot exploration. Also, we add penalty loss for violating higher ranked constraints to accelerate the learning procedure. We have evaluated our method on different redundant robotic tasks in simulation and show that our null space based reinforcement learning method can explore and learn safely and efficiently.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] Intelligent Mission Supervisor Design for Null-space-based Behavioral Control System: A Reinforcement Learning Approach
    Huang, Jie
    Mei, Hengquan
    Zhang, Zhenyi
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 5861 - 5866
  • [32] Integrating safety constraints into adversarial training for robust deep reinforcement learning
    Meng, Jinling
    Zhu, Fei
    Ge, Yangyang
    Zhao, Peiyao
    INFORMATION SCIENCES, 2023, 619 : 310 - 323
  • [33] Off-Policy Conservative Distributional Reinforcement Learning With Safety Constraints
    Zhang, Hengrui
    Lin, Youfang
    Han, Sheng
    Wang, Shuo
    Lv, Kai
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2025, 55 (03): : 2033 - 2045
  • [34] Sample and Feedback Efficient Hierarchical Reinforcement Learning from Human Preferences
    Pinsler, Robert
    Akrour, Riad
    Osa, Takayuki
    Peters, Jan
    Neumann, Gerhard
    2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2018, : 596 - 601
  • [35] Towards efficient RAN slicing: A deep hierarchical reinforcement learning approach
    Sun, Xiaochuan
    Qin, Zhenteng
    Zhang, Qi
    Li, Yingqi
    PHYSICAL COMMUNICATION, 2024, 66
  • [36] An Energy-Efficient Hardware Accelerator for Hierarchical Deep Reinforcement Learning
    Shiri, Aidin
    Prakash, Bharat
    Mazumder, Arnab Neelim
    Waytowich, Nicholas R.
    Oates, Tim
    Mohsenin, Tinoosh
    2021 IEEE 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS), 2021,
  • [37] Active Visual SLAM Based on Hierarchical Reinforcement Learning
    Chen, Wensong
    Li, Wei
    Yang, Andong
    Hu, Yu
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2023, : 7155 - 7162
  • [38] Online hierarchical reinforcement learning based on interrupting Option
    Zhu F.
    Xu Z.-P.
    Liu Q.
    Fu Y.-C.
    Wang H.
    Tongxin Xuebao, 6 (65-74): : 65 - 74
  • [39] Dynamic hierarchical reinforcement learning based on probability model
    Dai, Zhao-Hui
    Yuan, Jiao-Hong
    Wu, Min
    Chen, Xin
    Kongzhi Lilun Yu Yingyong/Control Theory and Applications, 2011, 28 (11): : 1595 - 1600
  • [40] A Hierarchical Framework for Quadruped Locomotion Based on Reinforcement Learning
    Tan, Wenhao
    Fang, Xing
    Zhang, Wei
    Song, Ran
    Chen, Teng
    Zheng, Yu
    Li, Yibin
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 8462 - 8468