Entropy-Guided Distributional Reinforcement Learning with Controlling Uncertainty in Robotic Tasks

被引:0
|
作者
Cho, Hyunjin [1 ]
Kim, Hyunseok [1 ]
机构
[1] Dong A Univ, Dept Comp Engn, Busan 49315, South Korea
来源
APPLIED SCIENCES-BASEL | 2025年 / 15卷 / 05期
关键词
distributional reinforcement learning; offline reinforcement learning; robot learning;
D O I
10.3390/app15052773
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
This study proposes a novel approach to enhance the stability and performance of reinforcement learning (RL) in long-horizon tasks. Overestimation bias in value function estimation and high uncertainty within environments make it difficult to determine the optimal action. To address this, we improve the truncated quantile critics algorithm by managing uncertainty in robotic applications. Our dynamic method adjusts the discount factor based on policy entropy, allowing for fine-tuning that reflects the agent's learning status. This enables the existing algorithm to learn stably even in scenarios with limited training data, ensuring more robust adaptation. By leveraging policy entropy loss, this approach effectively boosts confidence in predicting future rewards. Our experiments demonstrated an 11% increase in average evaluation return compared to traditional fixed-discount-factor approaches in the DeepMind Control Suite and Gymnasium robotics environments. This approach significantly enhances sample efficiency and adaptability in complex long-horizon tasks, highlighting the effectiveness of entropy-guided RL in navigating challenging and uncertain environments.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Uncertainty Estimation with Distributional Reinforcement Learning for Applications in Intelligent Transportation Systems: A Case Study
    Wang, Pengyue
    Li, Yan
    Shekhar, Shashi
    Northrop, William F.
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 3822 - 3827
  • [32] Modelling and controlling uncertainty in optimal disassembly planning through reinforcement learning
    Reveliotis, SA
    2004 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1- 5, PROCEEDINGS, 2004, : 2625 - 2632
  • [33] Efficient hindsight reinforcement learning using demonstrations for robotic tasks with sparse rewards
    Zuo, Guoyu
    Zhao, Qishen
    Lu, Jiahao
    Li, Jiangeng
    INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2020, 17 (01)
  • [34] Deep Reinforcement Learning for an Anthropomorphic Robotic Arm Under Sparse Reward Tasks
    Cheng, Hao
    Duan, Feng
    Zheng, Haosi
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2021, PT II, 2021, 13014 : 79 - 89
  • [35] Evaluations of the Gap between Supervised and Reinforcement Lifelong Learning on Robotic Manipulation Tasks
    Yang, Fan
    Yang, Chao
    Liu, Huaping
    Sun, Fuchun
    CONFERENCE ON ROBOT LEARNING, VOL 164, 2021, 164 : 547 - 556
  • [36] Learning vision-based robotic manipulation tasks sequentially in offline reinforcement learning settings
    Yadav, Sudhir Pratap
    Nagar, Rajendra
    Shah, Suril V.
    ROBOTICA, 2024, 42 (06) : 1715 - 1730
  • [37] Uncertainty-aware hierarchical reinforcement learning for long-horizon tasks
    Wenning Hu
    Hongbin Wang
    Ming He
    Nianbin Wang
    Applied Intelligence, 2023, 53 : 28555 - 28569
  • [38] Uncertainty-aware hierarchical reinforcement learning for long-horizon tasks
    Hu, Wenning
    Wang, Hongbin
    He, Ming
    Wang, Nianbin
    APPLIED INTELLIGENCE, 2023, 53 (23) : 28555 - 28569
  • [39] Uncertainty-Guided Active Reinforcement Learning with Bayesian Neural Networks
    Wu, Xinyang
    El-Shamouty, Mohamed
    Nitsche, Christof
    Huber, Marco F.
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 5751 - 5757
  • [40] User-guided reinforcement learning of robot assistive tasks for an intelligent environment
    Wang, Y
    Huber, M
    Papudesi, VN
    Cook, DJ
    IROS 2003: PROCEEDINGS OF THE 2003 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-4, 2003, : 424 - 429