Entropy-Guided Distributional Reinforcement Learning with Controlling Uncertainty in Robotic Tasks

被引:0
|
作者
Cho, Hyunjin [1 ]
Kim, Hyunseok [1 ]
机构
[1] Dong A Univ, Dept Comp Engn, Busan 49315, South Korea
来源
APPLIED SCIENCES-BASEL | 2025年 / 15卷 / 05期
关键词
distributional reinforcement learning; offline reinforcement learning; robot learning;
D O I
10.3390/app15052773
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
This study proposes a novel approach to enhance the stability and performance of reinforcement learning (RL) in long-horizon tasks. Overestimation bias in value function estimation and high uncertainty within environments make it difficult to determine the optimal action. To address this, we improve the truncated quantile critics algorithm by managing uncertainty in robotic applications. Our dynamic method adjusts the discount factor based on policy entropy, allowing for fine-tuning that reflects the agent's learning status. This enables the existing algorithm to learn stably even in scenarios with limited training data, ensuring more robust adaptation. By leveraging policy entropy loss, this approach effectively boosts confidence in predicting future rewards. Our experiments demonstrated an 11% increase in average evaluation return compared to traditional fixed-discount-factor approaches in the DeepMind Control Suite and Gymnasium robotics environments. This approach significantly enhances sample efficiency and adaptability in complex long-horizon tasks, highlighting the effectiveness of entropy-guided RL in navigating challenging and uncertain environments.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] An Entropy-Guided Adaptive Co-construction Method of State and Action Spaces in Reinforcement Learning
    Nagayoshi, Masato
    Murao, Hajime
    Tamaki, Hisashi
    NEURAL INFORMATION PROCESSING (ICONIP 2014), PT I, 2014, 8834 : 119 - 126
  • [2] Entropy-guided Vocabulary Augmentation of Multilingual Language Models for Low-resource Tasks
    Nag, Arijit
    Samanta, Bidisha
    Mukherjee, Animesh
    Ganguly, Niloy
    Chakrabarti, Soumen
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8619 - 8629
  • [3] Entropy-guided contrastive learning for semi-supervised medical image segmentation
    Xie, Junsong
    Wu, Qian
    Zhu, Renju
    IET IMAGE PROCESSING, 2024, 18 (02) : 312 - 326
  • [4] An Entropy-Guided Reinforced Partial Convolutional Network for Zero-Shot Learning
    Li, Yun
    Liu, Zhe
    Yao, Lina
    Wang, Xianzhi
    McAuley, Julian
    Chang, Xiaojun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (08) : 5175 - 5186
  • [5] Distributional reinforcement learning with epistemic and aleatoric uncertainty estimation
    Liu, Qi
    Li, Yanjie
    Chen, Shiyu
    Lin, Ke
    Shi, Xiongtao
    Lou, Yunjiang
    INFORMATION SCIENCES, 2023, 644
  • [6] EDRL: Entropy-guided disentangled representation learning for unsupervised domain adaptation in semantic segmentation
    Wang, Runze
    Zhou, Qin
    Zheng, Guoyan
    COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2023, 240
  • [7] SENTINEL: Taming Uncertainty with Ensemble based Distributional Reinforcement Learning
    Eriksson, Hannes
    Basu, Debabrota
    Alibeigi, Mina
    Dimitrakakis, Christos
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, VOL 180, 2022, 180 : 631 - 640
  • [8] Leveraging Deep Reinforcement Learning for Reaching Robotic Tasks
    Katyal, Kapil
    Wang, I-Jeng
    Burlina, Philippe
    2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, : 490 - 491
  • [9] Offline reinforcement learning with anderson acceleration for robotic tasks
    Zuo, Guoyu
    Huang, Shuai
    Li, Jiangeng
    Gong, Daoxiong
    APPLIED INTELLIGENCE, 2022, 52 (09) : 9885 - 9898
  • [10] Exploiting Symmetries in Reinforcement Learning of Bimanual Robotic Tasks
    Amadio, Fabio
    Colome, Adria
    Torras, Carme
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (02): : 1838 - 1845