Efficient Uncertainty Propagation for Reinforcement Learning with Limited Data

被引:0
|
作者
Hans, Alexander [1 ]
Udluft, Steffen [1 ]
机构
[1] Siemens AG, Corp Technol Informat & Commun, Learning Syst, D-81739 Munich, Germany
关键词
Reinforcement learning; model-based; uncertainty; Bayesian modeling;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In a typical reinforcement learning (RL) setting details of the environment are not given explicitly but have to be estimated from observations. Most R,L approaches only optimize the expected value. However, if the number of observations is limited considering expected values only can lead to false conclusions. Instead, it is crucial to also account for the estimator's uncertainties. In this paper, we present a method to incorporate those uncertainties and propagate them to the conclusions. By being only approximate, the method is computationally feasible. Furthermore, we describe a Bayesian approach to design the estimators. Our experiments show that the method considerably increases the robustness of the derived policies compared to the standard approach.
引用
收藏
页码:70 / 79
页数:10
相关论文
共 50 条
  • [1] Uncertainty Propagation for Efficient Exploration in Reinforcement Learning
    Hans, Alexander
    Udluft, Steffen
    [J]. ECAI 2010 - 19TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2010, 215 : 361 - 366
  • [2] Data-Efficient Pipeline for Offline Reinforcement Learning with Limited Data
    Nie, Allen
    Flet-Berliac, Yannis
    Jordan, Deon R.
    Steenbergen, William
    Brunskill, Emma
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [3] Uncertainty Propagation for Quality Assurance in Reinforcement Learning
    Schneegass, Daniel
    Udluft, Steffen
    Martinetz, Thomas
    [J]. 2008 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-8, 2008, : 2588 - 2595
  • [4] Online inverse reinforcement learning with limited data
    Self, Ryan
    Mahmud, S. M. Nahid
    Hareland, Katrine
    Kamalapurkar, Rushikesh
    [J]. 2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 603 - 608
  • [5] Uncertainty Estimation based Intrinsic Reward For Efficient Reinforcement Learning
    Chen, Chao
    Wan, Tianjiao
    Shi, Peichang
    Ding, Bo
    Gao, Zijian
    Feng, Dawei
    [J]. 2022 IEEE 13TH INTERNATIONAL CONFERENCE ON JOINT CLOUD COMPUTING (JCC 2022), 2022, : 1 - 8
  • [6] Efficient Online Reinforcement Learning with Offline Data
    Ball, Philip J.
    Smith, Laura
    Kostrikov, Ilya
    Levine, Sergey
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [7] Data-Efficient Hierarchical Reinforcement Learning
    Nachum, Ofir
    Gu, Shixiang
    Lee, Honglak
    Levine, Sergey
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [8] Data Efficient Reinforcement Learning for Legged Robots
    Yang, Yuxiang
    Caluwaerts, Ken
    Iscen, Atil
    Zhang, Tingnan
    Tan, Jie
    Sindhwani, Vikas
    [J]. CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [9] Efficient Uncertainty Propagation in Model-Based Reinforcement Learning Unmanned Surface Vehicle Using Unscented Kalman Filter
    Wang, Jincheng
    Xia, Lei
    Peng, Lei
    Li, Huiyun
    Cui, Yunduan
    [J]. DRONES, 2023, 7 (04)
  • [10] Uncertainty-Aware Data Augmentation for Offline Reinforcement Learning
    Su, Yunjie
    Kong, Yilun
    Wang, Xueqian
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,