Risk-Averse Model Uncertainty for Distributionally Robust Safe Reinforcement Learning

被引:0
|
作者
Queeney, James [1 ]
Benosman, Mouhacine [2 ]
机构
[1] Boston Univ, Div Syst Engn, Boston, MA 02215 USA
[2] Mitsubishi Elect Res Labs, Cambridge, MA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many real-world domains require safe decision making in uncertain environments. In this work, we introduce a deep reinforcement learning framework for approaching this important problem. We consider a distribution over transition models, and apply a risk-averse perspective towards model uncertainty through the use of coherent distortion risk measures. We provide robustness guarantees for this framework by showing it is equivalent to a specific class of distributionally robust safe reinforcement learning problems. Unlike existing approaches to robustness in deep reinforcement learning, however, our formulation does not involve minimax optimization. This leads to an efficient, model-free implementation of our approach that only requires standard data collection from a single training environment. In experiments on continuous control tasks with safety constraints, we demonstrate that our framework produces robust performance and safety at deployment time across a range of perturbed test environments.
引用
收藏
页数:22
相关论文
共 50 条
  • [1] A risk-averse distributionally robust project scheduling model to address
    Bruni, Maria Elena
    Hazir, Oencu
    [J]. EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2024, 318 (02) : 398 - 407
  • [2] Efficient Risk-Averse Reinforcement Learning
    Greenberg, Ido
    Chow, Yinlam
    Ghavamzadeh, Mohammad
    Mannor, Shie
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [3] Risk-averse supply chain management via robust reinforcement learning
    Wang, Jing
    Swartz, Christopher L.E.
    Huang, Kai
    [J]. Computers and Chemical Engineering, 2025, 192
  • [4] Risk-averse Reinforcement Learning for Algorithmic Trading
    Shen, Yun
    Huang, Ruihong
    Yan, Chang
    Obermayer, Klaus
    [J]. 2014 IEEE CONFERENCE ON COMPUTATIONAL INTELLIGENCE FOR FINANCIAL ENGINEERING & ECONOMICS (CIFER), 2014, : 391 - 398
  • [5] Risk-averse Reinforcement Learning for Portfolio Optimization
    Enkhsaikhan, Bayaraa
    Jo, Ohyun
    [J]. ICT EXPRESS, 2024, 10 (04): : 857 - 862
  • [6] STABILITY OF A CLASS OF RISK-AVERSE MULTISTAGE STOCHASTIC PROGRAMS AND THEIR DISTRIBUTIONALLY ROBUST COUNTERPARTS
    Jiang, Jie
    Chen, Zhiping
    Hu, He
    [J]. JOURNAL OF INDUSTRIAL AND MANAGEMENT OPTIMIZATION, 2021, 17 (05) : 2415 - 2440
  • [7] Risk-Averse Stochastic Programming and Distributionally Robust Optimization Via Operator Splitting
    Welington de Oliveira
    [J]. Set-Valued and Variational Analysis, 2021, 29 : 861 - 891
  • [8] Risk-Averse Stochastic Programming and Distributionally Robust Optimization Via Operator Splitting
    de Oliveira, Welington
    [J]. SET-VALUED AND VARIATIONAL ANALYSIS, 2021, 29 (04) : 861 - 891
  • [9] The Value of Randomized Strategies in Distributionally Robust Risk-Averse Network Interdiction Problems
    Sadana, Utsav
    Delage, Erick
    [J]. INFORMS JOURNAL ON COMPUTING, 2023, 35 (01) : 216 - 232
  • [10] Risk-Averse Bayes-Adaptive Reinforcement Learning
    Rigter, Marc
    Lacerda, Bruno
    Hawes, Nick
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34