Bayesian reinforcement learning reliability analysis

被引:0
|
作者
Zhou, Tong [1 ]
Guo, Tong [2 ]
Dang, Chao [3 ]
Beer, Michael [3 ,4 ,5 ]
机构
[1] Department of Civil and Environmental Engineering, The Hong Kong Polytechnic University, Hong Kong
[2] School of Civil Engineering, Southeast University, Nanjing,211189, China
[3] Institute for Risk and Reliability, Leibniz University Hannover, Hannover,30167, Germany
[4] Institute for Risk and Reliability, University of Liverpool, Liverpool,L69 7ZF, United Kingdom
[5] International Joint Research Center for Resilient Infrastructure & International Joint Research Center for Engineering Reliability and Stochastic Mechanics, Tongji University, Shanghai,200092, China
基金
中国国家自然科学基金;
关键词
Bayesian networks - Computational efficiency - Inference engines - Learning systems - Markov processes - Reinforcement learning - Statistics - Uncertainty analysis;
D O I
暂无
中图分类号
学科分类号
摘要
A Bayesian reinforcement learning reliability method that combines Bayesian inference for the failure probability estimation and reinforcement learning-guided sequential experimental design is proposed. The reliability-oriented sequential experimental design is framed as a finite-horizon Markov decision process (MDP), with the associated utility function defined by a measure of epistemic uncertainty about Kriging-estimated failure probability, referred to as integrated probability of misclassification (IPM). On this basis, a one-step Bayes optimal learning function termed integrated probability of misclassification reduction (IPMR), along with a compatible convergence criterion, is defined. Three effective strategies are implemented to accelerate IPMR-informed sequential experimental design: (i) Analytical derivation of the inner expectation in IPMR, simplifying it to a single expectation. (ii) Substitution of IPMR with its upper bound IPMRU to avoid element-wise computation of its integrand. (iii) Rational pruning of both quadrature set and candidate pool in IPMRU to alleviate computer memory constraint. The efficacy of the proposed approach is demonstrated on two benchmark examples and two numerical examples. Results indicate that IPMRU facilitates a much more rapid reduction of IPM compared to other existing learning functions, while requiring much less computational time than IPMR itself. Therefore, the proposed reliability method offers a substantial advantage in both computational efficiency and accuracy, especially in complex dynamic reliability problems. © 2024
引用
收藏
相关论文
共 50 条
  • [31] Robust reinforcement learning with bayesian optimisation and quadrature
    Paul, Supratik
    Chatzilygeroudis, Konstantinos
    Ciosek, Kamil
    Mouret, Jean-Baptiste
    Osborne, Michael A.
    Whiteson, Shimon
    Journal of Machine Learning Research, 2020, 21
  • [32] Improving the Efficiency of Bayesian Inverse Reinforcement Learning
    Michini, Bernard
    How, Jonathan P.
    2012 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2012, : 3651 - 3656
  • [33] A Bayesian Approach to Robust Inverse Reinforcement Learning
    Wei, Ran
    Zeng, Siliang
    Li, Chenliang
    Garcia, Alfredo
    McDonald, Anthony
    Hong, Mingyi
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [34] Approximate planning for bayesian hierarchical reinforcement learning
    Ngo Anh Vien
    Hung Ngo
    Sungyoung Lee
    TaeChoong Chung
    Applied Intelligence, 2014, 41 : 808 - 819
  • [35] Dual control for approximate Bayesian reinforcement learning
    Klenske, Edgar D.
    Hennig, Philipp
    Journal of Machine Learning Research, 2016, 17 : 1 - 30
  • [36] Dual Control for Approximate Bayesian Reinforcement Learning
    Klenske, Edgar D.
    Hennig, Philipp
    JOURNAL OF MACHINE LEARNING RESEARCH, 2016, 17
  • [37] Online reinforcement learning control by Bayesian inference
    Xia, Zhongpu
    Zhao, Dongbin
    IET CONTROL THEORY AND APPLICATIONS, 2016, 10 (12): : 1331 - 1338
  • [38] Monte Carlo Bayesian Hierarchical Reinforcement Learning
    Ngo Anh Vien
    Hung Ngo
    Ertel, Wolfgang
    AAMAS'14: PROCEEDINGS OF THE 2014 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2014, : 1551 - 1552
  • [39] A Bayesian approach to conceptualization using reinforcement learning
    Amizadeh, Saeed
    Ahmadabadi, Majid Nili
    Araabi, Babak N.
    Siegwart, Roland
    2007 IEEE/ASME INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS, VOLS 1-3, 2007, : 749 - +
  • [40] Variational Bayesian Reinforcement Learning with Regret Bounds
    O'Donoghue, Brendan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,