Bayesian reinforcement learning reliability analysis

被引:0
|
作者
Zhou, Tong [1 ]
Guo, Tong [2 ]
Dang, Chao [3 ]
Beer, Michael [3 ,4 ,5 ]
机构
[1] Department of Civil and Environmental Engineering, The Hong Kong Polytechnic University, Hong Kong
[2] School of Civil Engineering, Southeast University, Nanjing,211189, China
[3] Institute for Risk and Reliability, Leibniz University Hannover, Hannover,30167, Germany
[4] Institute for Risk and Reliability, University of Liverpool, Liverpool,L69 7ZF, United Kingdom
[5] International Joint Research Center for Resilient Infrastructure & International Joint Research Center for Engineering Reliability and Stochastic Mechanics, Tongji University, Shanghai,200092, China
基金
中国国家自然科学基金;
关键词
Bayesian networks - Computational efficiency - Inference engines - Learning systems - Markov processes - Reinforcement learning - Statistics - Uncertainty analysis;
D O I
暂无
中图分类号
学科分类号
摘要
A Bayesian reinforcement learning reliability method that combines Bayesian inference for the failure probability estimation and reinforcement learning-guided sequential experimental design is proposed. The reliability-oriented sequential experimental design is framed as a finite-horizon Markov decision process (MDP), with the associated utility function defined by a measure of epistemic uncertainty about Kriging-estimated failure probability, referred to as integrated probability of misclassification (IPM). On this basis, a one-step Bayes optimal learning function termed integrated probability of misclassification reduction (IPMR), along with a compatible convergence criterion, is defined. Three effective strategies are implemented to accelerate IPMR-informed sequential experimental design: (i) Analytical derivation of the inner expectation in IPMR, simplifying it to a single expectation. (ii) Substitution of IPMR with its upper bound IPMRU to avoid element-wise computation of its integrand. (iii) Rational pruning of both quadrature set and candidate pool in IPMRU to alleviate computer memory constraint. The efficacy of the proposed approach is demonstrated on two benchmark examples and two numerical examples. Results indicate that IPMRU facilitates a much more rapid reduction of IPM compared to other existing learning functions, while requiring much less computational time than IPMR itself. Therefore, the proposed reliability method offers a substantial advantage in both computational efficiency and accuracy, especially in complex dynamic reliability problems. © 2024
引用
收藏
相关论文
共 50 条
  • [21] Bayesian Reinforcement Learning and Bayesian Deep Learning for Blockchains With Mobile Edge Computing
    Asheralieva, Alia
    Niyato, Dusit
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2021, 7 (01) : 319 - 335
  • [22] Stability and Reliability Analysis for Multiple WT Using Deep Reinforcement Learning
    Padmaja, S. M.
    Kumar, Akhilesh
    Dhanesh, Lavanya
    Kamesh, Vinjamuri Venkata
    Geetha, H.
    Ramya, D.
    Raja Sekhar, G. G.
    Saha, Nutan
    ELECTRIC POWER COMPONENTS AND SYSTEMS, 2024, 52 (02) : 308 - 321
  • [23] An active learning Bayesian ensemble surrogate model for structural reliability analysis
    Xiao, Tianli
    Park, Chanseok
    Ouyang, Linhan
    Ma, Yizhong
    QUALITY AND RELIABILITY ENGINEERING INTERNATIONAL, 2022, 38 (07) : 3579 - 3597
  • [24] Structural reliability analysis by line sampling: A Bayesian active learning treatment
    Dang, Chao
    Valdebenito, Marcos A.
    Faes, Matthias G. R.
    Song, Jingwen
    Wei, Pengfei
    Beer, Michael
    STRUCTURAL SAFETY, 2023, 104
  • [25] Reliability Analysis of Structures by Active Learning Enhanced Sparse Bayesian Regression
    Roy, Atin
    Chakraborty, Subrata
    Adhikari, Sondipon
    JOURNAL OF ENGINEERING MECHANICS, 2023, 149 (05)
  • [26] Shaping Bayesian Network Based Reinforcement Learning
    Song, Jiong
    Jin, Zhao
    2012 INTERNATIONAL CONFERENCE ON INDUSTRIAL CONTROL AND ELECTRONICS ENGINEERING (ICICEE), 2012, : 742 - 745
  • [27] Robust Reinforcement Learning with Bayesian Optimisation and Quadrature
    Paul, Supratik
    Chatzilygeroudis, Konstantinos
    Ciosek, Kamil
    Mouret, Jean-Baptiste
    Osborne, Michael A.
    Whiteson, Shimon
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [28] Approximate planning for bayesian hierarchical reinforcement learning
    Ngo Anh Vien
    Hung Ngo
    Lee, Sungyoung
    Chung, TaeChoong
    APPLIED INTELLIGENCE, 2014, 41 (03) : 808 - 819
  • [29] A Bayesian Posterior Updating Algorithm in Reinforcement Learning
    Xiong, Fangzhou
    Liu, Zhiyong
    Yang, Xu
    Sun, Biao
    Chiu, Charles
    Qiao, Hong
    NEURAL INFORMATION PROCESSING, ICONIP 2017, PT V, 2017, 10638 : 418 - 426
  • [30] Surveillance Evasion Through Bayesian Reinforcement Learning
    Cornell University, United States
    arXiv, 1600,