A maintenance planning framework using online and offline deep reinforcement learning

被引:5
|
作者
Bukhsh, Zaharah A. [1 ]
Molegraaf, Hajo [2 ]
Jansen, Nils [3 ]
机构
[1] Eindhoven Univ Technol, Eindhoven, Netherlands
[2] Rolsch Assetmanagement, Enschede, Netherlands
[3] Radboud Univ Nijmegen, Nijmegen, Netherlands
基金
欧洲研究理事会;
关键词
Deep reinforcement learning; Maintenance planning; Water distribution systems; Conservative Q-learning; Deep Q-networks; Offline DRL; BUDGET ALLOCATION;
D O I
10.1007/s00521-023-08560-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cost-effective asset management is an area of interest across several industries. Specifically, this paper develops a deep reinforcement learning (DRL) solution to automatically determine an optimal rehabilitation policy for continuously deteriorating water pipes. We approach the problem of rehabilitation planning in an online and offline DRL setting. In online DRL, the agent interacts with a simulated environment of multiple pipes with distinct lengths, materials, and failure rate characteristics. We train the agent using deep Q-learning (DQN) to learn an optimal policy with minimal average costs and reduced failure probability. In offline learning, the agent uses static data, e.g., DQN replay data, to learn an optimal policy via a conservative Q-learning algorithm without further interactions with the environment. We demonstrate that DRL-based policies improve over standard preventive, corrective, and greedy planning alternatives. Additionally, learning from the fixed DQN replay dataset in an offline setting further improves the performance. The results warrant that the existing deterioration profiles of water pipes consisting of large and diverse states and action trajectories provide a valuable avenue to learn rehabilitation policies in the offline setting, which can be further fine-tuned using the simulator.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] UAV online path planning technology based on deep reinforcement learning
    Fan, Jiaxuan
    Wang, Zhenya
    Ren, Jinlei
    Lu, Ying
    Liu, Yiheng
    [J]. 2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 5382 - 5386
  • [22] Sample Efficient Offline-to-Online Reinforcement Learning
    Guo, Siyuan
    Zou, Lixin
    Chen, Hechang
    Qu, Bohao
    Chi, Haotian
    Yu, Philip S.
    Chang, Yi
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (03) : 1299 - 1310
  • [23] Path Planning for Construction Machines by Offline Reinforcement Learning
    Nakayama, Tatsuya
    Kashi, Haruki
    Uchimura, Yutaka
    [J]. IEEJ Transactions on Industry Applications, 2024, 144 (05) : 367 - 373
  • [24] Offline replay supports planning in human reinforcement learning
    Momennejad, Ida
    Otto, A. Ross
    Daw, Nathaniel D.
    Norman, Kenneth A.
    [J]. ELIFE, 2018, 7
  • [25] An Offline-Transfer-Online Framework for Cloud-Edge Collaborative Distributed Reinforcement Learning
    Zeng, Tianyu
    Zhang, Xiaoxi
    Duan, Jingpu
    Yu, Chao
    Wu, Chuan
    Chen, Xu
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (05) : 720 - 731
  • [26] An Optimal Disassembly Sequence Planning for Complex Products using Enhanced Deep Reinforcement Learning Framework
    Mirothali Chand
    Chandrasekar Ravi
    [J]. SN Computer Science, 5 (5)
  • [27] Mild Hybrid Electric Vehicle Powertrain Control Using Offline-Online Hybrid Deep Reinforcement Learning Strategy
    Yao, Zhengyu
    Yoon, Hwan-Sik
    [J]. SAE INTERNATIONAL JOURNAL OF ELECTRIFIED VEHICLES, 2023, 12 (03): : 331 - 341
  • [28] Agile Cache Replacement in Edge Computing via Offline-Online Deep Reinforcement Learning
    Wang, Zhe
    Hu, Jia
    Min, Geyong
    Zhao, Zhiwei
    Wang, Zi
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (04) : 663 - 674
  • [29] Offline and Online Deep Learning for Image Recognition
    Nguyen Huu Phong
    Ribeiro, Bernardete
    [J]. PROCEEDINGS OF 2017 4TH EXPERIMENT@INTERNATIONAL CONFERENCE (EXP.AT'17), 2017, : 171 - 175
  • [30] Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning
    You, Changxi
    Lu, Jianbo
    Filev, Dimitar
    Tsiotras, Panagiotis
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 114 : 1 - 18