Delay-aware model-based reinforcement learning for continuous control

被引:26
|
作者
Chen, Baiming [1 ]
Xu, Mengdi [2 ]
Li, Liang [1 ]
Zhao, Ding [2 ]
机构
[1] Tsinghua Univ, Beijing 100084, Peoples R China
[2] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
关键词
Model-based reinforcement learning; Markov decision process; Continuous control; Delayed system; FINITE SPECTRUM ASSIGNMENT; DEEP NEURAL-NETWORKS; SMITH PREDICTOR; SYSTEMS; INTEGRATOR; STABILITY; ROBOT;
D O I
10.1016/j.neucom.2021.04.015
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Action delays degrade the performance of reinforcement learning in many real-world systems. This paper proposes a formal definition of delay-aware Markov Decision Process and proves it can be transformed into standard MDP with augmented states using the Markov reward process. We develop a delay-aware model-based reinforcement learning framework that can incorporate the multi-step delay into the learned system models without learning effort. Experiments with the Gym and MuJoCo platforms show that the proposed delay-aware model-based algorithm is more efficient in training and transferable between systems with various durations of delay compared with state-of-the-art model-free reinforce-ment learning methods. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:119 / 128
页数:10
相关论文
共 50 条
  • [31] Fault Tolerant Control combining Reinforcement Learning and Model-based Control
    Bhan, Luke
    Quinones-Grueiro, Marcos
    Biswas, Gautam
    [J]. 5TH CONFERENCE ON CONTROL AND FAULT-TOLERANT SYSTEMS (SYSTOL 2021), 2021, : 31 - 36
  • [32] DC2: Delay-aware Compression Control for Distributed Machine Learning
    Abdelmoniem, Ahmed M.
    Canini, Marco
    [J]. IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2021), 2021,
  • [33] Sliding mode heading control for AUV based on continuous hybrid model-free and model-based reinforcement learning
    Wang, Dianrui
    Shen, Yue
    Wan, Junhe
    Sha, Qixin
    Li, Guangliang
    Chen, Guanzhong
    He, Bo
    [J]. APPLIED OCEAN RESEARCH, 2022, 118
  • [34] Efficient Exploration in Continuous-time Model-based Reinforcement Learning
    Treven, Lenart
    Hubotter, Jonas
    Sukhija, Bhavya
    Dorfler, Florian
    Krause, Andreas
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [35] Delay-aware reliable broadcast scheme based on power control for VANETs
    GUO Weijie
    HUANG Liusheng
    SUN Quan
    XU Hongli
    ZHANG Haoran
    [J]. TheJournalofChinaUniversitiesofPostsandTelecommunications., 2014, 21 (01) - 35
  • [36] Delay-Aware Resource Allocation in Fog-Assisted IoT Networks Through Reinforcement Learning
    Fan, Qiang
    Bai, Jianan
    Zhang, Hongxia
    Yi, Yang
    Liu, Lingjia
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (07): : 5189 - 5199
  • [37] Uncertainty-Aware Model-Based Offline Reinforcement Learning for Automated Driving
    Diehl, Christopher
    Sievernich, Timo Sebastian
    Kruger, Martin
    Hoffmann, Frank
    Bertram, Torsten
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (02) : 1167 - 1174
  • [38] Safe Model-Based Reinforcement Learning With an Uncertainty-Aware Reachability Certificate
    Yu, Dongjie
    Zou, Wenjun
    Yang, Yujie
    Ma, Haitong
    Li, Shengbo Eben
    Yin, Yuming
    Chen, Jianyu
    Duan, Jingliang
    [J]. IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2023, : 1 - 14
  • [39] Uncertainty-Aware Contact-Safe Model-Based Reinforcement Learning
    Kuo, Cheng-Yu
    Schaarschmidt, Andreas
    Cui, Yunduan
    Asfour, Tamim
    Matsubara, Takamitsu
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 3918 - 3925
  • [40] Cognitive Control Predicts Use of Model-based Reinforcement Learning
    Otto, A. Ross
    Skatova, Anya
    Madlon-Kay, Seth
    Daw, Nathaniel D.
    [J]. JOURNAL OF COGNITIVE NEUROSCIENCE, 2015, 27 (02) : 319 - 333