Multiphase Autonomous Docking via Model-Based and Hierarchical Reinforcement Learning

被引:0
|
作者
Aborizk, Anthony [1 ]
Fitz-Coy, Norman [1 ]
机构
[1] Univ Florida, Dept Mechan & Aerospace Engn, Gainesville, FL 32611 USA
基金
美国国家科学基金会;
关键词
Reinforcement Learning; Linear Quadratic Regulator; Satellites; Structural Reliability Analysis; Space Autonomous Logistics; Autonomous Systems; Planets; Spacecraft Mission Design; Linear Quadratic Gaussian; Research Facilities and Instrumentation;
D O I
10.2514/1.A35683
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
With the rise of traffic around Earth's orbit, spacecraft mission designs have placed an unprecedented demand on the capabilities of autonomous systems. In the early 2000s, the state-of-the-art autonomous spacecraft controllers were designed for static and uncluttered environments. A little over a decade later, the challenges facing spacecraft autonomy now include cluttered, dynamic environments with time-varying constraints, logical modes, fault tolerances, uncertain dynamics, and complex maneuvers. With this rise in complexity, many areas of research have been investigating more experimental control strategies, such as reinforcement learning (RL), as a potential solution to this problem. The research presented herein aims to expand on efforts to quantify the use of RL in autonomous rendezvous, proximity operations, and docking (ARPOD) environments, with consideration to the inherent drawbacks of the more common algorithms present in the field. We present hierarchical model-based RL as a solution to an autonomous docking problem. This algorithm can learn satellite parameters, extrapolate trajectory information, and learn uncertain dynamics via data collection. By using gradient-free model predictive control logic, the algorithm can handle nondifferentiable objectives and complex constraints. Lastly, the hierarchical structure demonstrates an ability to generate feasible trajectories in the presence of integrated third-party subcontrollers commonly found in spacecraft. This study highlights the ability of the hierarchical algorithm to combine and manipulate third-party subpolicies to achieve trajectories not previously trained on.
引用
收藏
页码:993 / 1005
页数:13
相关论文
共 50 条
  • [41] Continual Model-Based Reinforcement Learning with Hypernetworks
    Huang, Yizhou
    Xie, Kevin
    Bharadhwaj, Homanga
    Shkurti, Florian
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 799 - 805
  • [42] Model-Based Reinforcement Learning in Robotics: A Survey
    Sun S.
    Lan X.
    Zhang H.
    Zheng N.
    [J]. Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2022, 35 (01): : 1 - 16
  • [43] Model-based Reinforcement Learning and the Eluder Dimension
    Osband, Ian
    Van Roy, Benjamin
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [44] Transferring Instances for Model-Based Reinforcement Learning
    Taylor, Matthew E.
    Jong, Nicholas K.
    Stone, Peter
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PART II, PROCEEDINGS, 2008, 5212 : 488 - 505
  • [45] Consistency of Fuzzy Model-Based Reinforcement Learning
    Busoniu, Lucian
    Ernst, Damien
    De Schutter, Bart
    Babuska, Robert
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-5, 2008, : 518 - +
  • [46] Abstraction Selection in Model-Based Reinforcement Learning
    Jiang, Nan
    Kulesza, Alex
    Singh, Satinder
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 37, 2015, 37 : 179 - 188
  • [47] Asynchronous Methods for Model-Based Reinforcement Learning
    Zhang, Yunzhi
    Clavera, Ignasi
    Tsai, Boren
    Abbeel, Pieter
    [J]. CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [48] Online Constrained Model-based Reinforcement Learning
    van Niekerk, Benjamin
    Damianou, Andreas
    Rosman, Benjamin
    [J]. CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI2017), 2017,
  • [49] Calibrated Model-Based Deep Reinforcement Learning
    Malik, Ali
    Kuleshov, Volodymyr
    Song, Jiaming
    Nemer, Danny
    Seymour, Harlan
    Ermon, Stefano
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [50] Learning to see via epiretinal implant stimulation in silico with model-based deep reinforcement learning
    Lavoie, Jacob
    Besrour, Marwan
    Lemaire, William
    Rouat, Jean
    Fontaine, Rejean
    Plourde, Eric
    [J]. BIOMEDICAL PHYSICS & ENGINEERING EXPRESS, 2024, 10 (02)