Multiphase Autonomous Docking via Model-Based and Hierarchical Reinforcement Learning

被引:0
|
作者
Aborizk, Anthony [1 ]
Fitz-Coy, Norman [1 ]
机构
[1] Univ Florida, Dept Mechan & Aerospace Engn, Gainesville, FL 32611 USA
基金
美国国家科学基金会;
关键词
Reinforcement Learning; Linear Quadratic Regulator; Satellites; Structural Reliability Analysis; Space Autonomous Logistics; Autonomous Systems; Planets; Spacecraft Mission Design; Linear Quadratic Gaussian; Research Facilities and Instrumentation;
D O I
10.2514/1.A35683
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
With the rise of traffic around Earth's orbit, spacecraft mission designs have placed an unprecedented demand on the capabilities of autonomous systems. In the early 2000s, the state-of-the-art autonomous spacecraft controllers were designed for static and uncluttered environments. A little over a decade later, the challenges facing spacecraft autonomy now include cluttered, dynamic environments with time-varying constraints, logical modes, fault tolerances, uncertain dynamics, and complex maneuvers. With this rise in complexity, many areas of research have been investigating more experimental control strategies, such as reinforcement learning (RL), as a potential solution to this problem. The research presented herein aims to expand on efforts to quantify the use of RL in autonomous rendezvous, proximity operations, and docking (ARPOD) environments, with consideration to the inherent drawbacks of the more common algorithms present in the field. We present hierarchical model-based RL as a solution to an autonomous docking problem. This algorithm can learn satellite parameters, extrapolate trajectory information, and learn uncertain dynamics via data collection. By using gradient-free model predictive control logic, the algorithm can handle nondifferentiable objectives and complex constraints. Lastly, the hierarchical structure demonstrates an ability to generate feasible trajectories in the presence of integrated third-party subcontrollers commonly found in spacecraft. This study highlights the ability of the hierarchical algorithm to combine and manipulate third-party subpolicies to achieve trajectories not previously trained on.
引用
收藏
页码:993 / 1005
页数:13
相关论文
共 50 条
  • [1] Guided Cooperation in Hierarchical Reinforcement Learning via Model-Based Rollout
    Wang, Haoran
    Tang, Zeshen
    Sun, Yaoru
    Wang, Fang
    Zhang, Siyu
    Chen, Yeming
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [2] An Efficient Approach to Model-Based Hierarchical Reinforcement Learning
    Li, Zhuoru
    Narayan, Akshay
    Leong, Tze-Yun
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3583 - 3589
  • [3] Model-based hierarchical reinforcement learning and human action control
    Botvinick, Matthew
    Weinstein, Ari
    [J]. PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY B-BIOLOGICAL SCIENCES, 2014, 369 (1655)
  • [4] Hierarchical Model-Based Imitation Learning for Planning in Autonomous Driving
    Bronstein, Eli
    Palatucci, Mark
    Notz, Dominik
    White, Brandyn
    Kuefler, Alex
    Lu, Yiren
    Paul, Supratik
    Nikdel, Payam
    Mougin, Paul
    Chen, Hongge
    Fu, Justin
    Abrams, Austin
    Shah, Punit
    Racah, Evan
    Frenkel, Benjamin
    Whiteson, Shimon
    Anguelov, Dragomir
    [J]. 2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 8652 - 8659
  • [5] Model-Based Reinforcement Learning via Imagination with Derived Memory
    Mu, Yao
    Zhuang, Yuzheng
    Wang, Bin
    Zhu, Guangxiang
    Liu, Wulong
    Chen, Jianyu
    Luo, Ping
    Li, Shengbo Eben
    Zhang, Chongjie
    Hao, Jianye
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [6] Model-Based Probabilistic Pursuit via Inverse Reinforcement Learning
    Shkurti, Florian
    Kakodkar, Nikhil
    Dudek, Gregory
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2018, : 7804 - 7811
  • [7] Model-Based Reinforcement Learning via Proximal Policy Optimization
    Sun, Yuewen
    Yuan, Xin
    Liu, Wenzhang
    Sun, Changyin
    [J]. 2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 4736 - 4740
  • [8] AUTONOMOUS PORT NAVIGATION WITH RANGING SENSORS USING MODEL-BASED REINFORCEMENT LEARNING
    Herremans, Siemen
    Anwar, Ali
    Troch, Arne
    Ravijts, Ian
    Vangeneugden, Maarten
    Mercelis, Siegfried
    Hellinckx, Peter
    [J]. PROCEEDINGS OF ASME 2023 42ND INTERNATIONAL CONFERENCE ON OCEAN, OFFSHORE & ARCTIC ENGINEERING, OMAE2023, VOL 5, 2023,
  • [9] Autonomous Model Management via Reinforcement Learning
    Liebman, Elad
    Zavesky, Eric
    Stone, Peter
    [J]. AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2017, : 1601 - 1603
  • [10] A survey on model-based reinforcement learning
    Fan-Ming Luo
    Tian Xu
    Hang Lai
    Xiong-Hui Chen
    Weinan Zhang
    Yang Yu
    [J]. Science China Information Sciences, 2024, 67