Hierarchical Reinforcement Learning Framework for Stochastic Spaceflight Campaign Design

被引:9
|
作者
Takubo, Yuji [1 ]
Chen, Hao [1 ]
Ho, Koki [1 ]
机构
[1] Georgia Inst Technol, Aerosp Engn, Atlanta, GA 30332 USA
关键词
LOGISTICS; SYSTEM;
D O I
10.2514/1.A35122
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
This paper develops a hierarchical reinforcement learning architecture for multimission spaceflight campaign design under uncertainty, including vehicle design, infrastructure deployment planning, and space transportation scheduling. This problem involves a high-dimensional design space and is challenging especially with uncertainty present. To tackle this challenge, the developed framework has a hierarchical structure with reinforcement learning and network-based mixed-integer linear programming (MILP), where the former optimizes campaign-level decisions (e.g., design of the vehicle used throughout the campaign, destination demand assigned to each mission in the campaign), whereas the latter optimizes the detailed mission-level decisions (e.g., when to launch what from where to where). The framework is applied to a set of human lunar exploration campaign scenarios with uncertain in situ resource utilization performance as a case study. The main value of this work is its integration of the rapidly growing reinforcement learning research and the existing MILP-based space logistics methods through a hierarchical framework to handle the otherwise intractable complexity of space mission design under uncertainty. This unique framework is expected to be a critical steppingstone for the emerging research direction of artificial intelligence for space mission design.
引用
收藏
页码:421 / 433
页数:13
相关论文
共 50 条
  • [41] A deep reinforcement learning framework for solving two-stage stochastic programs
    Yilmaz, Dogacan
    Buyuktahtakin, I. Esra
    OPTIMIZATION LETTERS, 2024, 18 (09) : 1993 - 2020
  • [42] TASAC: A twin-actor reinforcement learning framework with a stochastic with an to batch control
    Joshi, Tanuja
    Kodamana, Hariprasad
    Kandath, Harikumar
    Kaisare, Niket
    CONTROL ENGINEERING PRACTICE, 2023, 134
  • [43] DeepBLOC: A Framework for Securing CPS through Deep Reinforcement Learning on Stochastic Games
    Tahsini, Alireza
    Dunstatter, Noah
    Guirguis, Mina
    Ahmed, Chuadhry Mujeeb
    2020 IEEE CONFERENCE ON COMMUNICATIONS AND NETWORK SECURITY (CNS), 2020,
  • [44] From deterministic to stochastic: an interpretable stochastic model-free reinforcement learning framework for portfolio optimization
    Song, Zitao
    Wang, Yining
    Qian, Pin
    Song, Sifan
    Coenen, Frans
    Jiang, Zhengyong
    Su, Jionglong
    APPLIED INTELLIGENCE, 2023, 53 (12) : 15188 - 15203
  • [45] From deterministic to stochastic: an interpretable stochastic model-free reinforcement learning framework for portfolio optimization
    Zitao Song
    Yining Wang
    Pin Qian
    Sifan Song
    Frans Coenen
    Zhengyong Jiang
    Jionglong Su
    Applied Intelligence, 2023, 53 : 15188 - 15203
  • [46] Reinforcement learning and stochastic optimisation
    Sebastian Jaimungal
    Finance and Stochastics, 2022, 26 : 103 - 129
  • [47] STOCHASTIC DYNAMICS OF REINFORCEMENT LEARNING
    BRESSLOFF, PC
    NETWORK-COMPUTATION IN NEURAL SYSTEMS, 1995, 6 (02) : 289 - 307
  • [48] Reinforcement learning and stochastic optimisation
    Jaimungal, Sebastian
    FINANCE AND STOCHASTICS, 2022, 26 (01) : 103 - 129
  • [49] A NEW APPROACH TO THE DESIGN OF REINFORCEMENT SCHEMES FOR LEARNING AUTOMATA - STOCHASTIC ESTIMATOR LEARNING ALGORITHMS
    PAPADIMITRIOU, GI
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 1994, 6 (04) : 649 - 654
  • [50] A NEW APPROACH TO THE DESIGN OF REINFORCEMENT SCHEMES FOR LEARNING AUTOMATA - STOCHASTIC ESTIMATOR LEARNING ALGORITHM
    VASILAKOS, AV
    PAPADIMITRIOU, GI
    NEUROCOMPUTING, 1995, 7 (03) : 275 - 297