Driving force planning in shield tunneling based on Markov decision processes

被引:0
|
作者
HU XiangTao
机构
关键词
shield tunneling; Markov decision process; automatic deviation rectifying; interval arithmetic; driving force planning;
D O I
暂无
中图分类号
U455.43 [盾构法(全断面开挖)];
学科分类号
0814 ; 081406 ;
摘要
In shield tunneling, the control system needs very reliable capability of deviation rectifying in order to ensure that the tunnel trajectory meets the permissible criterion. To this goal, we present an approach that adopts Markov decision process (MDP) theory to plan the driving force with explicit representation of the uncertainty during excavation. The shield attitudes of possi- ble world and driving forces during excavation are scattered as a state set and an action set, respectively. In particular, an evaluation function is proposed with consideration of the stability of driving force and the deviation of shield attitude. Unlike the deterministic approach, the driving forces based on MDP model lead to an uncertain effect and the attitude is known only with an imprecise probability. We consider the case that the transition probability varies in a given domain estimated by field data, and discuss the optimal policy based on the interval arithmetic. The validity of the approach is discussed by comparing the driving force planning with the actual operating data from the field records of Line 9 in Tianjin. It is proved that the MDP model is reasonable enough to predict the driving force for automatic deviation rectifying.
引用
收藏
页码:1022 / 1030
页数:9
相关论文
共 50 条
  • [21] Learning and Planning in Average-Reward Markov Decision Processes
    Wan, Yi
    Naik, Abhishek
    Sutton, Richard S.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139 : 7665 - 7676
  • [22] Planning in entropy-regularized Markov decision processes and games
    Grill, Jean-Bastien
    Domingues, Omar D.
    Menard, Pierre
    Munos, Remi
    Valko, Michal
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [23] Ground Delay Program Planning Using Markov Decision Processes
    Cox, Jonathan
    Kochenderfer, Mykel J.
    JOURNAL OF AEROSPACE INFORMATION SYSTEMS, 2016, 13 (03): : 134 - 142
  • [24] Simple Regret Optimization in Online Planning for Markov Decision Processes
    Feldman, Zohar
    Domshlak, Carmel
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2014, 51 : 165 - 205
  • [25] Probabilistic Logic Markov Decision Processes for Modeling Driving Behaviors in Self-driving Cars
    Aviles, Hector
    Negrete, Marco
    Machucho, Ruben
    Rivera, Karelly
    Trejo, David
    Vargas, Hector
    ADVANCES IN ARTIFICIAL INTELLIGENCE-IBERAMIA 2022, 2022, 13788 : 366 - 377
  • [26] Generating Explanations Based on Markov Decision Processes
    Elizalde, Francisco
    Sucar, Enrique
    Noguez, Julieta
    Reyes, Alberto
    MICAI 2009: ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2009, 5845 : 51 - +
  • [27] Markov decision processes
    White, D.J.
    Journal of the Operational Research Society, 1995, 46 (06):
  • [28] Markov Decision Processes
    Bäuerle N.
    Rieder U.
    Jahresbericht der Deutschen Mathematiker-Vereinigung, 2010, 112 (4) : 217 - 243
  • [29] Unifying nondeterministic and probabilistic planning through imprecise Markov Decision Processes
    Trevizan, Felipe W.
    Cozman, Fabio G.
    de Barros, Leliane N.
    ADVANCES IN ARTIFICIAL INTELLIGENCE - IBERAMIA-SBIA 2006, PROCEEDINGS, 2006, 4140 : 502 - 511
  • [30] Robust motion planning using Markov Decision Processes and quadtree decomposition
    Burlet, J
    Aycard, O
    Fraichard, T
    2004 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1- 5, PROCEEDINGS, 2004, : 2820 - 2825