Path Planning Using Wasserstein Distributionally Robust Deep Q-learning

被引:0
|
作者
Alpturk, Cem [1 ]
Renganathan, Venkatraman [2 ]
机构
[1] Lund Univ, Dept Automat Control, LTH, Lund, Sweden
[2] Lund Univ, Dept Automat Control, LTH, Lund, Sweden
基金
欧洲研究理事会;
关键词
D O I
10.23919/ECC57647.2023.10178154
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We investigate the problem of risk averse robot path planning using the deep reinforcement learning and distributionally robust optimization perspectives. Our problem formulation involves modelling the robot as a stochastic linear dynamical system, assuming that a collection of process noise samples is available. We cast the risk averse motion planning problem as a Markov decision process and propose a continuous reward function design that explicitly takes into account the risk of collision with obstacles while encouraging the robot's motion towards the goal. We learn the risk-averse robot control actions through Lipschitz approximated Wasserstein distributionally robust deep Q-learning to hedge against the noise uncertainty. The learned control actions result in a safe and risk averse trajectory from the source to the goal, avoiding all the obstacles. Various supporting numerical simulations are presented to demonstrate our proposed approach.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Distributionally Robust Q-Learning
    Liu, Zijian
    Bai, Qinxun
    Blanchet, Jose
    Dong, Perry
    Xu, Wei
    Zhou, Zhengqing
    Zhou, Zhengyuan
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [2] Model based path planning using Q-Learning
    Sharma, Avinash
    Gupta, Kanika
    Kumar, Anirudha
    Sharma, Aishwarya
    Kumar, Rajesh
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON INDUSTRIAL TECHNOLOGY (ICIT), 2017, : 837 - 842
  • [3] Wasserstein distributionally robust shortest path problem
    Wang, Zhuolin
    You, Keyou
    Song, Shiji
    Zhang, Yuli
    [J]. EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2020, 284 (01) : 31 - 43
  • [4] Automatic Path Planning for Spraying Drones Based on Deep Q-Learning
    Huang, Ya-Yu
    Li, Zi-Wen
    Yang, Chun-Hao
    Huang, Yueh-Min
    [J]. JOURNAL OF INTERNET TECHNOLOGY, 2023, 24 (03): : 565 - 575
  • [5] A novel deep learning driven robot path planning strategy: Q-learning approach
    Hu, Junli
    [J]. INTERNATIONAL JOURNAL OF COMPUTER APPLICATIONS IN TECHNOLOGY, 2023, 71 (03) : 237 - 243
  • [6] Distributionally Safe Path Planning: Wasserstein Safe RRT
    Lathrop, Paul
    Boardman, Beth
    Martinez, Sonia
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (01) : 430 - 437
  • [7] Path planning of mobile robots with Q-learning
    Cetin, Halil
    Durdu, Akif
    [J]. 2014 22ND SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2014, : 2162 - 2165
  • [8] A Deep Q-learning based Path Planning and Navigation System for Firefighting Environments
    Bhattarai, Manish
    Martinez-Ramon, Manel
    [J]. ICAART: PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 2, 2021, : 267 - 277
  • [9] Cooperative Path Planning for Single Leader Using Q-learning Method
    Zhang, Lichuan
    Wu, Dongwei
    Ren, Ranzhen
    Xing, Runfa
    [J]. GLOBAL OCEANS 2020: SINGAPORE - U.S. GULF COAST, 2020,
  • [10] Path planning of UAV using guided enhancement Q-learning algorithm
    Zhou, Bin
    Guo, Yan
    Li, Ning
    Zhong, Xijian
    [J]. Hangkong Xuebao/Acta Aeronautica et Astronautica Sinica, 2021, 42 (09):