Path Planning Using Wasserstein Distributionally Robust Deep Q-learning

被引:0
|
作者
Alpturk, Cem [1 ]
Renganathan, Venkatraman [2 ]
机构
[1] Lund Univ, Dept Automat Control, LTH, Lund, Sweden
[2] Lund Univ, Dept Automat Control, LTH, Lund, Sweden
基金
欧洲研究理事会;
关键词
D O I
10.23919/ECC57647.2023.10178154
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We investigate the problem of risk averse robot path planning using the deep reinforcement learning and distributionally robust optimization perspectives. Our problem formulation involves modelling the robot as a stochastic linear dynamical system, assuming that a collection of process noise samples is available. We cast the risk averse motion planning problem as a Markov decision process and propose a continuous reward function design that explicitly takes into account the risk of collision with obstacles while encouraging the robot's motion towards the goal. We learn the risk-averse robot control actions through Lipschitz approximated Wasserstein distributionally robust deep Q-learning to hedge against the noise uncertainty. The learned control actions result in a safe and risk averse trajectory from the source to the goal, avoiding all the obstacles. Various supporting numerical simulations are presented to demonstrate our proposed approach.
引用
收藏
页数:8
相关论文
共 50 条
  • [41] An Online Home Energy Management System using Q-Learning and Deep Q-Learning
    İzmitligil, Hasan
    Karamancıoğlu, Abdurrahman
    [J]. Sustainable Computing: Informatics and Systems, 2024, 43
  • [42] Path planning of a mobile robot in a free-space environment using Q-learning
    Jianxun Jiang
    Jianbin Xin
    [J]. Progress in Artificial Intelligence, 2019, 8 : 133 - 142
  • [43] Path planning of a mobile robot in a free-space environment using Q-learning
    Jiang, Jianxun
    Xin, Jianbin
    [J]. PROGRESS IN ARTIFICIAL INTELLIGENCE, 2019, 8 (01) : 133 - 142
  • [44] High-Level Path Planning for an Autonomous Sailboat Robot Using Q-Learning
    da Silva Junior, Andouglas Goncalves
    dos Santos, Davi Henrique
    Fernandes de Negreiros, Alvaro Pinto
    Boas de Souza Silva, Joao Moreno Vilas
    Garcia Goncalves, Luiz Marcos
    [J]. SENSORS, 2020, 20 (06)
  • [45] Making Deep Q-learning Methods Robust to Time Discretization
    Tallec, Corentin
    Blier, Leonard
    Ollivier, Yann
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [46] An Optimized Path Planning Method for Container Ships in Bohai Bay Based on Improved Deep Q-Learning
    Gao, Xuanyu
    Dong, Yitao
    Han, Yi
    [J]. IEEE ACCESS, 2023, 11 : 91275 - 91292
  • [47] A DEEP Q-LEARNING NETWORK FOR SHIP STOWAGE PLANNING PROBLEM
    Shen, Yifan
    Zhao, Ning
    Xia, Mengjue
    Du, Xueqiang
    [J]. POLISH MARITIME RESEARCH, 2017, 24 : 102 - 109
  • [48] An Autonomous Path Finding Robot Using Q-Learning
    Babu, Madhu
    Krishna, Vamshi U.
    Shahensha, S. K.
    [J]. PROCEEDINGS OF THE 10TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS AND CONTROL (ISCO'16), 2016,
  • [49] Hybrid Path Planning Algorithm of the Mobile Agent Based on Q-Learning
    Gao, Tengteng
    Li, Caihong
    Liu, Guoming
    Guo, Na
    Wang, Di
    Li, Yongdi
    [J]. AUTOMATIC CONTROL AND COMPUTER SCIENCES, 2022, 56 (02) : 130 - 142
  • [50] Indoor Emergency Path Planning Based on the Q-Learning Optimization Algorithm
    Xu, Shenghua
    Gu, Yang
    Li, Xiaoyan
    Chen, Cai
    Hu, Yingyi
    Sang, Yu
    Jiang, Wenxing
    [J]. ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION, 2022, 11 (01)