A Q-learning approach based on human reasoning for navigation in a dynamic environment

被引:7
|
作者
Yuan, Rupeng [1 ]
Zhang, Fuhai [1 ]
Wang, Yu [1 ]
Fu, Yili [1 ]
Wang, Shuguo [1 ]
机构
[1] Harbin Inst Technol, State Key Lab Robot & Syst, Harbin 150001, Heilongjiang, Peoples R China
基金
中国国家自然科学基金; 黑龙江省自然科学基金;
关键词
Autonomous navigation; Mobile robot; Dynamic environment; Q-learning; OBSTACLE AVOIDANCE; ROBOTS;
D O I
10.1017/S026357471800111X
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
A Q-learning approach is often used for navigation in static environments where state space is easy to define. In this paper, a new Q-learning approach is proposed for navigation in dynamic environments by imitating human reasoning. As a model-free method, a Q-learning method does not require the environmental model in advance. The state space and the reward function in the proposed approach are defined according to human perception and evaluation, respectively. Specifically, approximate regions instead of accurate measurements are used to define states. Moreover, due to the limitation of robot dynamics, actions for each state are calculated by introducing a dynamic window that takes robot dynamics into account. The conducted tests show that the obstacle avoidance rate of the proposed approach can reach 90.5% after training, and the robot can always operate below the dynamics limitation.
引用
收藏
页码:445 / 468
页数:24
相关论文
共 50 条
  • [31] A path planning approach for unmanned surface vehicles based on dynamic and fast Q-learning
    Hao, Bing
    Du, He
    Yan, Zheping
    [J]. OCEAN ENGINEERING, 2023, 270
  • [32] Cyclic error correction based Q-learning for mobile robots navigation
    Rongkuan Tang
    Hongliang Yuan
    [J]. International Journal of Control, Automation and Systems, 2017, 15 : 1790 - 1798
  • [33] A Multiagent Dynamic Assessment Approach for Water Quality Based on Improved Q-Learning Algorithm
    Ni, Jianjun
    Ren, Li
    Liu, Minghua
    Zhu, Daqi
    [J]. MATHEMATICAL PROBLEMS IN ENGINEERING, 2013, 2013
  • [34] Q-learning based univector field navigation method for mobile robots
    Vien, Ngo Anh
    Viet, Nguyen Hoang
    Park, HyunJeong
    Lee, SeungGwan
    Chung, TaeChoong
    [J]. ADVANCES AND INNOVATIONS IN SYSTEMS, COMPUTING SCIENCES AND SOFTWARE ENGINEERING, 2007, : 463 - +
  • [35] Cyclic Error Correction based Q-learning for Mobile Robots Navigation
    Tang, Rongkuan
    Yuan, Hongliang
    [J]. INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2017, 15 (04) : 1790 - 1798
  • [36] DQDWA: Dynamic Weight Coefficients Based on Q-learning for Dynamic Window Approach Considering Environmental Situations
    Kobayashi, Masato
    Zushi, Hiroka
    Nakamura, Tomoaki
    Motoi, Naoki
    [J]. 2023 IEEE/ASME INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS, AIM, 2023, : 1141 - 1146
  • [37] A novel contour extraction approach based on Q-learning
    Liang, Jun-Bin
    Xu, Jian-Min
    [J]. PROCEEDINGS OF 2006 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2006, : 3807 - +
  • [38] Adaptive packet scheduling in IoT environment based on Q-learning
    Kim, Donghyun
    Lee, Taeho
    Kim, Sejun
    Lee, Byungjun
    Youn, Hee Yong
    [J]. JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2020, 11 (06) : 2225 - 2235
  • [39] Q-Learning based SFC deployment on Edge Computing Environment
    Pandey, Suman
    Hong, James Won-Ki
    Yoo, Jae-Hyoung
    [J]. APNOMS 2020: 2020 21ST ASIA-PACIFIC NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM (APNOMS), 2020, : 220 - 226
  • [40] Adaptive packet scheduling in IoT environment based on Q-learning
    Donghyun Kim
    Taeho Lee
    Sejun Kim
    Byungjun Lee
    Hee Yong Youn
    [J]. Journal of Ambient Intelligence and Humanized Computing, 2020, 11 : 2225 - 2235