A Q-learning approach based on human reasoning for navigation in a dynamic environment

被引:7
|
作者
Yuan, Rupeng [1 ]
Zhang, Fuhai [1 ]
Wang, Yu [1 ]
Fu, Yili [1 ]
Wang, Shuguo [1 ]
机构
[1] Harbin Inst Technol, State Key Lab Robot & Syst, Harbin 150001, Heilongjiang, Peoples R China
基金
中国国家自然科学基金; 黑龙江省自然科学基金;
关键词
Autonomous navigation; Mobile robot; Dynamic environment; Q-learning; OBSTACLE AVOIDANCE; ROBOTS;
D O I
10.1017/S026357471800111X
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
A Q-learning approach is often used for navigation in static environments where state space is easy to define. In this paper, a new Q-learning approach is proposed for navigation in dynamic environments by imitating human reasoning. As a model-free method, a Q-learning method does not require the environmental model in advance. The state space and the reward function in the proposed approach are defined according to human perception and evaluation, respectively. Specifically, approximate regions instead of accurate measurements are used to define states. Moreover, due to the limitation of robot dynamics, actions for each state are calculated by introducing a dynamic window that takes robot dynamics into account. The conducted tests show that the obstacle avoidance rate of the proposed approach can reach 90.5% after training, and the robot can always operate below the dynamics limitation.
引用
收藏
页码:445 / 468
页数:24
相关论文
共 50 条
  • [21] Mounting of auction agent under dynamic environment by Q-learning and SARSA learning
    Katou, T
    Nagasaka, K
    [J]. 7TH WORLD MULTICONFERENCE ON SYSTEMICS, CYBERNETICS AND INFORMATICS, VOL V, PROCEEDINGS: COMPUTER SCIENCE AND ENGINEERING: I, 2003, : 472 - 475
  • [22] Dynamic scheduling with fuzzy clustering based Q-learning
    Wang, Guo-Lei
    Lin, Lin
    Zhong, Shi-Sheng
    [J]. Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2009, 15 (04): : 751 - 757
  • [23] A Deep Q-Learning Approach for Dynamic Management of Heterogeneous Processors
    Gupta, Ujjwal
    Mandal, Sumit K.
    Mao, Manqing
    Chakrabarti, Chaitali
    Ogras, Umit Y.
    [J]. IEEE COMPUTER ARCHITECTURE LETTERS, 2019, 18 (01) : 14 - 17
  • [24] Mobile Robot Navigation: Neural Q-Learning
    Yun, Soh Chin
    Parasuraman, S.
    Ganapathy, V.
    [J]. ADVANCES IN COMPUTING AND INFORMATION TECHNOLOGY, VOL 3, 2013, 178 : 259 - +
  • [25] Mobile robot navigation: neural Q-learning
    Parasuraman, S.
    Yun, Soh Chin
    [J]. INTERNATIONAL JOURNAL OF COMPUTER APPLICATIONS IN TECHNOLOGY, 2012, 44 (04) : 303 - 311
  • [26] Path Navigation For Indoor Robot With Q-Learning
    Huang, Lvwen
    He, Dongjian
    Zhang, Zhiyong
    Zhang, Peng
    [J]. INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2016, 22 (02): : 317 - 323
  • [27] Dynamic Pricing Decision for Perishable Goods: A Q-learning Approach
    Cheng, Yan
    [J]. 2008 4TH INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS, NETWORKING AND MOBILE COMPUTING, VOLS 1-31, 2008, : 11965 - 11969
  • [28] A Dynamic Grid-based Q-learning for Noise Covariance Adaptation in EKF and its Application in Navigation
    Dai, Xiang
    Fourati, Hassen
    Prieur, Christophe
    [J]. 2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 4984 - 4989
  • [29] Q-learning based Reinforcement Learning Approach for Lane Keeping
    Feher, Arpad
    Aradi, Szilard
    Becsi, Tamas
    [J]. 2018 18TH IEEE INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND INFORMATICS (CINTI), 2018, : 31 - 35
  • [30] A novel dynamic integration approach for multiple load forecasts based on Q-learning algorithm
    Ma, Minhua
    Jin, Bingjie
    Luo, Shuxin
    Guo, Shaoqing
    Huang, Hongwei
    [J]. INTERNATIONAL TRANSACTIONS ON ELECTRICAL ENERGY SYSTEMS, 2020, 30 (07):