Double action Q-learning for obstacle avoidance in a dynamically changing environment

被引:0
|
作者
Ngai, DCK [1 ]
Yung, NHC [1 ]
机构
[1] Univ Hong Kong, Dept Elect & Elect Engn, Hong Kong, Hong Kong, Peoples R China
关键词
Q-learning; reinforcement learning; temporal differences; obstacle avoidance;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a new method for solving the reinforcement learning problem in a dynamically changing environment, as in vehicle navigation, in which the Markov Decision Process used in traditional reinforcement learning is modified so that the response of the environment is taken into consideration for determining the agent's next state. This is achieved by changing the action-value function to handle three parameters at a time, namely, the current state, action taken by the agent, and action taken by the environment. As it considers the actions by the agent and environment, it is termed "Double Action". Based on the Q-learning method, the proposed method is implemented and the update rule is modified to handle all of the three parameters. Preliminary results show that the proposed method has the sum of rewards (negative) 89.5% less than that of the traditional method. Apart form that, our new method also has the total number of collisions and mean steps used in one episode 89.5% and 15.5% lower than that of the traditional method respectively.
引用
收藏
页码:211 / 216
页数:6
相关论文
共 50 条
  • [1] Performance evaluation of Double Action Q-learning in moving obstacle avoidance problem
    Ngai, DCK
    Yung, NHC
    INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS, VOL 1-4, PROCEEDINGS, 2005, : 865 - 870
  • [2] Fuzzy Q-learning obstacle avoidance algorithm of humanoid robot in unknown environment
    Wen, Shuhuan
    Chen, Jianhua
    Li, Zhen
    Rad, Ahmad B.
    Othman, Kamal Mohammed
    2018 37TH CHINESE CONTROL CONFERENCE (CCC), 2018, : 5186 - 5190
  • [3] Q-Learning for Autonomous Mobile Robot Obstacle Avoidance
    Ribeiro, Tiago
    Goncalves, Fernando
    Garcia, Ines
    Lopes, Gil
    Fernando Ribeiro, A.
    2019 19TH IEEE INTERNATIONAL CONFERENCE ON AUTONOMOUS ROBOT SYSTEMS AND COMPETITIONS (ICARSC 2019), 2019, : 243 - 249
  • [4] Obstacle Avoidance of Hexapod Robots Using Fuzzy Q-Learning
    Hong, Jun
    Tang, Kaiqiang
    Chen, Chunlin
    2017 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2017, : 1262 - 1267
  • [5] NAO robot obstacle avoidance based on fuzzy Q-learning
    Wen, Shuhuan
    Hu, Xueheng
    Li, Zhen
    Lam, Hak Keung
    Sun, Fuchun
    Fang, Bin
    INDUSTRIAL ROBOT-THE INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH AND APPLICATION, 2020, 47 (06): : 801 - 811
  • [6] Double Deep Q-Learning and Faster R-CNN-Based Autonomous Vehicle Navigation and Obstacle Avoidance in Dynamic Environment
    Bin Issa, Razin
    Das, Modhumonty
    Rahman, Md. Saferi
    Barua, Monika
    Rhaman, Md. Khalilur
    Ripon, Kazi Shah Nawaz
    Alam, Md. Golam Rabiul
    SENSORS, 2021, 21 (04) : 1 - 24
  • [7] Q-Learning of Bee-Like Robots through Obstacle Avoidance
    Rasheed, Jawairia
    Irfan, Haroon
    2024 12TH INTERNATIONAL CONFERENCE ON CONTROL, MECHATRONICS AND AUTOMATION, ICCMA, 2024, : 166 - 170
  • [8] Improved Q-Learning Applied to Dynamic Obstacle Avoidance and Path Planning
    Wang, Chunlei
    Yang, Xiao
    Li, He
    IEEE ACCESS, 2022, 10 : 92879 - 92888
  • [9] Obstacle Avoidance for AUV by Q-Learning based Guidance Vector Field
    Wu, Keqiao
    Yao, Peng
    PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), 2020, : 702 - 707
  • [10] Autonomous quadrotor obstacle avoidance based on dueling double deep recurrent Q-learning with monocular vision
    Ou, Jiajun
    Guo, Xiao
    Zhu, Ming
    Lou, Wenjie
    NEUROCOMPUTING, 2021, 441 : 300 - 310