Fast-maneuvering target seeking based on double-action Q-learning

被引:0
|
作者
Ngai, Daniel C. K. [1 ]
Yung, Nelson H. C. [1 ]
机构
[1] Univ Hong Kong, Dept Elect & Elect Engn, Hong Kong, Hong Kong, Peoples R China
来源
MACHINE LEARNING AND DATA MINING IN PATTERN RECOGNITION, PROCEEDINGS | 2007年 / 4571卷
关键词
moving object navigation; reinforcement learning; Q-learning;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a reinforcement learning method called DAQL is proposed to solve the problem of seeking and homing onto a fast maneuvering target, within the context of mobile robots. This Q-learning based method considers both target and obstacle actions when determining its own action decisions, which enables the agent to learn more effectively in a dynamically changing environment. It particularly suits fast-maneuvering target cases, in which maneuvers of the target are unknown a priori. Simulation result depicts that the proposed method is able to choose a less convoluted path to reach the target when compared to the ideal proportional navigation (IPN) method in handling fast maneuvering and randomly moving target. Furthermore, it can learn to adapt to the physical limitation of the system and do not require specific initial conditions to be satisfied for successful navigation towards the moving target.
引用
收藏
页码:653 / +
页数:3
相关论文
共 50 条
  • [1] Maneuvering Target Tracking Using Q-learning Based Kalman Filter
    Bekhtaoui, Z.
    Meche, A.
    Dahmani, M.
    Meraim, K. Abed
    2017 5TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING - BOUMERDES (ICEE-B), 2017,
  • [2] Infrared Fast-Maneuvering Target Tracking Based on Robust Exact Differentiator with Improved Particle Filter
    Zhang, Wanxin
    Huang, Bingrui
    Meng, Sijie
    Zhu, Jihong
    JOURNAL OF ROBOTICS, 2022, 2022
  • [3] Action Candidate Based Clipped Double Q-learning for Discrete and Continuous Action Tasks
    Jiang, Haobo
    Xie, Jin
    Yang, Jian
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 7979 - 7986
  • [4] Q-Learning with probability based action policy
    Ugurlu, Ekin Su
    Biricik, Goksel
    2006 IEEE 14TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS, VOLS 1 AND 2, 2006, : 210 - +
  • [5] Double Gumbel Q-Learning
    Hui, David Yu-Tung
    Courville, Aaron
    Bacon, Pierre-Luc
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [6] Weighted Double Q-learning
    Zhang, Zongzhang
    Pan, Zhiyuan
    Kochenderfer, Mykel J.
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3455 - 3461
  • [7] A novel double-mGBDT-based Q-learning
    Fu, Qiming
    Ma, Shuai
    Tian, Dawei
    Chen, JianPing
    Gao, Zhen
    Zhong, Shan
    INTERNATIONAL JOURNAL OF MODELLING IDENTIFICATION AND CONTROL, 2021, 37 (3-4) : 232 - 239
  • [8] Double action Q-learning for obstacle avoidance in a dynamically changing environment
    Ngai, DCK
    Yung, NHC
    2005 IEEE Intelligent Vehicles Symposium Proceedings, 2005, : 211 - 216
  • [9] Action Candidate Driven Clipped Double Q-Learning for Discrete and Continuous Action Tasks
    Jiang, Haobo
    Li, Guangyu
    Xie, Jin
    Yang, Jian
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (04) : 5269 - 5279
  • [10] Deep Reinforcement Learning with Double Q-Learning
    van Hasselt, Hado
    Guez, Arthur
    Silver, David
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2094 - 2100