Fast-maneuvering target seeking based on double-action Q-learning

被引:0
|
作者
Ngai, Daniel C. K. [1 ]
Yung, Nelson H. C. [1 ]
机构
[1] Univ Hong Kong, Dept Elect & Elect Engn, Hong Kong, Hong Kong, Peoples R China
来源
MACHINE LEARNING AND DATA MINING IN PATTERN RECOGNITION, PROCEEDINGS | 2007年 / 4571卷
关键词
moving object navigation; reinforcement learning; Q-learning;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a reinforcement learning method called DAQL is proposed to solve the problem of seeking and homing onto a fast maneuvering target, within the context of mobile robots. This Q-learning based method considers both target and obstacle actions when determining its own action decisions, which enables the agent to learn more effectively in a dynamically changing environment. It particularly suits fast-maneuvering target cases, in which maneuvers of the target are unknown a priori. Simulation result depicts that the proposed method is able to choose a less convoluted path to reach the target when compared to the ideal proportional navigation (IPN) method in handling fast maneuvering and randomly moving target. Furthermore, it can learn to adapt to the physical limitation of the system and do not require specific initial conditions to be satisfied for successful navigation towards the moving target.
引用
收藏
页码:653 / +
页数:3
相关论文
共 50 条
  • [21] Q-learning based on neural network in learning action selection of mobile robot
    Qiao, Junfei
    Hou, Zhanjun
    Ruan, Xiaogang
    2007 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION AND LOGISTICS, VOLS 1-6, 2007, : 263 - 267
  • [22] A Q-Learning Based Target Coverage Algorithm for Wireless Sensor Networks
    Xiong, Peng
    He, Dan
    Lu, Tiankun
    MATHEMATICS, 2025, 13 (03)
  • [23] Hexagon-based Q-learning to find a hidden target object
    Yoon, HU
    Sim, KB
    COMPUTATIONAL INTELLIGENCE AND SECURITY, PT 1, PROCEEDINGS, 2005, 3801 : 428 - 433
  • [24] Q-learning in continuous state and action spaces
    Gaskett, C
    Wettergreen, D
    Zelinsky, A
    ADVANCED TOPICS IN ARTIFICIAL INTELLIGENCE, 1999, 1747 : 417 - 428
  • [25] Variational quantum compiling with double Q-learning
    He, Zhimin
    Li, Lvzhou
    Zheng, Shenggen
    Li, Yongyao
    Situ, Haozhen
    NEW JOURNAL OF PHYSICS, 2021, 23 (03):
  • [26] Double Q-Learning for Radiation Source Detection
    Liu, Zheng
    Abbaszadeh, Shiva
    SENSORS, 2019, 19 (04)
  • [27] Expected Lenient Q-learning: a fast variant of the Lenient Q-learning algorithm for cooperative stochastic Markov games
    Amhraoui, Elmehdi
    Masrour, Tawfik
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (07) : 2781 - 2797
  • [28] Target transfer Q-learning and its convergence analysis
    Wang Y.
    Liu Y.
    Chen W.
    Ma Z.-M.
    Liu T.-Y.
    Neurocomputing, 2020, 392 : 11 - 22
  • [29] An Index Policy Based on Sarsa and Q-Learning for Heterogeneous Smart Target Tracking
    Hao, Yuhang
    Wang, Zengfu
    Fu, Jing
    Pan, Quan
    Yun, Tao
    IEEE SENSORS JOURNAL, 2024, 24 (21) : 36127 - 36142
  • [30] Adaptive Q-learning path planning algorithm based on virtual target guidance
    Li Z.
    Hu X.
    Zhang Y.
    Xu J.
    Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2024, 30 (02): : 553 - 568