Fast-maneuvering target seeking based on double-action Q-learning

被引:0
|
作者
Ngai, Daniel C. K. [1 ]
Yung, Nelson H. C. [1 ]
机构
[1] Univ Hong Kong, Dept Elect & Elect Engn, Hong Kong, Hong Kong, Peoples R China
来源
MACHINE LEARNING AND DATA MINING IN PATTERN RECOGNITION, PROCEEDINGS | 2007年 / 4571卷
关键词
moving object navigation; reinforcement learning; Q-learning;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a reinforcement learning method called DAQL is proposed to solve the problem of seeking and homing onto a fast maneuvering target, within the context of mobile robots. This Q-learning based method considers both target and obstacle actions when determining its own action decisions, which enables the agent to learn more effectively in a dynamically changing environment. It particularly suits fast-maneuvering target cases, in which maneuvers of the target are unknown a priori. Simulation result depicts that the proposed method is able to choose a less convoluted path to reach the target when compared to the ideal proportional navigation (IPN) method in handling fast maneuvering and randomly moving target. Furthermore, it can learn to adapt to the physical limitation of the system and do not require specific initial conditions to be satisfied for successful navigation towards the moving target.
引用
收藏
页码:653 / +
页数:3
相关论文
共 50 条
  • [31] Cooperative Q-Learning Based on Learning Automata
    Yang, Mao
    Tian, Yantao
    Qi, Xinyue
    2009 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION AND LOGISTICS ( ICAL 2009), VOLS 1-3, 2009, : 1972 - 1977
  • [32] A novel double-action actuator based on ferrofluid and permanent magnets
    Olaru, Radu
    Petrescu, Camelia
    Hertanu, Radu
    JOURNAL OF INTELLIGENT MATERIAL SYSTEMS AND STRUCTURES, 2012, 23 (14) : 1623 - 1630
  • [33] State and Action Space Segmentation Algorithm in Q-learning
    Notsu, Akira
    Ichihashi, Hidetomo
    Honda, Katsuhiro
    2008 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-8, 2008, : 2384 - 2389
  • [34] Accelerated Q-Learning for Fail State and Action Spaces
    Park, In-Won
    Kim, Jong-Hwan
    Park, Kui-Hong
    2008 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), VOLS 1-6, 2008, : 763 - +
  • [35] Fuzzy Q-learning in continuous state and action space
    Xu M.-L.
    Xu W.-B.
    Journal of China Universities of Posts and Telecommunications, 2010, 17 (04): : 100 - 109
  • [37] Double Deep Q-Learning Based Channel Estimation for Industrial Wireless Networks
    Bhardwaj, Sanjay
    Lee, Jae-Min
    Kim, Dong-Seong
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 1318 - 1320
  • [38] Q-LEARNING BASED THERAPY MODELING
    Jacak, Witold
    Proell, Karin
    EMSS 2009: 21ST EUROPEAN MODELING AND SIMULATION SYMPOSIUM, VOL II, 2009, : 204 - +
  • [39] Energy management strategy for hybrid electric vehicles based on double Q-learning
    Han, Lijin
    Yang, Ke
    Zhang, Xin
    Yang, Ningkang
    Liu, Hui
    Liu, Jiaxin
    INTERNATIONAL CONFERENCE ON MECHANICAL DESIGN AND SIMULATION (MDS 2022), 2022, 12261
  • [40] A Double Q-Learning Routing in Delay Tolerant Networks
    Yuan, Fan
    Wu, Jaogao
    Zhou, Hongyu
    Liu, Linfeng
    ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,