Model based path planning using Q-Learning

被引:0
|
作者
Sharma, Avinash [1 ]
Gupta, Kanika [1 ]
Kumar, Anirudha [1 ]
Sharma, Aishwarya [1 ]
Kumar, Rajesh [1 ]
机构
[1] Malaviya Natl Inst Technol, Dept Elect Engn, Jaipur 302017, Rajasthan, India
关键词
Model Based Control; Q-learning; Reinforcement Learning; Neural Network; Grid-World; REINFORCEMENT; ALGORITHM;
D O I
暂无
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Though the classical robotics is highly proficient in accomplishing a lot of complex tasks, still it is far from exhibiting the human-like natural intelligence in terms of flexibility and reliability to work in dynamic scenarios. In order to render these qualities in the robots, reinforcement learning could prove to be quite effective. By employing learning based training provided by reinforcement learning methods, a robot can be made to learn to work in previously unforeseen situations. Still this learning task can be quite cumbersome due to its requirement of the huge amount of training data which makes the training quite inefficient in the real world scenarios. The paper proposes a model based path planning method using the epsilon greedy based Q-learning. The scenario was modeled using a grid-world based simulator which is being used in the initial training of the agent. The trained policy is then improved to learn the real world dynamics by using the real world samples. This study proves the efficiency and reliability of the simulator-based training methodology.
引用
收藏
页码:837 / 842
页数:6
相关论文
共 50 条
  • [41] A path planning approach for unmanned surface vehicles based on dynamic and fast Q-learning
    Hao, Bing
    Du, He
    Yan, Zheping
    [J]. OCEAN ENGINEERING, 2023, 270
  • [42] Application of artificial neural network based on Q-learning for mobile robot path planning
    Li, Caihong
    Zhang, Jingyuan
    Li, Yibin
    [J]. 2006 IEEE INTERNATIONAL CONFERENCE ON INFORMATION ACQUISITION, VOLS 1 AND 2, CONFERENCE PROCEEDINGS, 2006, : 978 - 982
  • [43] Optimal path planning method based on epsilon-greedy Q-learning algorithm
    Bulut, Vahide
    [J]. JOURNAL OF THE BRAZILIAN SOCIETY OF MECHANICAL SCIENCES AND ENGINEERING, 2022, 44 (03)
  • [44] RSMDP-BASED ROBUST Q-LEARNING FOR OPTIMAL PATH PLANNING IN A DYNAMIC ENVIRONMENT
    Zhang, Yunfei
    Li, Weilin
    de Silva, Clarence W.
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2016, 31 (04): : 290 - 300
  • [45] UAV path planning algorithm based on Deep Q-Learning to search for a lost in the ocean
    Boulares, Mehrez
    Fehri, Afef
    Jemni, Mohamed
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2024, 179
  • [46] RSMDP-based robust Q-learning for optimal path planning in a dynamic environment
    Zhang, Yunfei
    Li, Weilin
    De Silva, Clarence W.
    [J]. International Journal of Robotics and Automation, 2016, 31 (04) : 290 - 300
  • [47] Car-Like Robot Path Planning Based on Voronoi and Q-Learning Algorithms
    Alhassow, Mustafa Mohammed
    Ata, Oguz
    Atilla, Dogu Cagdas
    [J]. 2021 7TH INTERNATIONAL CONFERENCE ON ENGINEERING AND EMERGING TECHNOLOGIES (ICEET 2021), 2021, : 591 - 594
  • [48] The Optimization of Path Planning for Multi-robot System using Boltzmann Policy based Q-Learning Algorithm
    Wang, Zeying
    Shi, Zhiguo
    Li, Yuankai
    Tu, Jun
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO), 2013, : 1199 - 1204
  • [49] An Autonomous Path Finding Robot Using Q-Learning
    Babu, Madhu
    Krishna, Vamshi U.
    Shahensha, S. K.
    [J]. PROCEEDINGS OF THE 10TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS AND CONTROL (ISCO'16), 2016,
  • [50] An optimized Q-Learning algorithm for mobile robot local path planning
    Zhou, Qian
    Lian, Yang
    Wu, Jiayang
    Zhu, Mengyue
    Wang, Haiyong
    Cao, Jinli
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 286