Model based path planning using Q-Learning

被引:0
|
作者
Sharma, Avinash [1 ]
Gupta, Kanika [1 ]
Kumar, Anirudha [1 ]
Sharma, Aishwarya [1 ]
Kumar, Rajesh [1 ]
机构
[1] Malaviya Natl Inst Technol, Dept Elect Engn, Jaipur 302017, Rajasthan, India
关键词
Model Based Control; Q-learning; Reinforcement Learning; Neural Network; Grid-World; REINFORCEMENT; ALGORITHM;
D O I
暂无
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Though the classical robotics is highly proficient in accomplishing a lot of complex tasks, still it is far from exhibiting the human-like natural intelligence in terms of flexibility and reliability to work in dynamic scenarios. In order to render these qualities in the robots, reinforcement learning could prove to be quite effective. By employing learning based training provided by reinforcement learning methods, a robot can be made to learn to work in previously unforeseen situations. Still this learning task can be quite cumbersome due to its requirement of the huge amount of training data which makes the training quite inefficient in the real world scenarios. The paper proposes a model based path planning method using the epsilon greedy based Q-learning. The scenario was modeled using a grid-world based simulator which is being used in the initial training of the agent. The trained policy is then improved to learn the real world dynamics by using the real world samples. This study proves the efficiency and reliability of the simulator-based training methodology.
引用
收藏
页码:837 / 842
页数:6
相关论文
共 50 条
  • [1] Q-learning based Path Planning Method for UAVs using Priority Shifting
    de Carvalho, Kevin B.
    de Oliveira, Iure Rosa L.
    Villa, Daniel K. D.
    Caldeira, Alexandre G.
    Sarcinelli-Filho, Mario
    Brandao, Alexandre S.
    [J]. 2022 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS (ICUAS), 2022, : 421 - 426
  • [2] The Method Based on Q-Learning Path Planning in Migrating Workflow
    Xiao, Song
    Wang, Xiao-lin
    [J]. PROCEEDINGS 2013 INTERNATIONAL CONFERENCE ON MECHATRONIC SCIENCES, ELECTRIC ENGINEERING AND COMPUTER (MEC), 2013, : 2204 - 2208
  • [3] A Path Planning Algorithm for UAV Based on Improved Q-Learning
    Yan, Chao
    Xiang, Xiaojia
    [J]. 2018 2ND INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION SCIENCES (ICRAS), 2018, : 46 - 50
  • [4] Ship Local Path Planning Based on Improved Q-Learning
    Gong, Ming-Fan
    Xu, Hai-Xiang
    Feng, Hui
    Wang, Yong
    Xue, Xue-Hua
    [J]. Chuan Bo Li Xue/Journal of Ship Mechanics, 2022, 26 (06): : 824 - 833
  • [5] A Path Planning Algorithm for Space Manipulator Based on Q-Learning
    Li, Taiguo
    Li, Quanhong
    Li, Wenxi
    Xia, Jiagao
    Tang, Wenhua
    Wang, Weiwen
    [J]. PROCEEDINGS OF 2019 IEEE 8TH JOINT INTERNATIONAL INFORMATION TECHNOLOGY AND ARTIFICIAL INTELLIGENCE CONFERENCE (ITAIC 2019), 2019, : 1566 - 1571
  • [6] Mobile robot path planning based on Q-learning algorithm
    Li, Shaochuan
    Wang, Xuiqing
    Hu, Liwei
    Liu, Ying
    [J]. 2019 WORLD ROBOT CONFERENCE SYMPOSIUM ON ADVANCED ROBOTICS AND AUTOMATION (WRC SARA 2019), 2019, : 160 - 165
  • [7] Coverage Path Planning Optimization Based on Q-Learning Algorithm
    Piardi, Luis
    Lima, Jose
    Pereira, Ana, I
    Costa, Paulo
    [J]. INTERNATIONAL CONFERENCE ON NUMERICAL ANALYSIS AND APPLIED MATHEMATICS (ICNAAM-2018), 2019, 2116
  • [8] Path planning for autonomous mobile robot using transfer learning-based Q-learning
    Wu, Shengshuai
    Hu, Jinwen
    Zhao, Chunhui
    Pan, Quan
    [J]. PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), 2020, : 88 - 93
  • [9] Path planning of mobile robots with Q-learning
    Cetin, Halil
    Durdu, Akif
    [J]. 2014 22ND SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2014, : 2162 - 2165
  • [10] Cooperative Path Planning for Single Leader Using Q-learning Method
    Zhang, Lichuan
    Wu, Dongwei
    Ren, Ranzhen
    Xing, Runfa
    [J]. GLOBAL OCEANS 2020: SINGAPORE - U.S. GULF COAST, 2020,