Analysis of Q-learning on ANNs for Robot Control using Live Video Feed

被引:0
|
作者
Murali, Nihal [1 ]
Gupta, Kunal [1 ]
Bhanot, Surekha [1 ]
机构
[1] BITS Pilani, Dept Elect & Elect Engn, Pilani Campus, Pilani 333031, Rajasthan, India
关键词
Artificial neural networks; Hardware implementation; Q-learning; Raw image inputs; Reinforcement learning; Robot learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Training of artificial neural networks (ANNs) using reinforcement learning (RL) techniques is being widely discussed in the robot learning literature. The high model complexity of ANNs along with the model-free nature of RL algorithms provides a desirable combination for many robotics applications. There is a huge need for algorithms that generalize using raw sensory inputs, such as vision, without any hand-engineered features or domain heuristics. In this paper, the standard control problem of line following robot was used as a test-bed, and an ANN controller for the robot was trained on images from a live video feed using Q-learning. A virtual agent was first trained in simulation environment and then deployed onto a robot's hardware. The robot successfully learns to traverse a wide range of curves and displays excellent generalization ability. Qualitative analysis of the evolution of policies, performance and weights of the network provide insights into the nature and convergence of the learning algorithm.
引用
收藏
页码:524 / 529
页数:6
相关论文
共 50 条
  • [1] MOTION CONTROL OF A ROBOT BY MEANS OF Q-LEARNING USING THE EXAMPLE OF LOCOMOTION
    Bussmann, Tobias
    Schilberg, Daniel
    PROCEEDINGS OF ASME 2023 INTERNATIONAL MECHANICAL ENGINEERING CONGRESS AND EXPOSITION, IMECE2023, VOL 3, 2023,
  • [2] Robot behavioral selection using Q-learning
    Martinson, E
    Stoytchev, A
    Arkin, R
    2002 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-3, PROCEEDINGS, 2002, : 970 - 977
  • [3] Balance Control of Robot With CMAC Based Q-learning
    Li Ming-ai
    Jiao Li-fang
    Qiao Jun-fei
    Ruan Xiao-gang
    2008 CHINESE CONTROL AND DECISION CONFERENCE, VOLS 1-11, 2008, : 2668 - 2672
  • [4] Autonomous reconfiguration of robot shape by using Q-learning
    Shiba, Satoshi
    Uchida, Masafumi
    Nozawa, Akio
    Asano, Hirotoshi
    Onogaki, Hitoshi
    Mizuno, Tota
    Ide, Hideto
    Yokoyama, Syuichi
    ARTIFICIAL LIFE AND ROBOTICS, 2009, 14 (02) : 213 - 218
  • [5] Autonomous Warehouse Robot using Deep Q-Learning
    Peyas, Ismot Sadik
    Hasan, Zahid
    Tushar, Md Rafat Rahman
    Al Musabbir
    Azni, Raisa Mehjabin
    Siddique, Shahnewaz
    2021 IEEE REGION 10 CONFERENCE (TENCON 2021), 2021, : 857 - 862
  • [6] An Autonomous Path Finding Robot Using Q-Learning
    Babu, Madhu
    Krishna, Vamshi U.
    Shahensha, S. K.
    PROCEEDINGS OF THE 10TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS AND CONTROL (ISCO'16), 2016,
  • [7] Autonomous Exploration for Mobile Robot using Q-learning
    Liu, Yang
    Liu, Huaping
    Wang, Bowen
    2017 2ND INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM), 2017, : 614 - 619
  • [8] Mobile robot navigation using neural Q-learning
    Yang, GS
    Chen, EK
    An, CW
    PROCEEDINGS OF THE 2004 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2004, : 48 - 52
  • [9] Hybrid control for robot navigation - A hierarchical Q-learning algorithm
    Chen, Chunlin
    Li, Han-Xiong
    Dong, Daoyi
    IEEE ROBOTICS & AUTOMATION MAGAZINE, 2008, 15 (02) : 37 - 47
  • [10] Control of the trajectory of a hexapod robot based on distributed Q-learning
    Youcef, Z
    Pierre, C
    PROCEEDINGS OF THE IEEE-ISIE 2004, VOLS 1 AND 2, 2004, : 277 - 282