Analysis of Q-learning on ANNs for Robot Control using Live Video Feed

被引:0
|
作者
Murali, Nihal [1 ]
Gupta, Kunal [1 ]
Bhanot, Surekha [1 ]
机构
[1] BITS Pilani, Dept Elect & Elect Engn, Pilani Campus, Pilani 333031, Rajasthan, India
关键词
Artificial neural networks; Hardware implementation; Q-learning; Raw image inputs; Reinforcement learning; Robot learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Training of artificial neural networks (ANNs) using reinforcement learning (RL) techniques is being widely discussed in the robot learning literature. The high model complexity of ANNs along with the model-free nature of RL algorithms provides a desirable combination for many robotics applications. There is a huge need for algorithms that generalize using raw sensory inputs, such as vision, without any hand-engineered features or domain heuristics. In this paper, the standard control problem of line following robot was used as a test-bed, and an ANN controller for the robot was trained on images from a live video feed using Q-learning. A virtual agent was first trained in simulation environment and then deployed onto a robot's hardware. The robot successfully learns to traverse a wide range of curves and displays excellent generalization ability. Qualitative analysis of the evolution of policies, performance and weights of the network provide insights into the nature and convergence of the learning algorithm.
引用
收藏
页码:524 / 529
页数:6
相关论文
共 50 条
  • [41] Q-Learning for Autonomous Mobile Robot Obstacle Avoidance
    Ribeiro, Tiago
    Goncalves, Fernando
    Garcia, Ines
    Lopes, Gil
    Fernando Ribeiro, A.
    2019 19TH IEEE INTERNATIONAL CONFERENCE ON AUTONOMOUS ROBOT SYSTEMS AND COMPETITIONS (ICARSC 2019), 2019, : 243 - 249
  • [42] Neural Q-Learning Based Mobile Robot Navigation
    Yun, Soh Chin
    Parasuraman, S.
    Ganapathy, V.
    Joe, Halim Kusuma
    MATERIALS SCIENCE AND INFORMATION TECHNOLOGY, PTS 1-8, 2012, 433-440 : 721 - +
  • [43] Bias-Corrected Q-Learning to Control Max-Operator Bias in Q-Learning
    Lee, Donghun
    Defourny, Boris
    Powell, Warren B.
    PROCEEDINGS OF THE 2013 IEEE SYMPOSIUM ON ADAPTIVE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING (ADPRL), 2013, : 93 - 99
  • [44] Falling-down Avoidance Control for Acrobat Robot by Q-Learning with Function Approximaion
    Suzuki, H.
    Yamakita, M.
    Hirano, S.
    Lno, Z. W.
    2008 PROCEEDINGS OF SICE ANNUAL CONFERENCE, VOLS 1-7, 2008, : 801 - +
  • [45] Walking control of semi-passive robot via a modified Q-learning algorithm
    Sun, Zhongkui
    Zhou, Yining
    Xu, Wei
    Wang, Yuexin
    INTERNATIONAL JOURNAL OF NON-LINEAR MECHANICS, 2024, 161
  • [46] An Online Home Energy Management System using Q-Learning and Deep Q-Learning
    İzmitligil H.
    Karamancıoğlu A.
    Sustainable Computing: Informatics and Systems, 2024, 43
  • [47] Emergency-Response Locomotion of Hexapod Robot with Heuristic Reinforcement Learning Using Q-Learning
    Yang, Ming-Chieh
    Samani, Hooman
    Zhu, Kening
    INTERACTIVE COLLABORATIVE ROBOTICS (ICR 2019), 2019, 11659 : 320 - 329
  • [48] Path planning for autonomous mobile robot using transfer learning-based Q-learning
    Wu, Shengshuai
    Hu, Jinwen
    Zhao, Chunhui
    Pan, Quan
    PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), 2020, : 88 - 93
  • [49] Learning Robot Grasping from a Random Pile with Deep Q-Learning
    Chen, Bin
    Su, Jianhua
    Wang, Lili
    Gu, Qipeng
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2021, PT II, 2021, 13014 : 142 - 152
  • [50] Q-learning for Waiting Time Control in CDN/V2V Live streaming
    Ma, Zhejiayu
    Roubia, Soufiane
    Giroire, Frederic
    Urvoy-Keller, Guillaume
    2023 IFIP NETWORKING CONFERENCE, IFIP NETWORKING, 2023,