Q-learning and robotics

被引:0
|
作者
Touzet, CF
Santos, JM
机构
关键词
Artificial Neural Networks; learning; robotics;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Because it allows the synthesis of behaviors despite the absence of a robot-world interaction model, Q-learning has become the most used learning algorithm for autonomous robotics in applications such as obstacle avoidance, wall following, go-to-the-nest, etc. This is mostly due to neural-based implementations such as multilayer perceptrons trained with backpropagation, or self-organizing maps. Such implementations provide an efficient generalization, i.e., fast learning, and designate the critic - the reinforcement function definition - as the real issue.
引用
收藏
页码:685 / 689
页数:5
相关论文
共 50 条
  • [1] Deep Q-Learning in Robotics: Improvement of Accuracy and Repeatability
    Sumanas, Marius
    Petronis, Algirdas
    Bucinskas, Vytautas
    Dzedzickis, Andrius
    Virzonis, Darius
    Morkvenaite-Vilkonciene, Inga
    [J]. SENSORS, 2022, 22 (10)
  • [2] Q-Learning with Double Progressive Widening: Application to Robotics
    Sokolovska, Nataliya
    Teytaud, Olivier
    Milone, Mario
    [J]. NEURAL INFORMATION PROCESSING, PT III, 2011, 7064 : 103 - +
  • [3] Q-learning with a growing RBF network for behavior learning in mobile robotics
    Li, J
    Duckett, T
    [J]. PROCEEDINGS OF THE SIXTH IASTED INTERNATIONAL CONFERENCE ON ROBOTICS AND APPLICATIONS, 2005, : 273 - 278
  • [4] LEGO© MINDSTORMS NXT AND Q-LEARNING: A TEACHING APPROACH FOR ROBOTICS IN ENGINEERING
    Martinez-Tenor, A.
    Fernandez-Madrigal, J. A.
    Cruz-Martin, A.
    [J]. ICERI2014: 7TH INTERNATIONAL CONFERENCE OF EDUCATION, RESEARCH AND INNOVATION, 2014, : 4836 - 4845
  • [5] Bootstrapping Q-Learning for Robotics From Neuro-Evolution Results
    Zimmer, Matthieu
    Doncieux, Stephane
    [J]. IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2018, 10 (01) : 102 - 119
  • [6] Q-LEARNING
    WATKINS, CJCH
    DAYAN, P
    [J]. MACHINE LEARNING, 1992, 8 (3-4) : 279 - 292
  • [7] Deep Reinforcement Learning: From Q-Learning to Deep Q-Learning
    Tan, Fuxiao
    Yan, Pengfei
    Guan, Xinping
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2017), PT IV, 2017, 10637 : 475 - 483
  • [8] Backward Q-learning: The combination of Sarsa algorithm and Q-learning
    Wang, Yin-Hao
    Li, Tzuu-Hseng S.
    Lin, Chih-Jui
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2013, 26 (09) : 2184 - 2193
  • [9] Constrained Deep Q-Learning Gradually Approaching Ordinary Q-Learning
    Ohnishi, Shota
    Uchibe, Eiji
    Yamaguchi, Yotaro
    Nakanishi, Kosuke
    Yasui, Yuji
    Ishii, Shin
    [J]. FRONTIERS IN NEUROROBOTICS, 2019, 13
  • [10] Learning rates for Q-Learning
    Even-Dar, E
    Mansour, Y
    [J]. COMPUTATIONAL LEARNING THEORY, PROCEEDINGS, 2001, 2111 : 589 - 604