Emergence of Prediction by Reinforcement Learning Using a Recurrent Neural Network

被引:5
|
作者
Goto, Kenta [1 ]
Shibata, Katsunari [1 ]
机构
[1] Oita Univ, Dept Elect & Elect Engn, 700 Dannoharu, Oita 8701192, Japan
关键词
D O I
10.1155/2010/437654
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
To develop a robot that behaves flexibly in the real world, it is essential that it learns various necessary functions autonomously without receiving significant information from a human in advance. Among such functions, this paper focuses on learning "prediction" that is attracting attention recently from the viewpoint of autonomous learning. The authors point out that it is important to acquire through learning not only the way of predicting future information, but also the purposive extraction of prediction target from sensor signals. It is suggested that through reinforcement learning using a recurrent neural network, both emerge purposively and simultaneously without testing individually whether or not each piece of information is predictable. In a task where an agent gets a reward when it catches amoving object that can possibly become invisible, it was observed that the agent learned to detect the necessary factors of the object velocity before it disappeared, to relay the information among some hidden neurons, and finally to catch the object at an appropriate position and timing, considering the effects of bounces off a wall after the object became invisible.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Emergence of Higher Exploration in Reinforcement Learning Using a Chaotic Neural Network
    Goto, Yuki
    Shibata, Katsunari
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2016, PT I, 2016, 9947 : 40 - 48
  • [2] Category learning in a recurrent neural network with reinforcement learning
    Zhang, Ying
    Pan, Xiaochuan
    Wang, Yihong
    [J]. FRONTIERS IN PSYCHIATRY, 2022, 13
  • [3] Recurrent Neural Network and Reinforcement Learning Model for COVID-19 Prediction
    Kumar, R. Lakshmana
    Khan, Firoz
    Din, Sadia
    Band, Shahab S.
    Mosavi, Amir
    Ibeke, Ebuka
    [J]. FRONTIERS IN PUBLIC HEALTH, 2021, 9
  • [4] Reinforcement learning of recurrent neural network for temporal coding
    Kimura, Daichi
    Hayakawa, Yoshinori
    [J]. NEUROCOMPUTING, 2008, 71 (16-18) : 3379 - 3386
  • [5] Emergence of Multi-step Discrete State Transition through Reinforcement Learning with a Recurrent Neural Network
    Samsudin, Mohamad Faizal
    Sawatsubashi, Yoshito
    Shibata, Katsunari
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2012, PT II, 2012, 7664 : 583 - 590
  • [6] A model to explain the emergence of reward expectancy neurons using reinforcement learning and neural network
    Ishii, Shinya
    Shidara, Munetaka
    Shibata, Katsunari
    [J]. NEUROCOMPUTING, 2006, 69 (10-12) : 1327 - 1331
  • [7] A reinforcement learning unit matching recurrent neural network for the state trend prediction of rolling bearings
    Li, Feng
    Chen, Yong
    Wang, Jiaxu
    Zhou, Xueming
    Tang, Baoping
    [J]. MEASUREMENT, 2019, 145 : 191 - 203
  • [8] Car Sales Prediction Using Gated Recurrent Units Neural Networks with Reinforcement Learning
    Zhu, Bowen
    Dong, Huailong
    Zhang, Jing
    [J]. INTELLIGENCE SCIENCE AND BIG DATA ENGINEERING: BIG DATA AND MACHINE LEARNING, PT II, 2019, 11936 : 312 - 324
  • [9] A recurrent control neural network for data efficient reinforcement learning
    Schaefer, Anton Maximilian
    Udluft, Steffen
    Zimmermann, Hans-Georg
    [J]. 2007 IEEE INTERNATIONAL SYMPOSIUM ON APPROXIMATE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING, 2007, : 151 - +
  • [10] Emergence of Color Constancy Illusion through Reinforcement Learning with a Neural Network
    Shibata, Katsunari
    Kurizaki, Shunsuke
    [J]. 2012 IEEE INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING AND EPIGENETIC ROBOTICS (ICDL), 2012,