Constructivist Approach to State Space Adaptation in Reinforcement Learning

被引:4
|
作者
Guerian, Maxime [1 ]
Cardozo, Nicolas [2 ]
Dusparic, Ivana [1 ]
机构
[1] Trinity Coll Dublin, Sch Comp Sci & Stat, Dublin, Ireland
[2] Univ Los Andes, Syst & Comp Engn Dept, Bogota, Colombia
基金
爱尔兰科学基金会;
关键词
dynamic adaptation; reinforcement learning;
D O I
10.1109/SASO.2019.00016
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement Learning (RL) is increasingly used to achieve adaptive behaviours in Internet of Things systems relying on large amounts of sensor data. To address the need for self-adaptation in such environments, techniques for detecting environment changes and re-learning behaviours appropriate to those changes have been proposed. However, with the heterogeneity of sensor inputs, the problem of self-adaptation permeates one level deeper; in order for the learnt behaviour to adapt, the underlying environment representation needs to adapt first. The granularity of the RL state space might need to be adapted to learn more efficiently, or to match the new granularity of input data. This paper proposes an implementation of Constructivist RL (Con-RL), enabling RL to learn and continuously adapt its state space representations. We propose a Multi-Layer Growing Neural Gas (ML-GNG) technique, as an extension of the GNG clustering algorithm, to autonomously learn suitable state spaces based on sensor data and learnt actions at runtime. We also create and continuously update a repository of state spaces, selecting the most appropriate one to use at each time step. We evaluate Con-RL in two scenarios: the canonical RL mountain car single-agent scenario, and a large-scale multi-agent car and ride-sharing scenario. We demonstrate its ability to adapt to new sensor inputs, to increase the speed of learning through state space optimization, and to maintain stable long-term performance.
引用
收藏
页码:52 / 61
页数:10
相关论文
共 50 条
  • [1] Budgeted Reinforcement Learning in Continuous State Space
    Carrara, Nicolas
    Leurent, Edouard
    Laroche, Romain
    Urvoy, Tanguy
    Maillard, Odalric-Ambrym
    Pietquin, Olivier
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] Adaptive state space formation in reinforcement learning
    Samejima, K
    Omori, T
    [J]. ICONIP'98: THE FIFTH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING JOINTLY WITH JNNS'98: THE 1998 ANNUAL CONFERENCE OF THE JAPANESE NEURAL NETWORK SOCIETY - PROCEEDINGS, VOLS 1-3, 1998, : 251 - 255
  • [3] Adaptive state space partitioning for reinforcement learning
    Lee, ISK
    Lau, HYK
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2004, 17 (06) : 577 - 588
  • [4] A reinforcement learning accelerated by state space reduction
    Senda, K
    Mano, S
    Fujii, S
    [J]. SICE 2003 ANNUAL CONFERENCE, VOLS 1-3, 2003, : 1992 - 1997
  • [5] A state space filter for reinforcement learning in POMDPs - Application to a continuous state space -
    Nagayoshi, Masato
    Murao, Hajime
    Tamaki, Hisashi
    [J]. 2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 3098 - +
  • [6] Reinforcement learning approach to locomotion adaptation in sloped environments
    Andre, Joao
    Costa, Lino
    Santos, Cristina
    [J]. 2014 IEEE INTERNATIONAL CONFERENCE ON AUTONOMOUS ROBOT SYSTEMS AND COMPETITIONS (ICARSC), 2014, : 164 - 169
  • [7] Adaptation Resources in Virtual Learning Environments under Constructivist Approach: A Systematic Review
    Bremgartner, Vitor
    de Magalhaes Netto, J. F.
    de Menezes, Credine Silva
    [J]. FRONTIERS IN EDUCATION CONFERENCE (FIE), 2015, 2015, : 1784 - 1791
  • [8] Reduction of state space on reinforcement learning by sensor selection
    Kishima, Yasutaka
    Kurashige, Kentarou
    [J]. 2012 INTERNATIONAL SYMPOSIUM ON MICRO-NANOMECHATRONICS AND HUMAN SCIENCE (MHS), 2012, : 138 - 143
  • [9] Reduction of state space in reinforcement learning by sensor selection
    Kishima, Yasutaka
    Kurashige, Kentarou
    [J]. ARTIFICIAL LIFE AND ROBOTICS, 2013, 18 (1-2) : 7 - 14
  • [10] A reinforcement learning approach for quantum state engineering
    Mackeprang, Jelena
    Dasari, Durga B. Rao
    Wrachtrup, Jorg
    [J]. QUANTUM MACHINE INTELLIGENCE, 2020, 2 (01)