Accelerating Learning in Constructive Predictive Frameworks with the Successor Representation

被引:0
|
作者
Sherstan, Craig [1 ]
Machado, Marlos C. [1 ]
Pilarski, Patrick M. [1 ,2 ]
机构
[1] Univ Alberta, Edmonton, AB, Canada
[2] DeepMind, London, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose using the Successor Representation (SR) to accelerate learning in a constructive knowledge system based on General Value Functions (GVFs). In real-world settings, like robotics for unstructured and dynamic environments, it is impossible to model all meaningful aspects of a system and its environment by hand. Instead, robots must learn and adapt to changes in their environment and task, incrementally constructing models from their own experience. GVFs, taken from the field of reinforcement learning (RL), are a way of modeling the world as predictive questions. One approach to such models proposes a massive network of interconnected and interdependent GVFs, which are incrementally added over time. It is reasonable to expect that new, incrementally added predictions can be learned more swiftly if the learning process leverages knowledge gained from past experience. The SR provides a means of capturing regularities that can be reused across multiple GVFs by separating the dynamics of the world from the prediction targets. As a primary contribution of this work, we show that using the SR can improve sample efficiency and learning speed of GVFs in a continual learning setting where new predictions are incrementally added and learned over time. We analyze our approach in a grid-world and then demonstrate its potential on data from a physical robot arm.
引用
收藏
页码:2997 / 3003
页数:7
相关论文
共 50 条
  • [1] Associative Learning of an Unnormalized Successor Representation
    Verosky, Niels J.
    [J]. NEURAL COMPUTATION, 2024, 36 (07) : 1410 - 1423
  • [2] The successor representation in human reinforcement learning
    I. Momennejad
    E. M. Russek
    J. H. Cheong
    M. M. Botvinick
    N. D. Daw
    S. J. Gershman
    [J]. Nature Human Behaviour, 2017, 1 : 680 - 692
  • [3] The successor representation in human reinforcement learning
    Momennejad, I.
    Russek, E. M.
    Cheong, J. H.
    Botvinick, M. M.
    Daw, N. D.
    Gershman, S. J.
    [J]. NATURE HUMAN BEHAVIOUR, 2017, 1 (09): : 680 - 692
  • [4] Temporal Abstraction in Reinforcement Learning with the Successor Representation
    Machado, Marlos C.
    Barreto, Andre
    Precup, Doina
    Bowling, Michael
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [5] Representation Learning for Constructive Comments Classification
    Uribe, Diego
    Cuan, Enrique
    Urquizo, Elisa
    [J]. 2020 INTERNATIONAL CONFERENCE ON MECHATRONICS, ELECTRONICS AND AUTOMOTIVE ENGINEERING (ICMEAE 2020), 2020, : 71 - 75
  • [7] A Probabilistic Successor Representation for Context-Dependent Learning
    Geerts, Jesse P. P.
    Gershman, Samuel J. J.
    Burgess, Neil
    Stachenfeld, Kimberly L. L.
    [J]. PSYCHOLOGICAL REVIEW, 2023, : 578 - 597
  • [8] Accelerating Imitation Learning with Predictive Models
    Cheng, Ching-An
    Yan, Xinyan
    Theodorou, Evangelos A.
    Boots, Byron
    [J]. 22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [9] Accelerating Deep Learning Frameworks with Micro-batches
    Oyama, Yosuke
    Ben-Nun, Tal
    Hoefler, Torsten
    Matsuoka, Satoshi
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON CLUSTER COMPUTING (CLUSTER), 2018, : 402 - 412
  • [10] A Deep Reinforcement Learning Approach to Marginalized Importance Sampling with the Successor Representation
    Fujimoto, Scott
    Meger, David
    Precup, Doina
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139