Compositional Models for Reinforcement Learning

被引:0
|
作者
Jong, Nicholas K. [1 ]
Stone, Peter [1 ]
机构
[1] Univ Texas Austin, Univ Stn C0500 1, Austin, TX 78712 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Innovations such as optimistic exploration, function approximation, and hierarchical decomposition have helped scale, reinforcement learning to more complex environments, but these three ideas have rarely been studied together. This paper develops a unified framework that formalizes these algorithmic contributions as, operators on learned models of the environment. Our formalism reveals some synergies among these innovations, and it, suggests a straight forward way to compose them. The resulting algorithm, Fitted R-MAXQ, is the first to combine the function approximation of fitted algorithms, the efficient; model-based exploration of R-MAX, and the hierarchical decompostion of MAXQ.
引用
收藏
页码:644 / 659
页数:16
相关论文
共 50 条
  • [1] Compositional Transfer in Hierarchical Reinforcement Learning
    Wulfmeier, Markus
    Abdolmaleki, Abbas
    Hafner, Roland
    Springenberg, Jost Tobias
    Neunert, Michael
    Hertweck, Tim
    Lampe, Thomas
    Siegel, Noah
    Heess, Nicolas
    Riedmiller, Martin
    [J]. ROBOTICS: SCIENCE AND SYSTEMS XVI, 2020,
  • [2] COMPOSUITE: A COMPOSITIONAL REINFORCEMENT LEARNING BENCHMARK
    Mendez, Jorge A.
    Hussing, Marcel
    Gummadi, Meghna
    Eaton, Eric
    [J]. CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 199, 2022, 199
  • [3] Compositional Reinforcement Learning from Logical Specifications
    Jothimurugan, Kishor
    Bansal, Suguman
    Bastani, Osbert
    Alur, Rajeev
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [4] Learning compositional categorization models
    Ommer, B
    Buhmann, JM
    [J]. COMPUTER VISION - ECCV 2006, PT 3, PROCEEDINGS, 2006, 3953 : 316 - 329
  • [5] Bounding the Optimal Value Function in Compositional Reinforcement Learning
    Adamczyk, Jacob
    Makarenko, Volodymyr
    Arriojas, Argenis
    Tiomkin, Stas
    Kulkarni, Rahul V.
    [J]. UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 22 - 32
  • [6] Compositional design of multicomponent alloys using reinforcement learning
    Xian, Yuehui
    Dang, Pengfei
    Tian, Yuan
    Jiang, Xue
    Zhou, Yumei
    Ding, Xiangdong
    Sun, Jun
    Lookman, Turab
    Xue, Dezhen
    [J]. Acta Materialia, 2024, 274
  • [7] Inference and Learning with Hierarchical Compositional Models
    Kokkinos, Iasonas
    Yuille, Alan
    [J]. 2009 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPR WORKSHOPS 2009), VOLS 1 AND 2, 2009, : 533 - 533
  • [8] Learning Compositional Sparse Bimodal Models
    Kumar, Suren
    Dhiman, Vikas
    Koch, Parker A.
    Corso, Jason J.
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (05) : 1032 - 1044
  • [9] Environment Generation for Zero-Shot Compositional Reinforcement Learning
    Gur, Izzeddin
    Jaques, Natasha
    Miao, Yingjie
    Choi, Jongwook
    Tiwari, Manoj
    Lee, Honglak
    Faust, Aleksandra
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [10] Integrating Reinforcement Learning with Models of Representation Learning
    Jones, Matt
    Canas, Fabian
    [J]. COGNITION IN FLUX, 2010, : 1258 - 1263