Anticipatory Learning Classifier Systems and Factored Reinforcement Learning

被引:0
|
作者
Sigaud, Olivier [1 ]
Butz, Martin V. [4 ]
Kozlova, Olga [1 ,2 ]
Meyer, Christophe [3 ]
机构
[1] Univ Paris 06, Inst Syst Intelligents & Robot, CNRS, UMR 7222, 4 Pl Jussieu, F-75005 Paris, France
[2] Thales Secur Solut & Serv Simulat, F-95523 Cergy Pontoise, France
[3] Thales Secur Solut & Serv, TrereSIS Res & Innovat Off, F-91767 Palaiseau, France
[4] Univ Wurzburg, D-97070 Wurzburg, Germany
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Factored Reinforcement Learning (FRL) is a new technique to solve Factored Markov Decision Problems (FMDPs) when the structure of the problem is not known in advance. Like Anticipatory Learning Classifier Systems (ALCSs), it is a model-based Reinforcement Learning approach that includes generalization mechanisms in the presence of a structured domain. In general, FRL and ALCSs are explicit, state-anticipatory approaches that learn generalized state transition models to improve system behavior based on model-based reinforcement learning techniques. In this contribution, we highlight the conceptual similarities and differences between FRL and ALCSs, focusing on the one hand on SPITI, an instance Of FRL method, and on ALCSs, MACS and XACS, on the other hand. Though FRL systems seem to benefit from a clearer theoretical grounding, an empirical comparison between SPITI and XACS On two benchmark problems reveals that the latter scales much better than the former when some combination of state variables do not occur. Based on this finding, we discuss the mechanisms in XACS that result in the better scalability and propose importing these mechanisms into FRL systems.
引用
收藏
页码:321 / +
页数:4
相关论文
共 50 条
  • [1] Learning classifier systems from a reinforcement learning perspective
    P. L. Lanzi
    [J]. Soft Computing, 2002, 6 (3) : 162 - 170
  • [2] Efficient reinforcement learning in factored MDPs
    Kearns, M
    Koller, D
    [J]. IJCAI-99: PROCEEDINGS OF THE SIXTEENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOLS 1 & 2, 1999, : 740 - 747
  • [3] Bayesian Reinforcement Learning in Factored POMDPs
    Katt, Sammie
    Oliehoek, Frans A.
    Amato, Christopher
    [J]. AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 7 - 15
  • [4] Reinforcement learning with factored states and actions
    Sallans, B
    Hinton, GE
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2004, 5 : 1063 - 1088
  • [5] Accurate and Interpretable Representations of Environments with Anticipatory Learning Classifier Systems
    Orhand, Romain
    Jeannin-Girardon, Anne
    Parrend, Pierre
    Collet, Pierre
    [J]. GENETIC PROGRAMMING (EUROGP 2022), 2022, : 245 - 261
  • [6] Pittsburgh Learning Classifier Systems for Explainable Reinforcement Learning: Comparing with XCS
    Bishop, Jordan T.
    Gallagher, Marcus
    Browne, Will N.
    [J]. PROCEEDINGS OF THE 2022 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE (GECCO'22), 2022, : 323 - 331
  • [7] Reinforcement Learning in Continuous Spaces by Using Learning Fuzzy Classifier Systems
    Chen, Gang
    Douch, Colin
    Zhang, Mengjie
    Pang, Shaoning
    [J]. NEURAL INFORMATION PROCESSING, PT II, 2015, 9490 : 320 - 328
  • [8] Two classifier systems for reinforcement learning of motion patterns
    Yamada, K
    Svinin, M
    Ohkura, K
    Hosoe, S
    Ueda, K
    [J]. MOBILE ROBOT TECHNOLOGY, PROCEEDINGS, 2001, : 1 - 6
  • [9] TeXDYNA: Hierarchical Reinforcement Learning in Factored MDPs
    Kozlova, Olga
    Sigaud, Olivier
    Meyer, Christophe
    [J]. FROM ANIMALS TO ANIMATS 11, 2010, 6226 : 489 - +
  • [10] Memory Efficient Factored Abstraction for Reinforcement Learning
    Sahin, Coskun
    Cilden, Erkin
    Polat, Faruk
    [J]. 2015 IEEE 2ND INTERNATIONAL CONFERENCE ON CYBERNETICS (CYBCONF), 2015, : 18 - 23