Learning intelligent behavior in a non-stationary and partially observable environment

被引:2
|
作者
Senkul, S [1 ]
Polat, F [1 ]
机构
[1] Middle E Tech Univ, Dept Comp Engn, TR-06531 Ankara, Turkey
关键词
agent learning; multi-agent systems; Q-learning; reinforcement learning;
D O I
10.1023/A:1019935502139
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Individual learning in an environment where more than one agent exist is a challenging task. In this paper, a single learning agent situated in an environment where multiple agents exist is modeled based on reinforcement learning. The environment is non-stationary and partially accessible from an agents' point of view. Therefore, learning activities of an agent is influenced by actions of other cooperative or competitive agents in the environment. A prey-hunter capture game that has the above characteristics is defined and experimented to simulate the learning process of individual agents. Experimental results show that there are no strict rules for reinforcement learning. We suggest two new methods to improve the performance of agents. These methods decrease the number of states while keeping as much state as necessary.
引用
收藏
页码:97 / 115
页数:19
相关论文
共 50 条
  • [1] Learning Intelligent Behavior in a Non-stationary and Partially Observable Environment
    SelÇuk şenkul
    Faruk Polat
    [J]. Artificial Intelligence Review, 2002, 18 : 97 - 115
  • [2] Learning Contextual Bandits in a Non-stationary Environment
    Wu, Qingyun
    Iyer, Naveen
    Wang, Hongning
    [J]. ACM/SIGIR PROCEEDINGS 2018, 2018, : 495 - 504
  • [3] The Parzen Kernel Approach to Learning in Non-stationary Environment
    Pietruczuk, Lena
    Rutkowski, Leszek
    Jaworski, Maciej
    Duda, Piotr
    [J]. PROCEEDINGS OF THE 2014 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2014, : 3319 - 3323
  • [4] Bilevel Online Deep Learning in Non-stationary Environment
    Han, Ya-nan
    Liu, Jian-wei
    Xiao, Bing-biao
    Wang, Xin-Tan
    Luo, Xiong-lin
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT II, 2021, 12892 : 347 - 358
  • [5] Learning Optimal Behavior in Environments with Non-stationary Observations
    Boone, Ilio
    Rens, Gavin
    [J]. ICAART: PROCEEDINGS OF THE 14TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 3, 2022, : 729 - 736
  • [6] Bargaining in a non-stationary environment
    Coles, MG
    Muthoo, A
    [J]. JOURNAL OF ECONOMIC THEORY, 2003, 109 (01) : 70 - 89
  • [7] Distributed recurrent self-organization for tracking the state of non-stationary partially observable dynamical systems
    Khouzam, Bassem
    Frezza-Buet, Herve
    [J]. BIOLOGICALLY INSPIRED COGNITIVE ARCHITECTURES, 2013, 3 : 87 - 104
  • [8] Models of Forecasting of Enterprise's Behavior in Non-Stationary External Environment
    Rayevnyeva, Olena
    Touzani, Tarik
    [J]. ESTUDIOS DE ECONOMIA APLICADA, 2020, 38 (04):
  • [9] Cascading Non-Stationary Bandits: Online Learning to Rank in the Non-Stationary Cascade Model
    Li, Chang
    de Rijke, Maarten
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2859 - 2865
  • [10] INVARIANT STRUCTURE IN NON-STATIONARY BEHAVIOR
    TREVINO, G
    [J]. JOURNAL OF SOUND AND VIBRATION, 1988, 125 (03) : 503 - 510