Model-Free Non-Stationarity Detection and Adaptation in Reinforcement Learning

被引:7
|
作者
Canonaco, Giuseppe [1 ]
Restelli, Marcello [1 ]
Roveri, Manuel [1 ]
机构
[1] Politecn Milan, Milan, Italy
关键词
D O I
10.3233/FAIA200200
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In most Reinforcement Learning (RL) studies, the considered task is assumed to be stationary, i.e., it does not change its behavior or its characteristics over time, as this allows to generate all the convergence properties of RL techniques. Unfortunately, this assumption does not hold in real-world scenarios where systems and environments typically evolve over time. For instance, in robotic applications, sensor or actuator faults would induce a sudden change in the RL settings, while in financial applications the evolution of the market can cause a more gradual variation over time. In this paper, we present an adaptive RL algorithm able to detect changes in the environment or in the reward function and react to these changes by adapting to the new conditions of the task. At first, we develop a figure of merit onto which a hypothesis test can be applied to detect changes between two different learning iterations. Then, we extended this test to sequentially operate over time by means of the CUmulative SUM (CUSUM) approach. Finally, the proposed changedetection mechanism is combined (following an adaptive-active approach) with a well known RL algorithm to make it able to deal with non-stationary tasks. We tested the proposed algorithm on two well-known continuous-control tasks to check its effectiveness in terms of non-stationarity detection and adaptation over a vanilla RL algorithm.
引用
下载
收藏
页码:1047 / 1054
页数:8
相关论文
共 50 条
  • [1] Non-stationarity Detection in Model-Free Reinforcement Learning via Value Function Monitoring
    Hussein, Maryem
    Keshk, Marwa
    Hussein, Aya
    ADVANCES IN ARTIFICIAL INTELLIGENCE, AI 2023, PT II, 2024, 14472 : 350 - 362
  • [2] Covariance matrix adaptation for model-free reinforcement learning
    Adaptation de la matrice de covariance pour l'apprentissage par renforcement direct
    2013, Lavoisier, 14 rue de Provigny, Cachan Cedex, F-94236, France (27)
  • [3] Non-stationarity in multiagent reinforcement learning in electricity market simulation
    Renshaw-Whitman, Charles
    Zobernig, Viktor
    Cremer, Jochen L.
    de Vries, Laurens
    ELECTRIC POWER SYSTEMS RESEARCH, 2024, 235
  • [4] Deep Reinforcement Learning amidst Continual Structured Non-Stationarity
    Xie, Annie
    Harrison, James
    Finn, Chelsea
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [5] REACTIVE EXPLORATION TO COPE WITH NON-STATIONARITY IN LIFELONG REINFORCEMENT LEARNING
    Steinparz, Christian
    Schmied, Thomas
    Paischer, Fabian
    Dinu, Marius-Constantin
    Patil, Vihang
    Bitto-Nemling, Angela
    Eghbal-zadeh, Hamid
    Hochreiter, Sepp
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 199, 2022, 199
  • [6] Meta-Reinforcement Learning by Tracking Task Non-stationarity
    Poiani, Riccardo
    Tirinzoni, Andrea
    Restelli, Marcello
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2899 - 2905
  • [7] Non-stationarity detection in natural images
    Raj, Raghu G.
    Bovik, Alan C.
    Geisler, Wilson S.
    2007 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, VOLS 1-7, 2007, : 1433 - 1436
  • [8] Learning to Optimize under Non-Stationarity
    Cheung, Wang Chi
    Simchi-Levi, David
    Zhu, Ruihao
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [9] Quantifying the impact of non-stationarity in reinforcement learning-based traffic signal control
    Alegre, Lucas N.
    Bazzan, Ana L. C.
    da Silva, Bruno C.
    PEERJ COMPUTER SCIENCE, 2021,
  • [10] Detection of Man-in-the-Middle Attacks in Model-Free Reinforcement Learning
    Rani, Rishi
    Franceschetti, Massimo
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211