Unified reinforcement Q-learning for mean field game and control problems

被引:24
|
作者
Angiuli, Andrea [1 ]
Fouque, Jean-Pierre [1 ]
Lauriere, Mathieu [2 ]
机构
[1] Univ Calif Santa Barbara, Dept Stat & Appl Probabil, South Hall 5504, Santa Barbara, CA 93106 USA
[2] Princeton Univ, Dept Operat Res & Financial Engn, Princeton, NJ 08544 USA
关键词
Q-learning; Mean field game; Mean field control; Timescales; Linear-quadratic control;
D O I
10.1007/s00498-021-00310-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We present a Reinforcement Learning (RL) algorithm to solve infinite horizon asymptotic Mean Field Game (MFG) and Mean Field Control (MFC) problems. Our approach can be described as a unified two-timescale Mean Field Q-learning: The same algorithm can learn either the MFG or the MFC solution by simply tuning the ratio of two learning parameters. The algorithm is in discrete time and space where the agent not only provides an action to the environment but also a distribution of the state in order to take into account the mean field feature of the problem. Importantly, we assume that the agent cannot observe the population's distribution and needs to estimate it in a model-free manner. The asymptotic MFG and MFC problems are also presented in continuous time and space, and compared with classical (non-asymptotic or stationary) MFG and MFC problems. They lead to explicit solutions in the linear-quadratic (LQ) case that are used as benchmarks for the results of our algorithm.
引用
收藏
页码:217 / 271
页数:55
相关论文
共 50 条
  • [21] Mildly Conservative Q-Learning for Offline Reinforcement Learning
    Lyu, Jiafei
    Ma, Xiaoteng
    Li, Xiu
    Lu, Zongqing
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [22] Adaptable Conservative Q-Learning for Offline Reinforcement Learning
    Qiu, Lyn
    Li, Xu
    Liang, Lenghan
    Sun, Mingming
    Yan, Junchi
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT III, 2024, 14427 : 200 - 212
  • [23] Multi-agent dueling Q-learning with mean field and value decomposition
    Ding, Shifei
    Du, Wei
    Ding, Ling
    Guo, Lili
    Zhang, Jian
    An, Bo
    PATTERN RECOGNITION, 2023, 139
  • [24] Reinforcement Q-learning for Closed-loop Hypnosis Depth Control in Anesthesia
    Calvi, Giulia
    Manzoni, Eleonora
    Rampazzo, Mirco
    2022 30TH MEDITERRANEAN CONFERENCE ON CONTROL AND AUTOMATION (MED), 2022, : 164 - 169
  • [25] Deep Reinforcement Q-Learning for Intelligent Traffic Signal Control with Partial Detection
    Romain Ducrocq
    Nadir Farhi
    International Journal of Intelligent Transportation Systems Research, 2023, 21 : 192 - 206
  • [26] Deep Reinforcement Q-Learning for Intelligent Traffic Signal Control with Partial Detection
    Ducrocq, Romain
    Farhi, Nadir
    INTERNATIONAL JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS RESEARCH, 2023, 21 (01) : 192 - 206
  • [27] Reinforcement Q-Learning for PDF Tracking Control of Stochastic Systems with Unknown Dynamics
    Yang, Weiqing
    Zhou, Yuyang
    Zhang, Yong
    Ren, Yan
    MATHEMATICS, 2024, 12 (16)
  • [28] Ant Colony Optimization Incorporated With Fuzzy Q-Learning for Reinforcement Fuzzy Control
    Juang, Chia-Feng
    Lu, Chun-Ming
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART A-SYSTEMS AND HUMANS, 2009, 39 (03): : 597 - 608
  • [29] An Efficient Hardware Implementation of Reinforcement Learning: The Q-Learning Algorithm
    Spano, Sergio
    Cardarilli, Gian Carlo
    Di Nunzio, Luca
    Fazzolari, Rocco
    Giardino, Daniele
    Matta, Marco
    Nannarelli, Alberto
    Re, Marco
    IEEE ACCESS, 2019, 7 : 186340 - 186351
  • [30] Q-learning based Reinforcement Learning Approach for Lane Keeping
    Feher, Arpad
    Aradi, Szilard
    Becsi, Tamas
    2018 18TH IEEE INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND INFORMATICS (CINTI), 2018, : 31 - 35