Dynamic distributed constraint optimization using multi-agent reinforcement learning

被引:9
|
作者
Shokoohi, Maryam [1 ]
Afsharchi, Mohsen [2 ]
Shah-Hoseini, Hamed [1 ]
机构
[1] Islamic Azad Univ, Dept Mechan Elect & Comp, Sci & Res Branch, Tehran, Iran
[2] Univ Zanjan, Dept Elect & Comp, Engn, Zanjan, Iran
关键词
Dynamic distributed constraint optimization problem; Reinforcement learning; Multi-agent systems; Weapon target assignment; Markov decision process; WEAPON-TARGET ASSIGNMENT; MISSILE DEFENSE; ALLOCATION;
D O I
10.1007/s00500-022-06820-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An inherent difficulty in dynamic distributed constraint optimization problems (dynamic DCOP) is the uncertainty of future events when making an assignment at the current time. This dependency is not well addressed in the research community. This paper proposes a reinforcement-learning-based solver for dynamic distributed constraint optimization. We show that reinforcement learning techniques are an alternative approach to solve the given problem over time and are computationally more efficient than sequential DCOP solvers. We also use the novel heuristic to obtain the correct results and describe a formalism that has been adopted to model dynamic DCOPs with cooperative agents. We evaluate this approach in dynamic weapon target assignment (dynamic WTA) problem, via experimental results. We observe that the system dynamic WTA problem remains a safe zone after convergence while satisfying the constraints. Moreover, in the experiment we have implemented the agents that finally converge to the correct assignment.
引用
收藏
页码:3601 / 3629
页数:29
相关论文
共 50 条
  • [11] Cooperative Optimization Strategy for Distributed Energy Resource System using Multi-Agent Reinforcement Learning
    Liu, Zhaoyang
    Xiang, Tianchun
    Wang, Tianhao
    Mu, Chaoxu
    [J]. 2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
  • [12] Dynamic Variable Speed Limit Zones Allocation Using Distributed Multi-Agent Reinforcement Learning
    Kusic, Kresimir
    Ivanjko, Edouard
    Vrbanic, Filip
    Greguric, Martin
    Dusparic, Ivana
    [J]. 2021 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2021, : 3238 - 3245
  • [13] Multi-Agent Reinforcement Learning for Convex Optimization
    Morcos, Amir
    West, Aaron
    Maguire, Brian
    [J]. ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS III, 2021, 11746
  • [14] A Distributed Multi-Agent Dynamic Area Coverage Algorithm Based on Reinforcement Learning
    Xiao, Jian
    Wang, Gang
    Zhang, Ying
    Cheng, Lei
    [J]. IEEE ACCESS, 2020, 8 : 33511 - 33521
  • [15] Multi-Agent Reinforcement Learning-Based Distributed Dynamic Spectrum Access
    Albinsaid, Hasan
    Singh, Keshav
    Biswas, Sudip
    Li, Chih-Peng
    [J]. IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2022, 8 (02) : 1174 - 1185
  • [16] Cooperative Multi-Agent Systems Using Distributed Reinforcement Learning Techniques
    Zemzem, Wiem
    Tagina, Moncef
    [J]. KNOWLEDGE-BASED AND INTELLIGENT INFORMATION & ENGINEERING SYSTEMS (KES-2018), 2018, 126 : 517 - 526
  • [17] Distributed reinforcement learning in multi-agent decision systems
    Giráldez, JI
    Borrajo, D
    [J]. PROGRESS IN ARTIFICIAL INTELLIGENCE-IBERAMIA 98, 1998, 1484 : 148 - 159
  • [18] Distributed localization for IoT with multi-agent reinforcement learning
    Jia, Jie
    Yu, Ruoying
    Du, Zhenjun
    Chen, Jian
    Wang, Qinghu
    Wang, Xingwei
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (09): : 7227 - 7240
  • [19] Distributed Coordination Guidance in Multi-Agent Reinforcement Learning
    Lau, Qiangfeng Peter
    Lee, Mong Li
    Hsu, Wynne
    [J]. 2011 23RD IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2011), 2011, : 456 - 463
  • [20] Distributed localization for IoT with multi-agent reinforcement learning
    Jie Jia
    Ruoying Yu
    Zhenjun Du
    Jian Chen
    Qinghu Wang
    Xingwei Wang
    [J]. Neural Computing and Applications, 2022, 34 : 7227 - 7240