Stochastic learning in multi-agent optimization: Communication and payoff-based approaches

被引:10
|
作者
Tatarenko, Tatiana [1 ]
机构
[1] Tech Univ Darmstadt, Control Methods & Robot Lab, Darmstadt, Germany
关键词
Non-convex optimization; Multi-agent systems; Game theory; Learning algorithms; Stochastic approximation; DISTRIBUTED OPTIMIZATION; POWER-CONTROL; CONSENSUS; GAMES;
D O I
10.1016/j.automatica.2018.10.001
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Game theory serves as a powerful tool for distributed optimization in multi-agent systems in different applications. In this paper we consider multi-agent systems that can be modeled by means of potential games whose potential function coincides with a global objective function to be maximized. In this approach, the agents correspond to the strategic decision makers and the optimization problem is equivalent to the problem of learning a potential function maximizer in the designed game. The paper deals with two different information settings in the system. Firstly, we consider systems, where agents have the access to the gradient of their utility functions. However, they do not possess the full information about the joint actions. Thus, to be able to move along the gradient toward a local optimum, they need to exchange the information with their neighbors by means of communication. The second setting refers to a payoff based approach. Here, we assume that at each iteration agents can only observe their own played actions and experienced payoffs. In both cases, the paper studies unconstrained non-concave optimization with a differentiable objective function. To develop the corresponding algorithms guaranteeing convergence to a local maximum of the potential function in absence of saddle points, we utilize the idea of the well-known Robbins-Monro procedure based on the theory of stochastic approximation. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1 / 12
页数:12
相关论文
共 50 条
  • [41] Communication-robust multi-agent learning by adaptable auxiliary multi-agent adversary generation
    Yuan, Lei
    Chen, Feng
    Zhang, Zongzhang
    Yu, Yang
    [J]. FRONTIERS OF COMPUTER SCIENCE, 2024, 18 (06)
  • [42] Collaborative optimization of multi-microgrids system with shared energy storage based on multi-agent stochastic game and reinforcement learning
    Wang, Yijian
    Cui, Yang
    Li, Yang
    Xu, Yang
    [J]. ENERGY, 2023, 280
  • [43] Partition-based multi-agent optimization in the presence of lossy and asynchronous communication
    Todescato, Marco
    Bof, Nicoletta
    Cavraro, Guido
    Carli, Ruggero
    Schenato, Luca
    [J]. AUTOMATICA, 2020, 111
  • [44] Cloud-Based Centralized/Decentralized Multi-Agent Optimization with Communication Delays
    Hale, Matthew T.
    Nedic, Angelia
    Egerstedt, Magnus
    [J]. 2015 54TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2015, : 700 - 705
  • [45] Targeted Multi-Agent Communication with Deep Metric Learning
    Miao, Hua
    Yu, Nanxiang
    [J]. ENGINEERING LETTERS, 2023, 31 (02) : 712 - 723
  • [46] Multi-Agent Deep Reinforcement Learning with Emergent Communication
    Simoes, David
    Lau, Nuno
    Reis, Luis Paulo
    [J]. 2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [47] Learning Individually Inferred Communication for Multi-Agent Cooperation
    Ding, Ziluo
    Huang, Tiejun
    Lu, Zongqing
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [48] Multi-Agent Path Finding with Prioritized Communication Learning
    Li, Wenhao
    Chen, Hongjun
    Jin, Bo
    Tan, Wenzhe
    Zha, Hongyuan
    Wang, Xiangfeng
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2022, 2022, : 10695 - 10701
  • [49] Multi-Agent Reinforcement Learning for Coordinating Communication and Control
    Mason, Federico
    Chiariotti, Federico
    Zanella, Andrea
    Popovski, Petar
    [J]. IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2024, 10 (04) : 1566 - 1581
  • [50] Improving coordination with communication in multi-agent reinforcement learning
    Szer, D
    Charpillet, F
    [J]. ICTAI 2004: 16TH IEEE INTERNATIONALCONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2004, : 436 - 440