Stochastic learning in multi-agent optimization: Communication and payoff-based approaches

被引:10
|
作者
Tatarenko, Tatiana [1 ]
机构
[1] Tech Univ Darmstadt, Control Methods & Robot Lab, Darmstadt, Germany
关键词
Non-convex optimization; Multi-agent systems; Game theory; Learning algorithms; Stochastic approximation; DISTRIBUTED OPTIMIZATION; POWER-CONTROL; CONSENSUS; GAMES;
D O I
10.1016/j.automatica.2018.10.001
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Game theory serves as a powerful tool for distributed optimization in multi-agent systems in different applications. In this paper we consider multi-agent systems that can be modeled by means of potential games whose potential function coincides with a global objective function to be maximized. In this approach, the agents correspond to the strategic decision makers and the optimization problem is equivalent to the problem of learning a potential function maximizer in the designed game. The paper deals with two different information settings in the system. Firstly, we consider systems, where agents have the access to the gradient of their utility functions. However, they do not possess the full information about the joint actions. Thus, to be able to move along the gradient toward a local optimum, they need to exchange the information with their neighbors by means of communication. The second setting refers to a payoff based approach. Here, we assume that at each iteration agents can only observe their own played actions and experienced payoffs. In both cases, the paper studies unconstrained non-concave optimization with a differentiable objective function. To develop the corresponding algorithms guaranteeing convergence to a local maximum of the potential function in absence of saddle points, we utilize the idea of the well-known Robbins-Monro procedure based on the theory of stochastic approximation. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1 / 12
页数:12
相关论文
共 50 条
  • [1] Stochastic Payoff-Based Learning in Multi-Agent Systems Modeled by Means of Potential Games
    Tatarenko, Tatiana
    [J]. 2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 5298 - 5303
  • [2] A simple payoff-based learning classifier system
    Bull, L
    [J]. PARALLEL PROBLEM SOLVING FROM NATURE - PPSN VIII, 2004, 3242 : 1032 - 1041
  • [3] Payoff-based communication adaptation based on network service availability
    Kravets, R
    Calvert, K
    Schwan, K
    [J]. IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA COMPUTING AND SYSTEMS, PROCEEDINGS, 1998, : 33 - 42
  • [4] Multi-agent reinforcement learning based on local communication
    Zhang, Wenxu
    Ma, Lei
    Li, Xiaonan
    [J]. CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2019, 22 (Suppl 6): : 15357 - 15366
  • [5] Multi-agent reinforcement learning based on local communication
    Wenxu Zhang
    Lei Ma
    Xiaonan Li
    [J]. Cluster Computing, 2019, 22 : 15357 - 15366
  • [6] Multi-Agent Reinforcement Learning With Distributed Targeted Multi-Agent Communication
    Xu, Chi
    Zhang, Hui
    Zhang, Ya
    [J]. 2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 2915 - 2920
  • [7] ANALYZING MYOPIC APPROACHES FOR MULTI-AGENT COMMUNICATION
    Becker, Raphen
    Carlin, Alan
    Lesser, Victor
    Zilberstein, Shlomo
    [J]. COMPUTATIONAL INTELLIGENCE, 2009, 25 (01) : 31 - 50
  • [8] Learning and communication in multi-agent systems
    Friedrich, H
    Kaiser, M
    Rogalla, O
    Dillmann, R
    [J]. DISTRIBUTED ARTIFICIAL INTELLIGENCE MEETS MACHINE LEARNING: LEARNING IN MULTI-AGENT ENVIRONMENTS, 1997, 1221 : 259 - 275
  • [9] Learning communication for multi-agent systems
    Giles, CL
    Jim, KC
    [J]. INNOVATIVE CONCPTS FOR AGENT-BASED SYSTEMS, 2002, 2564 : 377 - 390
  • [10] Analyzing myopic approaches for multi-agent communication
    Becker, R
    Lesser, V
    Zilberstein, S
    [J]. 2005 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON INTELLIGENT AGENT TECHNOLOGY, PROCEEDINGS, 2005, : 550 - 557