Multi-Agent Reinforcement Learning for Coordinating Communication and Control

被引:2
|
作者
Mason, Federico [1 ]
Chiariotti, Federico [1 ]
Zanella, Andrea [1 ]
Popovski, Petar [2 ]
机构
[1] Univ Padua, Dept Informat Engn, I-35131 Padua, Italy
[2] Aalborg Univ, Dept Elect Syst, Aalborg, Denmark
关键词
Optimization; Networked control systems; Sensors; Quality of service; Process control; Wireless communication; Robot sensing systems; Markov decision processes; networked control systems; goal-oriented communications; multi-agent reinforcement learning; NETWORKED CONTROL-SYSTEMS; INFORMATION; AGE; PROTOCOLS; DELAY;
D O I
10.1109/TCCN.2024.3384492
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
The automation of factories and manufacturing processes has been accelerating over the past few years, leading to an ever-increasing number of scenarios with networked agents whose coordination requires reliable wireless communication. In this context, goal-oriented communication adapts transmissions to the control task, prioritizing the more relevant information to decide which action to take. Instead, networked control models follow the opposite pathway, optimizing physical actions to address communication impairments. In this work, we propose a joint design that combines goal-oriented communication and networked control into a single optimization model, an extension of a multi-agent Partially Observable Markov Decision Process (POMDP), which we call Cyber-Physical POMDP. The proposed model is flexible enough to represent a large variety of scenarios and we illustrate its potential in two simple use cases with a single agent and a set of supporting sensors. Our results assess that the joint optimization of communication and control tasks radically improves the performance of networked control systems, particularly in the case of constrained resources, leading to implicit coordination of communication actions.
引用
收藏
页码:1566 / 1581
页数:16
相关论文
共 50 条
  • [1] Coordinating Multi-Agent Navigation by Learning Communication
    Hildreth, Dalto N.
    Guy, Stephen J.
    [J]. PROCEEDINGS OF THE ACM ON COMPUTER GRAPHICS AND INTERACTIVE TECHNIQUES, 2019, 2 (02)
  • [2] Multi-Agent Reinforcement Learning With Distributed Targeted Multi-Agent Communication
    Xu, Chi
    Zhang, Hui
    Zhang, Ya
    [J]. 2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 2915 - 2920
  • [3] Learning structured communication for multi-agent reinforcement learning
    Junjie Sheng
    Xiangfeng Wang
    Bo Jin
    Junchi Yan
    Wenhao Li
    Tsung-Hui Chang
    Jun Wang
    Hongyuan Zha
    [J]. Autonomous Agents and Multi-Agent Systems, 2022, 36
  • [4] Learning structured communication for multi-agent reinforcement learning
    Sheng, Junjie
    Wang, Xiangfeng
    Jin, Bo
    Yan, Junchi
    Li, Wenhao
    Chang, Tsung-Hui
    Wang, Jun
    Zha, Hongyuan
    [J]. AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2022, 36 (02)
  • [5] Multi-agent reinforcement learning for character control
    Li, Cheng
    Fussell, Levi
    Komura, Taku
    [J]. VISUAL COMPUTER, 2021, 37 (12): : 3115 - 3123
  • [6] Learning of Communication Codes in Multi-Agent Reinforcement Learning Problem
    Kasai, Tatsuya
    Tenmoto, Hiroshi
    Kamiya, Akimoto
    [J]. 2008 IEEE CONFERENCE ON SOFT COMPUTING IN INDUSTRIAL APPLICATIONS SMCIA/08, 2009, : 1 - +
  • [7] Multi-agent reinforcement learning for character control
    Cheng Li
    Levi Fussell
    Taku Komura
    [J]. The Visual Computer, 2021, 37 : 3115 - 3123
  • [8] Multi-agent reinforcement learning based on local communication
    Zhang, Wenxu
    Ma, Lei
    Li, Xiaonan
    [J]. CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2019, 22 (Suppl 6): : 15357 - 15366
  • [9] Multi-Agent Deep Reinforcement Learning with Emergent Communication
    Simoes, David
    Lau, Nuno
    Reis, Luis Paulo
    [J]. 2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [10] Improving coordination with communication in multi-agent reinforcement learning
    Szer, D
    Charpillet, F
    [J]. ICTAI 2004: 16TH IEEE INTERNATIONALCONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2004, : 436 - 440