Smart Underwater Pollution Detection Based on Graph-Based Multi-Agent Reinforcement Learning Towards AUV-Based Network ITS

被引:28
|
作者
Lin, Chuan [1 ]
Han, Guangjie [2 ,3 ]
Zhang, Tongwei [4 ]
Shah, Syed Bilal Hussain [5 ]
Peng, Yan [6 ]
机构
[1] Northeastern Univ, Software Coll, Shenyang 110819, Peoples R China
[2] Hohai Univ, Dept Internet Things Engn, Changzhou 213022, Peoples R China
[3] Chinese Acad Sci, Inst Acoust, State Key Lab Acoust, Beijing 100190, Peoples R China
[4] Natl Deep Sea Ctr, Dept Technol, Qingdao 266237, Peoples R China
[5] Manchester Metropolitan Univ, Sch Comp & Math, Manchester M15 6BH, Lancs, England
[6] Shanghai Univ, Artificial Intelligence Inst, Shanghai 200444, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Pollution; Task analysis; Optimization; Systems architecture; Heuristic algorithms; Computer architecture; Routing; Autonomous underwater vehicle; software-defined networking; 6G; graph-based soft actor critic; intelligent transportation systems; EFFICIENT DATA-COLLECTION; ENERGY-EFFICIENT; INTERNET; SCHEME; ARCHITECTURE;
D O I
10.1109/TITS.2022.3162850
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
The exploitation/utilization of marine resources and the rapid development of urbanization along coastal cities result in serious marine pollution, especially underwater diffusion pollution. It is a non-trivial task to detect the source of diffusion pollution, such that the disadvantageous effect of the pollution can be reduced. With the vision of 6G framework, we employ Autonomous Underwater Vehicle (AUV) flock and introduce the concept of AUV-based network. In particular, we utilize the Software-Defined Networking (SDN) technique to update the controllability of the AUV-based network, leading to the paradigm of SDN-enabled multi-AUVs network Intelligent Transportation Systems (SDNA-ITS). For SDNA-ITS, we utilize artificial potential field theories to model the control model. To optimize the system output, we introduce the graph-based Soft Actor-Critic (SAC) algorithm, i.e., a category of Multi-Agent Reinforcement Learning (MARL) mechanism where each AUV can be regarded as a node in a graph. In particular, we improve the optimization model based on Centralized Training Decentralized Execution (CTDE) architecture with the assistance of the SDN controller, by which each AUV can efficiently adjust its speed towards the diffusion source. Further, to achieve exact path planning for detecting the diffusion source, a dynamic detection scheme is proposed to output the united control policy to schedule the SDNA-ITS dynamically. Simulation results demonstrate that our approaches are available to detect the underwater diffusion source when the actual scenario is taken into account and perform better than some recent research products.
引用
收藏
页码:7494 / 7505
页数:12
相关论文
共 50 条
  • [1] Collaborative Information Dissemination with Graph-Based Multi-Agent Reinforcement Learning
    Galliera, Raffaele
    Venable, Kristen Brent
    Bassani, Matteo
    Suri, Niranjan
    ALGORITHMIC DECISION THEORY, ADT 2024, 2025, 15248 : 160 - 173
  • [2] Learning Heterogeneous Strategies via Graph-based Multi-agent Reinforcement Learning
    Li, Yang
    Luo, Xiangfeng
    Xie, Shaorong
    2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 709 - 713
  • [3] Decentralized graph-based multi-agent reinforcement learning using reward machines
    Hu, Jueming
    Xu, Zhe
    Wang, Weichang
    Qu, Guannan
    Pang, Yutian
    Liu, Yongming
    NEUROCOMPUTING, 2024, 564
  • [4] Multi-Agent Collaborative Exploration through Graph-based Deep Reinforcement Learning
    Luo, Tianze
    Subagdja, Budhitama
    Wang, Di
    Tan, Ah-Hwee
    2019 IEEE INTERNATIONAL CONFERENCE ON AGENTS (ICA), 2019, : 2 - 7
  • [5] A Graph-Based Soft Actor Critic Approach in Multi-Agent Reinforcement Learning
    Pan, Wei
    Liu, Cheng
    INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2023, 18 (01)
  • [6] Towards Self-X cognitive manufacturing network: An industrial knowledge graph-based multi-agent reinforcement learning approach
    Zheng, Pai
    Xia, Liqiao
    Li, Chengxi
    Li, Xinyu
    Liu, Bufan
    JOURNAL OF MANUFACTURING SYSTEMS, 2021, 61 : 16 - 26
  • [7] Graph-based multi-agent reinforcement learning for collaborative search and tracking of multiple UAVs
    Zhao, Bocheng
    Huo, Mingying
    Li, Zheng
    Feng, Wenyu
    Yu, Ze
    Qi, Naiming
    Wang, Shaohai
    CHINESE JOURNAL OF AERONAUTICS, 2025, 38 (03)
  • [8] Graph-based Selection-Activation Reinforcement Learning for Heterogenous Multi-agent Collaboration
    Chen, Hao-Xiang
    Zhang, Xi-Wen
    Shen, Jun-Nan
    Chinese Control Conference, CCC, 2024, : 5835 - 5840
  • [9] GRAPHCOMM: A GRAPH NEURAL NETWORK BASED METHOD FOR MULTI-AGENT REINFORCEMENT LEARNING
    Shen, Siqi
    Fu, Yongquan
    Su, Huayou
    Pan, Hengyue
    Qiao, Peng
    Dou, Yong
    Wang, Cheng
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3510 - 3514
  • [10] Graph-based multi-agent decision making
    Yu, Xiaohan
    Xu, Zeshui
    INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2012, 53 (04) : 502 - 512