Multi-vehicle Flocking Control with Deep Deterministic Policy Gradient Method

被引:0
|
作者
Xu, Zhao [1 ]
Lyu, Yang [2 ]
Pan, Quan [2 ]
Hu, Jinwen [2 ]
Zhao, Chunhui [2 ]
Liu, Shuai [3 ]
机构
[1] Northwestern Polytech Univ, Sch Elect & Informat, Xian 710072, Shaanxi, Peoples R China
[2] Northwestern Polytech Univ, Sch Automat, Minist Educ, Key Lab Informat Fus, Xian 710072, Shaanxi, Peoples R China
[3] Shandong Univ, Sch Control Sci & Engn, Jinan 250061, Shandong, Peoples R China
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Flocking control has been studied extensively along with the wide application of multi-vehicle systems. In this paper the Multi-vehicles System (MVS) flocking control with collision avoidance and communication preserving is considered based on the deep reinforcement learning framework. Specifically the deep deterministic policy gradient (DDPG) with centralized training and distributed execution process is implemented to obtain the flocking control policy. First, to avoid the dynamically changed observation of state, a three layers tensor based representation of the observation is used so that the state remains constant although the observation dimension is changing. A reward function is designed to guide the way-points tracking, collision avoidance and communication preserving. The reward function is augmented by introducing the local reward function of neighbors. Finally, a centralized training process which trains the shared policy based on common training set among all agents. The proposed method is tested under simulated scenarios with different setup.
引用
下载
收藏
页码:306 / 311
页数:6
相关论文
共 50 条
  • [41] Policy Space Noise in Deep Deterministic Policy Gradient
    Yan, Yan
    Liu, Quan
    NEURAL INFORMATION PROCESSING (ICONIP 2018), PT II, 2018, 11302 : 624 - 634
  • [42] Hybrid Formation Control for Multi-Robot Hunters Based on Multi-Agent Deep Deterministic Policy Gradient
    Hamed O.
    Hamlich M.
    Mendel, 2021, 27 (02) : 23 - 29
  • [43] New Results on Cooperative Multi-Vehicle Deterministic Learning Control: Design and Validation in Gazebo Simulation
    Dong, Xiaonan
    Chen, Xiaotian
    Yuan, Chengzhi
    Stegagno, Paolo
    2020 IEEE/ASME INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS (AIM), 2020, : 1413 - 1418
  • [44] Stability Analysis for Autonomous Vehicle Navigation Trained over Deep Deterministic Policy Gradient
    Cabezas-Olivenza, Mireya
    Zulueta, Ekaitz
    Sanchez-Chica, Ander
    Fernandez-Gamiz, Unai
    Teso-Fz-Betono, Adrian
    MATHEMATICS, 2023, 11 (01)
  • [45] Optimal Transportation Network Company Vehicle Dispatching via Deep Deterministic Policy Gradient
    Shi, Dian
    Li, Xuanheng
    Li, Ming
    Wang, Jie
    Li, Pan
    Pan, Miao
    WIRELESS ALGORITHMS, SYSTEMS, AND APPLICATIONS, WASA 2019, 2019, 11604 : 297 - 309
  • [46] Composite deep learning control for autonomous bicycles by using deep deterministic policy gradient
    He, Kanghui
    Dong, Chaoyang
    Yan, An
    Zheng, Qingyuan
    Liang, Bin
    Wang, Qing
    IECON 2020: THE 46TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2020, : 2766 - 2773
  • [47] Self-guided deep deterministic policy gradient with multi-actor
    Chen, Hongming
    Liu, Quan
    Zhong, Shan
    NEURAL COMPUTING & APPLICATIONS, 2021, 33 (15): : 9723 - 9732
  • [48] Self-guided deep deterministic policy gradient with multi-actor
    Hongming Chen
    Quan Liu
    Shan Zhong
    Neural Computing and Applications, 2021, 33 : 9723 - 9732
  • [49] Intrinsic Motivation for Deep Deterministic Policy Gradient in Multi-Agent Environments
    Cao, Xiaoge
    Lu, Tao
    Cai, Yinghao
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 1628 - 1633
  • [50] A multi-critic deep deterministic policy gradient UAV path planning
    Wu, Runjia
    Gu, Fangqing
    Huang, Jie
    2020 16TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND SECURITY (CIS 2020), 2020, : 6 - 10