Deep Reinforcement Learning-Based High Concurrent Computing Offloading for Heterogeneous Industrial Tasks

被引:0
|
作者
Liu, Xiao-Yu [1 ,2 ,3 ,4 ]
Xu, Chi [1 ,2 ,3 ]
Zeng, Peng [1 ,2 ,3 ]
Yu, Hai-Bin [1 ,2 ,3 ]
机构
[1] State Key Laboratory of Robotics, Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang,110016, China
[2] Key Laboratory of Networked Control System, Chinese Academy of Sciences, Shenyang,110016, China
[3] Institute of Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang,110169, China
[4] University of Chinese Academy of Sciences, Beijing,100049, China
来源
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Edge computing - Markov processes - Wireless networks - Deep learning - Accident prevention - Learning algorithms - Network security - Learning systems;
D O I
10.11897/SP.J.1016.2021.02369
中图分类号
学科分类号
摘要
With the rapid development of Industry 4.0, massive distributed intelligent industrial devices are interconnected by industrial wireless networks, and generate a large number of heterogeneous industrial tasks with different delay sensitivity and computing load during smart manufacturing. Real-time and efficient processing of industrial tasks is the key factor affecting the safety and efficiency of industrial manufacturing production. However, the limited local computing capacity of industrial devices cannot support the real-time and efficient processing of industrial tasks, and the common industrial cloud computing paradigm results in uncertain communication delays and additional network security issues. It is an effective method to offload industrial tasks to Multi-access Edge Computing servers deployed in base stations, access points and other network edge infrastructures through industrial wireless networks. Nevertheless, the limited time-frequency resources of industrial wireless networks cannot support the high concurrent computing offloading of industrial tasks. With consideration of the difficulty in modeling high concurrent computing offloading of industrial tasks, a Deep Reinforcement Learning-based Concurrent Access with Dynamic Priority (CADP-DRL) algorithm is proposed in this paper. Firstly, the industrial devices are assigned with dynamic priorities according to the delay sensitivity and computing load of their industrial tasks, and the access offloading probabilities of these industrial devices are changed dynamically depending on their priorities. Then, the Markov decision process is utilized to formulate the dynamic priority concurrent computing offloading problem. As both dynamic priority and concurrent computing offloading of massive industrial devices result in the explosion of state space, deep reinforcement learning is used to establish a mapping relationship from states to actions in the high-dimensional state space. Next, the long-term cumulative reward is maximized to obtain an effective dynamic priority concurrent computing offloading policy. Especially, with the aim at the multi-objective decision of dynamic priority and concurrent offloading, a novel compound reward function with joint priority reward and offloading reward is designed. The priority reward is used to ensure reliable offloading of high-priority industrial devices, and the offloading reward is employed to minimize offloading conflicts. In order to guarantee the independent and identical distribution of training data while accelerating the convergence of CADP-DRL, an experience replay with experience-weight is designed. Experiences are classified as high-weighted and low-weighted experiences depending on their weights, and stored in different experience memories respectively. Experiences are randomly sampled as the training data for CADP-DRL, and the sampling probabilities of experiences in different experience memories vary dynamically to break the time correlation among experiences while speeding up the convergence. The expensive training overhead of CADP-DRL is consumed in the offline training phase, and the trained CADP-DRL can make an effective computing offloading decision in real-time in the online execution phase. Slotted-Aloha algorithm is chosen as the benchmark algorithm in the field of communications, and DQN, DDQN and D3QN algorithms are chosen as the benchmark algorithms in the field of deep reinforcement learning. Extensive experiments show that, compared with these benchmark algorithms, CADP-DRL converges quickly, and performs well in generalization. Meanwhile, CADP-DRL always guarantees the highest successful offloading probabilities of high-priority industrial devices with minimum offloading conflicts. © 2021, Science Press. All right reserved.
引用
收藏
页码:2367 / 2381
相关论文
共 50 条
  • [1] Deep Reinforcement Learning-Based Computation Offloading in Vehicular Edge Computing
    Zhan, Wenhan
    Luo, Chunbo
    Wang, Jin
    Min, Geyong
    Duan, Hancong
    [J]. 2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [2] ADRLO: Adaptive deep reinforcement learning-based offloading for edge computing
    Li, Zhigang
    Wang, Yutong
    Zhang, Wentao
    Li, Shujie
    Sun, Xiaochuan
    [J]. PHYSICAL COMMUNICATION, 2023, 61
  • [3] Deep reinforcement learning-based multitask hybrid computing offloading for multiaccess edge computing
    Cai, Jun
    Fu, Hongtian
    Liu, Yan
    [J]. INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (09) : 6221 - 6243
  • [4] A multi-layer guided reinforcement learning-based tasks offloading in edge computing
    Robles-Enciso, Alberto
    Skarmeta, Antonio F.
    [J]. COMPUTER NETWORKS, 2023, 220
  • [5] Deep Reinforcement Learning-Based Offloading Decision Optimization in Mobile Edge Computing
    Zhang, Hao
    Wu, Wenjun
    Wang, Chaoyi
    Li, Meng
    Yang, Ruizhe
    [J]. 2019 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2019,
  • [6] A Clustering Offloading Decision Method for Edge Computing Tasks Based on Deep Reinforcement Learning
    Zhen Zhang
    Huanzhou Li
    Zhangguo Tang
    Dinglin Gu
    Jian Zhang
    [J]. New Generation Computing, 2023, 41 : 85 - 108
  • [7] A Clustering Offloading Decision Method for Edge Computing Tasks Based on Deep Reinforcement Learning
    Zhang, Zhen
    Li, Huanzhou
    Tang, Zhangguo
    Gu, Dinglin
    Zhang, Jian
    [J]. NEW GENERATION COMPUTING, 2023, 41 (01) : 85 - 108
  • [8] Computation offloading and tasks scheduling for the internet of vehicles in edge computing: A deep reinforcement learning-based pointer network approach
    Ju, Xiang
    Su, Shengchao
    Xu, Chaojie
    Wang, Haoxuan
    [J]. COMPUTER NETWORKS, 2023, 223
  • [9] Deep Reinforcement Learning-Based Task Offloading and Load Balancing for Vehicular Edge Computing
    Wu, Zhoupeng
    Jia, Zongpu
    Pang, Xiaoyan
    Zhao, Shan
    [J]. ELECTRONICS, 2024, 13 (08)
  • [10] Deep reinforcement learning-based online task offloading in mobile edge computing networks
    Wu, Haixing
    Geng, Jingwei
    Bai, Xiaojun
    Jin, Shunfu
    [J]. INFORMATION SCIENCES, 2024, 654