RLink: Accelerate On-Device Deep Reinforcement Learning with Inference Knowledge at the Edge

被引:0
|
作者
Zeng, Tianyu [1 ]
Zhang, Xiaoxi [1 ]
Feng, Daipeng [1 ]
Duan, Jingpu [2 ,3 ]
Zhou, Zhi [1 ]
Chen, Xu [1 ]
机构
[1] Sun Yat Sen Univ, Sch Comp Sci & Engn, Guangzhou, Peoples R China
[2] Southern Univ Sci & Technol, Inst Future Networks, Shenzhen, Peoples R China
[3] Peng Cheng Lab, Dept Communicat, Shenzhen, Peoples R China
基金
美国国家科学基金会;
关键词
Edge intelligence; distributed architecture; deep reinforcement learning; knowledge distillation; training acceleration;
D O I
10.1109/MSN60784.2023.00093
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep reinforcement learning (DRL) has been a successful paradigm in machine learning that enables solving complex control problems at the human level. However, the sampling and training efficiency of state-of-the-art DRL frameworks can not satisfy the stringent latency and throughput requirements of today's mobile environments. Existing distributed and offline reinforcement learning algorithms along with the libraries for training acceleration are inherently designed for DRL tasks performed in the cloud rather than on distributed mobile devices, on which the computing resources are highly constrained, heterogeneous, and possibly dynamically changing. With the rise of edge computing and intelligence services, this paper presents RLink, a novel distributed training library to accelerate on-device deep reinforcement learning with inference knowledge at the edge. We leverage knowledge distillation to realize lightweight interaction between our on-device training task and the remote models that can provide inference knowledge. In this way, RLink is designed to be event-driven and agnostic to heterogeneous deep reinforcement learning algorithms and libraries. To tackle the communication bottleneck, a novel asynchronous sampling algorithm is proposed to facilitate real-time training in RLink. Tuned for unstable-connected mobile devices, RLink is robust and efficient by using a semantic-aware communication pipeline for lossless data compression. Extensive experimental results show that, compared with state-of-the-art algorithms and libraries, RLink can accelerate deep reinforcement learning at the edge with up to decuple speedups in convergence and ideal computational performance.
引用
收藏
页码:628 / 635
页数:8
相关论文
共 50 条
  • [1] Knowledge Transfer for On-Device Deep Reinforcement Learning in Resource Constrained Edge Computing Systems
    Jang, Ingook
    Kim, Hyunseok
    Lee, Donghun
    Son, Young-Sung
    Kim, Seonghyun
    [J]. IEEE ACCESS, 2020, 8 : 146588 - 146597
  • [2] On-Device Deep Learning Inference for Efficient Activity Data Collection
    Mairittha, Nattaya
    Mairittha, Tittaya
    Inoue, Sozo
    [J]. SENSORS, 2019, 19 (15)
  • [3] A Hardware Prototype Targeting Distributed Deep Learning for On-device Inference
    Farcas, Allen-Jasmin
    Li, Guihong
    Bhardwaj, Kartikeya
    Marculescu, Radu
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 1600 - 1601
  • [4] RLINK: Deep reinforcement learning for user identity linkage
    Xiaoxue Li
    Yanan Cao
    Qian Li
    Yanmin Shang
    Yangxi Li
    Yanbing Liu
    Guandong Xu
    [J]. World Wide Web, 2021, 24 : 85 - 103
  • [5] RLINK: Deep reinforcement learning for user identity linkage
    Li, Xiaoxue
    Cao, Yanan
    Li, Qian
    Shang, Yanmin
    Li, Yangxi
    Liu, Yanbing
    Xu, Guandong
    [J]. WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2021, 24 (01): : 85 - 103
  • [6] On-Device Deep Learning Inference for System-on-Chip (SoC) Architectures
    Springer, Tom
    Eiroa-Lledo, Elia
    Stevens, Elizabeth
    Linstead, Erik
    [J]. ELECTRONICS, 2021, 10 (06) : 1 - 21
  • [7] Digital In-Memory Computing to Accelerate Deep Learning Inference on the Edge
    Perri, Stefania
    Zambelli, Cristian
    Ielmini, Daniele
    Silvano, Cristina
    [J]. 2024 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS, IPDPSW 2024, 2024, : 130 - 133
  • [8] Mistify: Automating DNN Model Porting for On-Device Inference at the Edge
    Guo, Peizhen
    Hu, Bo
    Hu, Wenjun
    [J]. PROCEEDINGS OF THE 18TH USENIX SYMPOSIUM ON NETWORKED SYSTEM DESIGN AND IMPLEMENTATION, 2021, : 705 - 720
  • [9] Deep-Learning-Based Pedestrian Inertial Navigation: Methods, Data Set, and On-Device Inference
    Chen, Changhao
    Zhao, Peijun
    Lu, Chris Xiaoxuan
    Wang, Wei
    Markham, Andrew
    Trigoni, Niki
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (05): : 4431 - 4441
  • [10] Mobile Edge NLU with On-device Inference for Humanitarian Assistance during Disasters
    Mabuntham, Karunchai
    Marurngsith, Worawan
    [J]. 2022 IEEE 5TH INTERNATIONAL CONFERENCE ON ELECTRONICS AND COMMUNICATION ENGINEERING, ICECE, 2022, : 239 - 243