A Safe Training Approach for Deep Reinforcement Learning-based Traffic Engineering

被引:0
|
作者
Wang, Linghao [1 ,2 ]
Wang, Miao [1 ]
Zhang, Yujun [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] Nanjing Inst Informat Superbahn, Nanjing, Peoples R China
基金
美国国家科学基金会;
关键词
Traffic Engineering; Safe Reinforcement Learning; Learning from Demonstration;
D O I
10.1109/ICC45855.2022.9838944
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Traffic engineering (TE) is fundamental and important in modern communication networks. Deep reinforcement learning (DRL)-based TE solutions can solve TE in a data-driven and model-free way thus have attracted much attention recently. However, most of these solutions ignore that TE is a real-world application and there are challenges applying DRL to real-world TE like: (1) Efficiency. Existing learning-from-scratch DRL agent needs long-time interactions to find solutions better than traditional methods. (2) Safety. Existing DRL-based solutions make TE decisions without considering safety constraints, poor decisions may be made and cause significant performance degradation. In this paper, we propose a safe training approach for DRL-based TE, which tries to address the above two problems. It focuses on making full use of data and ensuring safety so that DRL agent for TE can learn more quickly and possibly poor decisions will not be applied to real environment. We implemented the proposed method in ns-3 and simulation results show that our method performs better with faster convergence rate compared to other DRL-based methods while ensuring the safety of the performed TE decisions.
引用
收藏
页码:1450 / 1455
页数:6
相关论文
共 50 条
  • [21] Traffic Engineering Based on Deep Reinforcement Learning in Hybrid IP/SR Network
    Bo Chen
    Penghao Sun
    Peng Zhang
    Julong Lan
    Youjun Bu
    Juan Shen
    China Communications, 2021, 18 (10) : 204 - 213
  • [22] Traffic Engineering Based on Deep Reinforcement Learning in Hybrid IP/SR Network
    Chen, Bo
    Sun, Penghao
    Zhang, Peng
    Lan, Julong
    Bu, Youjun
    Shen, Juan
    CHINA COMMUNICATIONS, 2021, 18 (10) : 204 - 213
  • [23] ScaleDRL: A Scalable Deep Reinforcement Learning Approach for Traffic Engineering in SDN with Pinning Control
    Sun, Penghao
    Guo, Zehua
    Lan, Julong
    Li, Junfei
    Hu, Yuxiang
    Baker, Thar
    COMPUTER NETWORKS, 2021, 190
  • [24] Deep Reinforcement Learning-Based Traffic Sampling for Multiple Traffic Analyzers on Software-Defined Networks
    Kim, Sunghwan
    Yoon, Seunghyun
    Lim, Hyuk
    IEEE ACCESS, 2021, 9 : 47815 - 47827
  • [25] Deep Learning-based Approach on Risk Estimation of Urban Traffic Accidents
    Jin, Zhixiong
    Noh, Byeongjoon
    Cho, Haechan
    Yeo, Hwasoo
    2022 IEEE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2022, : 1446 - 1451
  • [26] EcoMRL: Deep reinforcement learning-based traffic signal control for urban air quality
    Jung, Jaeeun
    Kim, Inhi
    Yoon, Jinwon
    INTERNATIONAL JOURNAL OF SUSTAINABLE TRANSPORTATION, 2024,
  • [27] Deep Reinforcement Learning-based Quantization for Federated Learning
    Zheng, Sihui
    Dong, Yuhan
    Chen, Xiang
    2023 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC, 2023,
  • [28] A Deep Safe Reinforcement Learning Approach for Mapless Navigation
    Lv, Shaohua
    Li, Yanjie
    Liu, Qi
    Gao, Jianqi
    Pang, Xizheng
    Chen, Meiling
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (IEEE-ROBIO 2021), 2021, : 1520 - 1525
  • [29] From Local to Global: A Curriculum Learning Approach for Reinforcement Learning-based Traffic Signal Control
    Zheng, Nianzhao
    Li, Jialong
    Mao, Zhenyu
    Tei, Kenji
    2022 2ND IEEE INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND ARTIFICIAL INTELLIGENCE (SEAI 2022), 2022, : 253 - 258
  • [30] DEEP REINFORCEMENT LEARNING-BASED IRRIGATION SCHEDULING
    Yang, Y.
    Hu, J.
    Porter, D.
    Marek, T.
    Heflin, K.
    Kong, H.
    Sun, L.
    TRANSACTIONS OF THE ASABE, 2020, 63 (03) : 549 - 556