Reward Mechanism Design for Deep Reinforcement Learning-Based Microgrid Energy Management

被引:0
|
作者
Hu, Mingjie [1 ]
Han, Baohui [2 ]
Lv, Shilin [3 ]
Bao, Zhejing [2 ]
Lu, Lingxia [2 ]
Yu, Miao [2 ]
机构
[1] Zhejiang Univ, Polytech Inst, Hangzhou, Peoples R China
[2] Zhejiang Univ, Coll Elect Engn, Hangzhou, Peoples R China
[3] State Grid Zhejiang Elect Power Co Ltd, Hangzhou, Peoples R China
关键词
deep reinforcement learning; microgrid energy management; reward mechanism; cliff walking pattern; Leduc poker pattern;
D O I
10.1109/REPE59476.2023.10512009
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Deep Reinforcement Learning (DRL), with its data-driven nature and model-free advantage, has attracted great interest in the field of microgrid energy management system. The choice of reward mechanism plays a crucial role in the performance and effectiveness of DRL-based microgrid energy management. This paper aims to investigate the reward mechanism design by comparing the performances of DRL-based microgrid energy management under two different reward mechanisms, namely, cliff walking pattern and Leduc poker pattern. The reward mechanism incorporates auxiliary rewards alongside the primary reward to harmonize diverse objectives. Using a real microgrid dataset, the performance of DRL agents under different reward mechanisms are compared. The experimental results demonstrate that different reward mechanisms have a significant impact on the convergence speed and generalization ability of trained microgrid energy management policy.
引用
收藏
页码:201 / 205
页数:5
相关论文
共 50 条
  • [1] An Assessment of Multistage Reward Function Design for Deep Reinforcement Learning-Based Microgrid Energy Management
    Goh, Hui Hwang
    Huang, Yifeng
    Lim, Chee Shen
    Zhang, Dongdong
    Liu, Hui
    Dai, Wei
    Kurniawan, Tonni Agustiono
    Rahman, Saifur
    [J]. IEEE TRANSACTIONS ON SMART GRID, 2022, 13 (06) : 4300 - 4311
  • [2] Online Microgrid Energy Management Based on Safe Deep Reinforcement Learning
    Li, Hepeng
    Wang, Zhenhua
    Li, Lusi
    He, Haibo
    [J]. 2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
  • [3] A unified benchmark for deep reinforcement learning-based energy management: Novel training ideas with the unweighted reward
    Chen, Jiaxin
    Tang, Xiaolin
    Yang, Kai
    [J]. ENERGY, 2024, 307
  • [4] Novel Architecture of Energy Management Systems Based on Deep Reinforcement Learning in Microgrid
    Lee, Seongwoo
    Seon, Joonho
    Sun, Young Ghyu
    Kim, Soo Hyun
    Kyeong, Chanuk
    Kim, Dong In
    Kim, Jin Young
    [J]. IEEE TRANSACTIONS ON SMART GRID, 2024, 15 (02) : 1646 - 1658
  • [5] Deep Reinforcement Learning-based Image Captioning with Embedding Reward
    Ren, Zhou
    Wang, Xiaoyu
    Zhang, Ning
    Lv, Xutao
    Li, Li-Jia
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 1151 - 1159
  • [6] Deep reinforcement learning for energy management in a microgrid with flexible demand
    Nakabi, Taha Abdelhalim
    Toivanen, Pekka
    [J]. SUSTAINABLE ENERGY GRIDS & NETWORKS, 2021, 25
  • [7] A hybrid deep learning-based online energy management scheme for industrial microgrid
    Lu, Renzhi
    Bai, Ruichang
    Ding, Yuemin
    Wei, Min
    Jiang, Junhui
    Sun, Mingyang
    Xiao, Feng
    Zhang, Hai-Tao
    [J]. APPLIED ENERGY, 2021, 304
  • [8] Reinforcement learning-based scheduling strategy for energy storage in microgrid
    Zhou, Kunshu
    Zhou, Kaile
    Yang, Shanlin
    [J]. JOURNAL OF ENERGY STORAGE, 2022, 51
  • [9] Effect of immediate reward function on the performance of reinforcement learning-based energy management system
    Biswas, Atriya
    Wang, Yue
    Emadi, Ali
    [J]. 2022 IEEE/AIAA TRANSPORTATION ELECTRIFICATION CONFERENCE AND ELECTRIC AIRCRAFT TECHNOLOGIES SYMPOSIUM (ITEC+EATS 2022), 2022, : 1021 - 1026
  • [10] Energy management in solar microgrid via reinforcement learning using fuzzy reward
    Kofinas, Panagiotis
    Vouros, George
    Dounis, Anastasios I.
    [J]. ADVANCES IN BUILDING ENERGY RESEARCH, 2018, 12 (01) : 97 - 115