Reward Shaping-Based Actor-Critic Deep Reinforcement Learning for Residential Energy Management

被引:27
|
作者
Lu, Renzhi [1 ,2 ,3 ,4 ,5 ]
Jiang, Zhenyu [1 ]
Wu, Huaming [6 ]
Ding, Yuemin [7 ]
Wang, Dong [8 ]
Zhang, Hai-Tao [1 ,9 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Artificial Intelligence & Automat, Key Lab Image Proc & Intelligent Control, Wuhan 430074, Peoples R China
[2] North China Elect Power Univ, State Key Lab Alternate Elect Power Syst Renewable, Beijing 102206, Peoples R China
[3] East China Univ Sci & Technol, Key Lab Smart Mfg Energy Chem Proc, Minist Educ, Shanghai 200237, Peoples R China
[4] Chongqing Univ Posts & Telecommun, Key Lab Ind Internet Things & Networked Control, Chongqing 400065, Peoples R China
[5] Wuhan Univ Sci & Technol, Hubei Key Lab Mech Transmiss & Mfg Engn, Wuhan 430081, Peoples R China
[6] Tianjin Univ, Ctr Appl Math, Tianjin 300072, Peoples R China
[7] Univ Navarra, Dept Elect & Elect Engn, San Sebastian 20018, Spain
[8] Dalian Univ Technol, Key Lab Intelligent Control & Optimizat Ind Equipm, Minist Educ, Dalian 116024, Peoples R China
[9] Huazhong Univ Sci & Technol, State Key Lab Digital Mfg Equipment & Technol, Wuhan 430074, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Energy management; Home appliances; Energy consumption; Costs; Schedules; Informatics; Load modeling; Deep deterministic policy gradient; deep reinforcement learning; demand response; residential energy management; reward shaping (RS); DEMAND RESPONSE;
D O I
10.1109/TII.2022.3183802
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Residential energy consumption continues to climb steadily, requiring intelligent energy management strategies to reduce power system pressures and residential electricity bills. However, it is challenging to design such strategies due to the random nature of electricity pricing, appliance demand, and user behavior. This article presents a novel reward shaping (RS)-based actor-critic deep reinforcement learning (ACDRL) algorithm to manage the residential energy consumption profile with limited information about the uncertain factors. Specifically, the interaction between the energy management center and various residential loads is modeled as a Markov decision process that provides a fundamental mathematical framework to represent the decision-making in situations where outcomes are partially random and partially influenced by the decision-maker control signals, in which the key elements containing the agent, environment, state, action, and reward are carefully designed, and the electricity price is considered as a stochastic variable. An RS-ACDRL algorithm is then developed, incorporating both the actor and critic network and an RS mechanism, to learn the optimal energy consumption schedules. Several case studies involving real-world data are conducted to evaluate the performance of the proposed algorithm. Numerical results demonstrate that the proposed algorithm outperforms state-of-the-art RL methods in terms of learning speed, solution optimality, and cost reduction.
引用
收藏
页码:2662 / 2673
页数:12
相关论文
共 50 条
  • [1] Integrated Actor-Critic for Deep Reinforcement Learning
    Zheng, Jiaohao
    Kurt, Mehmet Necip
    Wang, Xiaodong
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT IV, 2021, 12894 : 505 - 518
  • [2] FUSION SPARSE AND SHAPING REWARD FUNCTION IN SOFT ACTOR-CRITIC DEEP REINFORCEMENT LEARNING FOR MOBILE ROBOT NAVIGATION
    Abu Bakar, Mohamad Hafiz
    Shamsudin, Abu Ubaidah
    Soomro, Zubair Adil
    Tadokoro, Satoshi
    Salaan, C. J.
    [J]. JOURNAL OF THE KOREAN SOCIETY OF RADIOLOGY, 2024, 86 (02): : 37 - 49
  • [3] FUSION SPARSE AND SHAPING REWARD FUNCTION IN SOFT ACTOR-CRITIC DEEP REINFORCEMENT LEARNING FOR MOBILE ROBOT NAVIGATION
    Bakar, Mohamad Hafiz Abu
    Shamsudin, Abu Ubaidah
    Soomro, Zubair Adil
    Tadokoro, Satoshi
    Salaan, C. J.
    [J]. JURNAL TEKNOLOGI-SCIENCES & ENGINEERING, 2024, 86 (02): : 37 - 49
  • [4] Stochastic Integrated Actor-Critic for Deep Reinforcement Learning
    Zheng, Jiaohao
    Kurt, Mehmet Necip
    Wang, Xiaodong
    [J]. IEEE Transactions on Neural Networks and Learning Systems, 2024, 35 (05) : 6654 - 6666
  • [5] Visual Navigation with Actor-Critic Deep Reinforcement Learning
    Shao, Kun
    Zhao, Dongbin
    Zhu, Yuanheng
    Zhang, Qichao
    [J]. 2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [6] Actor-Critic based Improper Reinforcement Learning
    Zaki, Mohammadi
    Mohan, Avinash
    Gopalan, Aditya
    Mannor, Shie
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [7] Averaged Soft Actor-Critic for Deep Reinforcement Learning
    Ding, Feng
    Ma, Guanfeng
    Chen, Zhikui
    Gao, Jing
    Li, Peng
    [J]. COMPLEXITY, 2021, 2021
  • [8] Deep Actor-Critic Reinforcement Learning for Anomaly Detection
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    [J]. 2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [9] Actor-Critic based Deep Reinforcement Learning Framework for Energy Management of Extended Range Electric Delivery Vehicles
    Wang, Pengyue
    Li, Yan
    Shekhar, Shashi
    Northrop, William F.
    [J]. 2019 IEEE/ASME INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS (AIM), 2019, : 1379 - 1384
  • [10] ACTOR-CRITIC DEEP REINFORCEMENT LEARNING FOR DYNAMIC MULTICHANNEL ACCESS
    Zhong, Chen
    Lu, Ziyang
    Gursoy, M. Cenk
    Velipasalar, Senem
    [J]. 2018 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2018), 2018, : 599 - 603