Reward Shaping-Based Actor-Critic Deep Reinforcement Learning for Residential Energy Management

被引:27
|
作者
Lu, Renzhi [1 ,2 ,3 ,4 ,5 ]
Jiang, Zhenyu [1 ]
Wu, Huaming [6 ]
Ding, Yuemin [7 ]
Wang, Dong [8 ]
Zhang, Hai-Tao [1 ,9 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Artificial Intelligence & Automat, Key Lab Image Proc & Intelligent Control, Wuhan 430074, Peoples R China
[2] North China Elect Power Univ, State Key Lab Alternate Elect Power Syst Renewable, Beijing 102206, Peoples R China
[3] East China Univ Sci & Technol, Key Lab Smart Mfg Energy Chem Proc, Minist Educ, Shanghai 200237, Peoples R China
[4] Chongqing Univ Posts & Telecommun, Key Lab Ind Internet Things & Networked Control, Chongqing 400065, Peoples R China
[5] Wuhan Univ Sci & Technol, Hubei Key Lab Mech Transmiss & Mfg Engn, Wuhan 430081, Peoples R China
[6] Tianjin Univ, Ctr Appl Math, Tianjin 300072, Peoples R China
[7] Univ Navarra, Dept Elect & Elect Engn, San Sebastian 20018, Spain
[8] Dalian Univ Technol, Key Lab Intelligent Control & Optimizat Ind Equipm, Minist Educ, Dalian 116024, Peoples R China
[9] Huazhong Univ Sci & Technol, State Key Lab Digital Mfg Equipment & Technol, Wuhan 430074, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Energy management; Home appliances; Energy consumption; Costs; Schedules; Informatics; Load modeling; Deep deterministic policy gradient; deep reinforcement learning; demand response; residential energy management; reward shaping (RS); DEMAND RESPONSE;
D O I
10.1109/TII.2022.3183802
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Residential energy consumption continues to climb steadily, requiring intelligent energy management strategies to reduce power system pressures and residential electricity bills. However, it is challenging to design such strategies due to the random nature of electricity pricing, appliance demand, and user behavior. This article presents a novel reward shaping (RS)-based actor-critic deep reinforcement learning (ACDRL) algorithm to manage the residential energy consumption profile with limited information about the uncertain factors. Specifically, the interaction between the energy management center and various residential loads is modeled as a Markov decision process that provides a fundamental mathematical framework to represent the decision-making in situations where outcomes are partially random and partially influenced by the decision-maker control signals, in which the key elements containing the agent, environment, state, action, and reward are carefully designed, and the electricity price is considered as a stochastic variable. An RS-ACDRL algorithm is then developed, incorporating both the actor and critic network and an RS mechanism, to learn the optimal energy consumption schedules. Several case studies involving real-world data are conducted to evaluate the performance of the proposed algorithm. Numerical results demonstrate that the proposed algorithm outperforms state-of-the-art RL methods in terms of learning speed, solution optimality, and cost reduction.
引用
收藏
页码:2662 / 2673
页数:12
相关论文
共 50 条
  • [31] Power Allocation in Dual Connectivity Networks Based on Actor-Critic Deep Reinforcement Learning
    Moein, Elham
    Hasibi, Ramin
    Shokri, Matin
    Rasti, Mehdi
    [J]. 17TH INTERNATIONAL SYMPOSIUM ON MODELING AND OPTIMIZATION IN MOBILE, AD HOC, AND WIRELESS NETWORKS (WIOPT 2019), 2019, : 170 - 177
  • [32] Swarm Reinforcement Learning Method Based on an Actor-Critic Method
    Iima, Hitoshi
    Kuroe, Yasuaki
    [J]. SIMULATED EVOLUTION AND LEARNING, 2010, 6457 : 279 - 288
  • [33] Manipulator Motion Planning based on Actor-Critic Reinforcement Learning
    Li, Qiang
    Nie, Jun
    Wang, Haixia
    Lu, Xiao
    Song, Shibin
    [J]. 2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 4248 - 4254
  • [34] Evaluating Correctness of Reinforcement Learning based on Actor-Critic Algorithm
    Kim, Youngjae
    Hussain, Manzoor
    Suh, Jae-Won
    Hong, Jang-Eui
    [J]. 2022 THIRTEENTH INTERNATIONAL CONFERENCE ON UBIQUITOUS AND FUTURE NETWORKS (ICUFN), 2022, : 320 - 325
  • [35] Multi-actor mechanism for actor-critic reinforcement learning
    Li, Lin
    Li, Yuze
    Wei, Wei
    Zhang, Yujia
    Liang, Jiye
    [J]. INFORMATION SCIENCES, 2023, 647
  • [36] Distributed Structured Actor-Critic Reinforcement Learning for Universal Dialogue Management
    Chen, Zhi
    Chen, Lu
    Liu, Xiaoyuan
    Yu, Kai
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 : 2400 - 2411
  • [37] An Actor-Critic Reinforcement Learning Approach for Energy Harvesting Communications Systems
    Masadeh, Ala'eddin
    Wang, Zhengdao
    Kamal, Ahmed E.
    [J]. 2019 28TH INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND NETWORKS (ICCCN), 2019,
  • [38] Fast Learning in an Actor-Critic Architecture with Reward and Punishment
    Balkenius, Christian
    Winberg, Stefan
    [J]. TENTH SCANDINAVIAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2008, 173 : 20 - 27
  • [39] Simultaneous Vibration Control and Energy Harvesting Using Actor-Critic Based Reinforcement Learning
    Loong, Cheng Ning
    Chang, C. C.
    Dimitrakoloulos, Elias G.
    [J]. ACTIVE AND PASSIVE SMART STRUCTURES AND INTEGRATED SYSTEMS XII, 2018, 10595
  • [40] Dynamic power management of an embedded sensor network based on actor-critic reinforcement based learning
    Sridhar, Prasanna
    Nanayakkara, Thrishantha
    Madni, Asad M.
    Jamshidi, Mo
    [J]. 2007 THIRD INTERNATIONAL CONFERENCE ON INFORMATION AND AUTOMATION FOR SUSTAINABILITY, 2007, : 71 - +