Combined heat and power system intelligent economic dispatch: A deep reinforcement learning approach

被引:80
|
作者
Zhou, Suyang [1 ,2 ]
Hu, Zijian [1 ]
Gu, Wei [1 ]
Jiang, Meng [3 ]
Chen, Meng [4 ]
Hong, Qiteng [5 ]
Booth, Campbell [5 ]
机构
[1] Southeast Univ, Sch Elect Engn, 2 Sipailou Xuanwu Qu, Nanjing, Peoples R China
[2] Jiangsu Key Lab Smart Grid Technol & Equipment, Nanjing, Peoples R China
[3] Univ Notre Dame, Dept Comp Sci & Engn, Notre Dame, IN 46556 USA
[4] Jiangsu Elect Power Co State Grid Corp China, Yangzhou Power Supply Co, Yangzhou, Jiangsu, Peoples R China
[5] Univ Strathclyde, Dept Elect & Elect Engn, Glasgow, Lanark, Scotland
基金
中国国家自然科学基金;
关键词
Combined heat and power economic dispatch; Deep reinforcement learning; Proximal policy optimization; ALGORITHM; OPTIMIZATION; MODEL; OPERATION;
D O I
10.1016/j.ijepes.2020.106016
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper proposed a Deep Reinforcement learning (DRL) approach for Combined Heat and Power (CHP) system economic dispatch which obtain adaptability for different operating scenarios and significantly decrease the computational complexity without affecting accuracy. In the respect of problem description, a vast of Combined Heat and Power (CHP) economic dispatch problems are modeled as a high-dimensional and non-smooth objective function with a large number of non-linear constraints for which powerful optimization algorithms and considerable time are required to solve it. In order to reduce the solution time, most engineering applications choose to linearize the optimization target and devices model. To avoid complicated linearization process, this paper models CHP economic dispatch problems as Markov Decision Process (MDP) that making the model highly encapsulated to preserve the input and output characteristics of various devices. Furthermore, we improve an advanced deep reinforcement learning algorithm: distributed proximal policy optimization (DPPO), to make it applicable to CHP economic dispatch problem. Based on this algorithm, the agent will be trained to explore optimal dispatch strategies for different operation scenarios and respond to system emergencies efficiently. In the utility phase, the trained agent will generate optimal control strategy in real time based on current system state. Compared with existing optimization methods, advantages of DRL methods are mainly reflected in the following three aspects: 1) Adaptability: under the premise of the same network topology, the trained agent can handle the economic scheduling problem in various operating scenarios without recalculation. 2) High encapsulation: The user only needs to input the operating state to get the control strategy, while the optimization algorithm needs to re-write the constraints and other formulas for different situations. 3) Time scale flexibility: It can be applied to both the day-ahead optimized scheduling and the real-time control. The proposed method is applied to two test system with different characteristics. The results demonstrate that the DRL method could handle with varieties of operating situations while get better optimization performance than most of other algorithms.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Optimal Dispatch of Combined Heat and Power System Based on Multi-agent Deep Reinforcement Learning
    Dong, Lei
    Liu, Yu
    Qiao, Ji
    Wang, Xinying
    Wang, Chunfei
    Pu, Tianjiao
    [J]. Dianwang Jishu/Power System Technology, 2021, 45 (12): : 4729 - 4737
  • [2] Intelligent Economic Dispatch for PV-PHS Integrated System: a Deep Reinforcement Learning-Based Approach
    Li, Tao
    Hu, Weihao
    Li, Jian
    Han, Xiaoyan
    Chen, Zhe
    [J]. Diangong Jishu Xuebao/Transactions of China Electrotechnical Society, 2020, 35 (13): : 2757 - 2768
  • [3] Low-carbon economic dispatch of the combined heat and power-virtual power plants: A improved deep reinforcement learning-based approach
    Tan, Yongqiang
    Shen, Yanxia
    Yu, Xinyan
    Lu, Xin
    [J]. IET RENEWABLE POWER GENERATION, 2023, 17 (04) : 982 - 1007
  • [4] Economic Emission Dispatch of Combined Heat and Power System
    Adhvaryyu, Pradosh Kumar
    Chattopadhyay, Pranab Kumar
    Bhattacharjya, Aniruddha
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING, INFORMATICS, COMMUNICATION AND ENERGY SYSTEMS (SPICES), 2015,
  • [5] A Benders decomposition approach for a combined heat and power economic dispatch
    Abdolmohammadi, Hamid Reza
    Kazemi, Ahad
    [J]. ENERGY CONVERSION AND MANAGEMENT, 2013, 71 : 21 - 31
  • [6] Artificial immune system for combined heat and power economic dispatch
    Basu, M.
    [J]. INTERNATIONAL JOURNAL OF ELECTRICAL POWER & ENERGY SYSTEMS, 2012, 43 (01) : 1 - 5
  • [7] A test system for combined heat and power economic dispatch problems
    Algie, C
    Wong, KP
    [J]. PROCEEDINGS OF THE 2004 IEEE INTERNATIONAL CONFERENCE ON ELECTRIC UTILITY DEREGULATION, RESTRUCTURING AND POWER TECHNOLOGIES, VOLS 1 AND 2, 2004, : 96 - 101
  • [8] Deep Reinforcement Learning for Economic Dispatch of Virtual Power Plant in Internet of Energy
    Lin, Lin
    Guan, Xin
    Peng, Yu
    Wang, Ning
    Maharjan, Sabita
    Ohtsuki, Tomoaki
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (07) : 6288 - 6301
  • [9] Solving large-scale combined heat and power economic dispatch problems by using deep reinforcement learning based crisscross optimization algorithm
    Meng, Anbo
    Rong, Jiayu
    Yin, Hao
    Luo, Jianqiang
    Tang, Yanshu
    Zhang, Haitao
    Li, Chen
    Zhu, Jianbin
    Yin, Yiding
    Li, Hanhong
    Liu, Jiawei
    [J]. APPLIED THERMAL ENGINEERING, 2024, 245
  • [10] Economic Emission Dispatch of Wind Power Integrated Combined Heat and Power System
    Adhvaryyu, Pradosh Kumar
    Chattopadhyay, Pranab Kumar
    Bhattacharjya, Aniruddha
    [J]. 2014 EIGHTEENTH NATIONAL POWER SYSTEMS CONFERENCE (NPSC), 2014,