Stochastic Optimal CPS Relaxed Control Methodology for Interconnected Power Systems Using Q-Learning Method

被引:26
|
作者
Yu, Tao [3 ]
Zhou, Bin [1 ]
Chan, Ka Wing [1 ]
Lu, En [2 ]
机构
[1] Hong Kong Polytech Univ, Dept Elect Engn, HKSAR, Hong Kong, Hong Kong, Peoples R China
[2] China So Power Grid Co, Guangdong Power Dispatching Ctr, Guangzhou 510600, Guangdong, Peoples R China
[3] S China Univ Technol, Coll Elect Engn, Guangzhou 510641, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
Q-learning algorithm; Reinforcement learning; Automatic generation control; Control performance standard; Markov decision process; Optimal control; China Southern Power Grid; DYNAMIC ANALYSIS; PERFORMANCE; (ACE)OVER-BAR(1); NERCS;
D O I
10.1061/(ASCE)EY.1943-7897.0000017
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
This paper presents the application and design of a novel stochastic optimal control methodology based on the Q-learning method for solving the automatic generation control (AGC) under the new control performance standards (CPS) for the North American Electric Reliability Council (NERC). The aims of CPS are to relax the control constraint requirements of AGC plant regulation and enhance the frequency dispatch support effect from interconnected control areas. The NERC's CPS-based AGC problem is a dynamic stochastic decision problem that can be modeled as a reinforcement learning (RL) problem based on the Markov decision process theory. In this paper, the Q-learning method is adopted as the RL core algorithm with CPS values regarded as the rewards from the interconnected power systems; the CPS control and relaxed control objectives are formulated as immediate reward functions by means of a linear weighted aggregative approach. By regulating a closed-loop CPS control rule to maximize the long-term discounted reward in the procedure of online learning, the optimal CPS control strategy can be gradually obtained. This paper also introduces a practical semisupervisory group prelearning method to improve the stability and convergence ability of Q-learning controllers during the prelearning process. Tests on the China Southern Power Grid demonstrate that the proposed control strategy can effectively enhance the robustness and relaxation property of AGC systems while CPS compliances are ensured. DOI:10.1061/(ASCE)EY.1943-7897.0000017. (C) 2011 American Society of Civil Engineers.
引用
收藏
页码:116 / 129
页数:14
相关论文
共 50 条
  • [21] Q-learning Approach for Optimal Power Dispatch of Microgrid
    Samadi, Esmat
    Badri, Ali
    Ebrahimpour, Reza
    2020 28TH IRANIAN CONFERENCE ON ELECTRICAL ENGINEERING (ICEE), 2020, : 935 - 939
  • [22] Decentralized Q-Learning for Uplink Power Control
    Dzulkifly, Sumayyah
    Giupponi, Lorenza
    Said, Fatin
    Dohler, Mischa
    2015 IEEE 20TH INTERNATIONAL WORKSHOP ON COMPUTER AIDED MODELLING AND DESIGN OF COMMUNICATION LINKS AND NETWORKS (CAMAD), 2015, : 54 - 58
  • [23] Q-learning optimal state estimation and control for discrete systems with unknown parameters
    Li J.-N.
    Ma S.-K.
    Kongzhi yu Juece/Control and Decision, 2021, 35 (12): : 2889 - 2897
  • [24] An Optimal Control Method for Expressways Entering Ramps Metering Based on Q-Learning
    Ji, Xiaofeng
    He, Zenghui
    ICICTA: 2009 SECOND INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTATION TECHNOLOGY AND AUTOMATION, VOL I, PROCEEDINGS, 2009, : 739 - 741
  • [25] Stochastic linear quadratic optimal tracking control for discrete-time systems with delays based on Q-learning algorithm
    Tan, Xufeng
    Li, Yuan
    Liu, Yang
    AIMS MATHEMATICS, 2023, 8 (05): : 10249 - 10265
  • [26] Optimal Consensus Control for Discrete-Time Systems with State Delay Using Q-learning Solution
    Zhang, Li
    Huo, Shicheng
    Zhang, Ya
    2022 IEEE 17TH INTERNATIONAL CONFERENCE ON CONTROL & AUTOMATION, ICCA, 2022, : 630 - 635
  • [27] A Novel Self-tuning CPS Controller Based on Q-learning Method
    Tao, Yu
    Bin, Zhou
    2008 IEEE POWER & ENERGY SOCIETY GENERAL MEETING, VOLS 1-11, 2008, : 1083 - 1088
  • [28] Fuzzy Q-learning Control for Temperature Systems
    Chen, Yeong-Chin
    Hung, Lon-Chen
    Syamsudin, Mariana
    22ND IEEE/ACIS INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ARTIFICIAL INTELLIGENCE, NETWORKING AND PARALLEL/DISTRIBUTED COMPUTING (SNPD 2021-FALL), 2021, : 148 - 151
  • [29] On-policy Q-learning for Adaptive Optimal Control
    Jha, Sumit Kumar
    Bhasin, Shubhendu
    2014 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning (ADPRL), 2014, : 301 - 306
  • [30] Input-Decoupled Q-Learning for Optimal Control
    Minh Q. Phan
    Seyed Mahdi B. Azad
    The Journal of the Astronautical Sciences, 2020, 67 : 630 - 656