Stochastic Optimal CPS Relaxed Control Methodology for Interconnected Power Systems Using Q-Learning Method

被引:25
|
作者
Yu, Tao [3 ]
Zhou, Bin [1 ]
Chan, Ka Wing [1 ]
Lu, En [2 ]
机构
[1] Hong Kong Polytech Univ, Dept Elect Engn, HKSAR, Hong Kong, Hong Kong, Peoples R China
[2] China So Power Grid Co, Guangdong Power Dispatching Ctr, Guangzhou 510600, Guangdong, Peoples R China
[3] S China Univ Technol, Coll Elect Engn, Guangzhou 510641, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
Q-learning algorithm; Reinforcement learning; Automatic generation control; Control performance standard; Markov decision process; Optimal control; China Southern Power Grid; DYNAMIC ANALYSIS; PERFORMANCE; (ACE)OVER-BAR(1); NERCS;
D O I
10.1061/(ASCE)EY.1943-7897.0000017
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
This paper presents the application and design of a novel stochastic optimal control methodology based on the Q-learning method for solving the automatic generation control (AGC) under the new control performance standards (CPS) for the North American Electric Reliability Council (NERC). The aims of CPS are to relax the control constraint requirements of AGC plant regulation and enhance the frequency dispatch support effect from interconnected control areas. The NERC's CPS-based AGC problem is a dynamic stochastic decision problem that can be modeled as a reinforcement learning (RL) problem based on the Markov decision process theory. In this paper, the Q-learning method is adopted as the RL core algorithm with CPS values regarded as the rewards from the interconnected power systems; the CPS control and relaxed control objectives are formulated as immediate reward functions by means of a linear weighted aggregative approach. By regulating a closed-loop CPS control rule to maximize the long-term discounted reward in the procedure of online learning, the optimal CPS control strategy can be gradually obtained. This paper also introduces a practical semisupervisory group prelearning method to improve the stability and convergence ability of Q-learning controllers during the prelearning process. Tests on the China Southern Power Grid demonstrate that the proposed control strategy can effectively enhance the robustness and relaxation property of AGC systems while CPS compliances are ensured. DOI:10.1061/(ASCE)EY.1943-7897.0000017. (C) 2011 American Society of Civil Engineers.
引用
收藏
页码:116 / 129
页数:14
相关论文
共 50 条
  • [1] Q-learning based dynamic optimal CPS control methodology for interconnected power systems
    College of Electric Power, South China University of Technology, Guangzhou 510640, China
    不详
    Zhongguo Dianji Gongcheng Xuebao, 2009, 19 (13-19):
  • [2] Stochastic Optimal Control of Unknown Linear Networked Control System using Q-Learning Methodology
    Xu, Hao
    Jagannathan, S.
    2011 AMERICAN CONTROL CONFERENCE, 2011, : 2819 - 2824
  • [3] Optimal Control for Interconnected Multi-Area Power Systems With Unknown Dynamics: An Off-Policy Q-Learning Method
    Wang, Jing
    Mi, Xuanrui
    Shen, Hao
    Park, Ju H.
    Shi, Kaibo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2024, 71 (05) : 2849 - 2853
  • [4] Reinforcement learning based CPS self-tuning control methodology for interconnected power systems
    Yu, Tao
    Zhou, Bin
    Dianli Xitong Baohu yu Kongzhi/Power System Protection and Control, 2009, 37 (10): : 33 - 38
  • [5] Artificial emotionnal Q-learning for automatic generation control of interconnected power grids
    Yin L.-F.
    Zheng B.-M.
    Yu T.
    Yu, Tao (taoyu1@scut.edu.cn), 1650, South China University of Technology (33): : 1650 - 1657
  • [6] An optimal control method for hybrid systems based on Q-learning for an intersection traffic signal control
    Zhao, Xiaohua
    Li, Zhenlong
    Chen, Yangzhou
    Li, Yunchi
    Gaojishu Tongxin/Chinese High Technology Letters, 2007, 17 (05): : 498 - 502
  • [7] Optimal Control Inspired Q-Learning for Switched Linear Systems
    Chen, Hua
    Zheng, Linfang
    Zhang, Wei
    2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 4003 - 4010
  • [8] Optimal Tracking Control of Nonlinear Multiagent Systems Using Internal Reinforce Q-Learning
    Peng, Zhinan
    Luo, Rui
    Hu, Jiangping
    Shi, Kaibo
    Nguang, Sing Kiong
    Ghosh, Bijoy Kumar
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (08) : 4043 - 4055
  • [9] Optimal control using adaptive resonance theory and Q-learning
    Kiumarsi, Bahare
    AlQaudi, Bakur
    Modares, Hamidreza
    Lewis, Frank L.
    Levine, Daniel S.
    NEUROCOMPUTING, 2019, 361 : 119 - 125
  • [10] Optimal operational control for industrial processes based on Q-learning method
    Li, Jinna
    Gao, Xize
    Yuan, Decheng
    Fan, Jialu
    PROCEEDINGS OF THE 36TH CHINESE CONTROL CONFERENCE (CCC 2017), 2017, : 2562 - 2567