Automatic voltage control considering demand response: Approximatively completed observed Markov decision process-based reinforcement learning scheme

被引:0
|
作者
Gu, Yaru [1 ]
Huang, Xueliang [1 ]
机构
[1] Southeast Univ, Sch Elect Engn, Nanjing, Peoples R China
关键词
Automatic voltage control; Partially observable system; Uncertainty; Differential increment incentive mechanism; Reinforcement learning algorithm; MODEL;
D O I
10.1016/j.ijepes.2024.110156
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
To fully utilize the voltage regulation capacity of flexible load and distributed generations (DGs), we propose a novel Approximatively Completed Observed Markov Decision Process-based (ACOMDP-based) Reinforcement Learning (RL) (namely, ACMRL) scheme for a multi-objective Automatic Voltage Control (AVC) problem considering Differential Increment Incentive Mechanism (DIIM)-based Incentive-Based Demand Response (IBDR). Firstly, we propose a DIIM to motivate high-flexibility consumers to achieve maximum potential in realtime voltage control while ensuring the best economy. Secondly, we characterize the multi-objective AVC problem as an ACOMDP model, transformed from the Partially Observable Markov Decision Process (POMDP) model, by introducing a novel hidden system state vector that incorporates the belief state, and the high confidence probability vector. The belief state and the high-confidence probability vector describe the probability distribution extracted from the historical observed state, portraying the precise state and the uncertainty existing in the state update process. Then, the ACOMDP block is inputted into the RL block, which adopts a modified underlying network architecture with the Asynchronous Advantage Actor-Critic (MA3C) algorithm embedded with the Shared Modular Policies(SMP) module. The MA3C-based RL block, characterized by enhanced communication efficiency, enables expedited generation of optimal decision-making actions even in the face of substantial uncertainty. Case studies are conducted in a practical district in Suzhou, China, and simulation results validate the superior performance of the proposed methodology.
引用
收藏
页数:13
相关论文
共 49 条
  • [1] Markov Decision Process Framework for Control-Based Reinforcement Learning
    Lu Y.
    Squillante M.S.
    Wah Wu C.
    Performance Evaluation Review, 2023, 51 (02): : 39 - 41
  • [2] Demand response-based autonomous voltage control for transmission and active distribution networks using modified partially observed Markov decision process model
    Gu, Yaru
    Huang, Xueliang
    IET GENERATION TRANSMISSION & DISTRIBUTION, 2023, 17 (23) : 5155 - 5170
  • [3] Automatic Construction of Markov Decision Process Models for Multi-Agent Reinforcement Learning
    Young, Darrell L.
    Eccles, Chris
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS II, 2020, 11413
  • [4] Corrective Voltage Control Scheme Considering Demand Response and Stochastic Wind Power
    Rabiee, Abbas
    Soroudi, Alireza
    Mohammadi-ivatloo, Behnam
    Parniani, Mostafa
    IEEE TRANSACTIONS ON POWER SYSTEMS, 2014, 29 (06) : 2965 - 2973
  • [5] A Finite Horizon Markov Decision Process Based Reinforcement Learning Control of a Rapid Thermal Processing system
    Pradeep, D. John
    Noel, Mathew Mithra
    JOURNAL OF PROCESS CONTROL, 2018, 68 : 218 - 225
  • [6] A Teacher-Student Markov Decision Process-based Framework for Online Correctional Learning
    Lourenco, Ines
    Winqvist, Rebecka
    Rojas, Cristian R.
    Wahlberg, Bo
    2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 3456 - 3461
  • [7] Model-based Bayesian Reinforcement Learning in Factored Markov Decision Process
    Wu, Bo
    Feng, Yanpeng
    Zheng, Hongyan
    JOURNAL OF COMPUTERS, 2014, 9 (04) : 845 - 850
  • [8] Best equivalent hydrogen consumption control for fuel cell vehicle based on Markov decision process-based
    Fu J.-T.
    Fu Z.-M.
    Song S.-Z.
    Kongzhi Lilun Yu Yingyong/Control Theory and Applications, 2021, 38 (08): : 1219 - 1228
  • [9] Decision Optimization Model of Incentive Demand Response Based on Deep Reinforcement Learning
    Xu H.
    Lu J.
    Yang Z.
    Li Y.
    Lu J.
    Huang H.
    Dianli Xitong Zidonghua/Automation of Electric Power Systems, 2021, 45 (14): : 97 - 103
  • [10] Task Offloading with Power Control for Mobile Edge Computing Using Reinforcement Learning-Based Markov Decision Process
    Zhang, Bingxin
    Zhang, Guopeng
    Sun, Weice
    Yang, Kun
    MOBILE INFORMATION SYSTEMS, 2020, 2020