Automatic voltage control considering demand response: Approximatively completed observed Markov decision process-based reinforcement learning scheme

被引:0
|
作者
Gu, Yaru [1 ]
Huang, Xueliang [1 ]
机构
[1] Southeast Univ, Sch Elect Engn, Nanjing, Peoples R China
关键词
Automatic voltage control; Partially observable system; Uncertainty; Differential increment incentive mechanism; Reinforcement learning algorithm; MODEL;
D O I
10.1016/j.ijepes.2024.110156
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
To fully utilize the voltage regulation capacity of flexible load and distributed generations (DGs), we propose a novel Approximatively Completed Observed Markov Decision Process-based (ACOMDP-based) Reinforcement Learning (RL) (namely, ACMRL) scheme for a multi-objective Automatic Voltage Control (AVC) problem considering Differential Increment Incentive Mechanism (DIIM)-based Incentive-Based Demand Response (IBDR). Firstly, we propose a DIIM to motivate high-flexibility consumers to achieve maximum potential in realtime voltage control while ensuring the best economy. Secondly, we characterize the multi-objective AVC problem as an ACOMDP model, transformed from the Partially Observable Markov Decision Process (POMDP) model, by introducing a novel hidden system state vector that incorporates the belief state, and the high confidence probability vector. The belief state and the high-confidence probability vector describe the probability distribution extracted from the historical observed state, portraying the precise state and the uncertainty existing in the state update process. Then, the ACOMDP block is inputted into the RL block, which adopts a modified underlying network architecture with the Asynchronous Advantage Actor-Critic (MA3C) algorithm embedded with the Shared Modular Policies(SMP) module. The MA3C-based RL block, characterized by enhanced communication efficiency, enables expedited generation of optimal decision-making actions even in the face of substantial uncertainty. Case studies are conducted in a practical district in Suzhou, China, and simulation results validate the superior performance of the proposed methodology.
引用
收藏
页数:13
相关论文
共 49 条
  • [41] Many-objective charging optimization for electric vehicles considering demand response and multi-uncertainties based on Markov chain and information gap decision theory
    Yan, Qingyou
    Lin, Hongyu
    Li, Jinmeng
    Ai, Xingbei
    Shi, Mengshu
    Zhang, Meijuan
    De Gejirifu
    SUSTAINABLE CITIES AND SOCIETY, 2022, 78
  • [42] Lessons Learned from Data-Driven Building Control Experiments: Contrasting Gaussian Process-based MPC, Bilevel DeePC, and Deep Reinforcement Learning
    Di Natale, Loris
    Lian, Yingzhao
    Maddalena, Emilio T.
    Shi, Jicheng
    Jones, Colin N.
    2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 1111 - 1117
  • [43] Reinforcement learning-based demand response strategy for thermal energy storage air-conditioning system considering room temperature and humidity setpoints
    Li, Zeyang
    Meng, Qinglong
    Wei, Ying'an
    Yan, Xiuying
    Lei, Yu
    Wu, Xiao
    Liu, Jiali
    Wang, Liqiang
    JOURNAL OF ENERGY STORAGE, 2023, 72
  • [44] A semi-Markov decision process-based joint call admission control for inter-RAT cell re-selection in next generation wireless networks
    Carvalho, Glaucio H. S.
    Woungang, Isaac
    Anpalagan, Alagan
    Coutinho, Rodolfo W. L.
    Costa, Joao C. W. A.
    COMPUTER NETWORKS, 2013, 57 (17) : 3545 - 3562
  • [45] Voltage unbalance mitigation with space vector-based PN-sequence control scheme of inverter-interfaced DGs considering demand-side management capability
    Ali, Nada
    Kumar, Deepak
    ELECTRICAL ENGINEERING, 2023, 105 (02) : 883 - 903
  • [46] Voltage unbalance mitigation with space vector-based PN-sequence control scheme of inverter-interfaced DGs considering demand-side management capability
    Nada Ali
    Deepak Kumar
    Electrical Engineering, 2023, 105 : 883 - 903
  • [47] Integration Scheme for Economic Load Dispatching and Optimization Control in Coal-Fired Plants Based on Sparse Gaussian Process Model and Deep Reinforcement Learning
    Dai, Bangwu
    Chang, Yuqing
    Wang, Fuli
    Chu, Fei
    Song, Shengjun
    INDUSTRIAL & ENGINEERING CHEMISTRY RESEARCH, 2023, 62 (39) : 16025 - 16036
  • [48] Dynamic energy management and control of networked microgrids based on load to grid services and incentive-based demand response programs: A multi-agent deep reinforcement learning approach
    Seylab, Masoumeh Rezazadeh
    Naderi, Mehdi S.
    Gharehpetian, Gevork B.
    SUSTAINABLE CITIES AND SOCIETY, 2024, 117
  • [49] A Policy optimization-based Deep Reinforcement Learning method for data-driven output voltage control of grid connected solid oxide fuel cell considering operation constraints
    Zeng, Shunqi
    Huang, Chunyan
    Wang, Fei
    Li, Xin
    Chen, Minghui
    ENERGY REPORTS, 2023, 10 : 1161 - 1168