Automatic voltage control considering demand response: Approximatively completed observed Markov decision process-based reinforcement learning scheme

被引:0
|
作者
Gu, Yaru [1 ]
Huang, Xueliang [1 ]
机构
[1] Southeast Univ, Sch Elect Engn, Nanjing, Peoples R China
关键词
Automatic voltage control; Partially observable system; Uncertainty; Differential increment incentive mechanism; Reinforcement learning algorithm; MODEL;
D O I
10.1016/j.ijepes.2024.110156
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
To fully utilize the voltage regulation capacity of flexible load and distributed generations (DGs), we propose a novel Approximatively Completed Observed Markov Decision Process-based (ACOMDP-based) Reinforcement Learning (RL) (namely, ACMRL) scheme for a multi-objective Automatic Voltage Control (AVC) problem considering Differential Increment Incentive Mechanism (DIIM)-based Incentive-Based Demand Response (IBDR). Firstly, we propose a DIIM to motivate high-flexibility consumers to achieve maximum potential in realtime voltage control while ensuring the best economy. Secondly, we characterize the multi-objective AVC problem as an ACOMDP model, transformed from the Partially Observable Markov Decision Process (POMDP) model, by introducing a novel hidden system state vector that incorporates the belief state, and the high confidence probability vector. The belief state and the high-confidence probability vector describe the probability distribution extracted from the historical observed state, portraying the precise state and the uncertainty existing in the state update process. Then, the ACOMDP block is inputted into the RL block, which adopts a modified underlying network architecture with the Asynchronous Advantage Actor-Critic (MA3C) algorithm embedded with the Shared Modular Policies(SMP) module. The MA3C-based RL block, characterized by enhanced communication efficiency, enables expedited generation of optimal decision-making actions even in the face of substantial uncertainty. Case studies are conducted in a practical district in Suzhou, China, and simulation results validate the superior performance of the proposed methodology.
引用
收藏
页数:13
相关论文
共 49 条
  • [21] A Continuous-Time Markov decision process-based resource allocation scheme in vehicular cloud for mobile video services
    Hou, Lu
    Zheng, Kan
    Chatzimisios, Periklis
    Feng, Yi
    COMPUTER COMMUNICATIONS, 2018, 118 : 140 - 147
  • [22] A homotopy-based reinforcement learning scheme to optimal control for Markov switched interconnected systems
    Liu, Jinxu
    Mi, Xuanrui
    Xia, Jianwei
    Su, Lei
    Shen, Hao
    JOURNAL OF CONTROL AND DECISION, 2024,
  • [23] Operation of Distributed Battery Considering Demand Response Using Deep Reinforcement Learning in Grid Edge Control
    Li, Wenying
    Tang, Ming
    Zhang, Xinzhen
    Gao, Danhui
    Wang, Jian
    ENERGIES, 2021, 14 (22)
  • [24] Online Learning Strategy Induction through Partially Observable Markov Decision Process-Based Cognitive Experience Model
    Gao, Huifan
    Ma, Biyang
    ELECTRONICS, 2024, 13 (19)
  • [25] Active flow control using deep reinforcement learning with time delays in Markov decision process and autoregressive policy
    Mao, Yiqian
    Zhong, Shan
    Yin, Hujun
    PHYSICS OF FLUIDS, 2022, 34 (05)
  • [26] Evaluation of Reinforcement Learning-Based False Data Injection Attack to Automatic Voltage Control
    Chen, Ying
    Huang, Shaowei
    Liu, Feng
    Wang, Zhisheng
    Sun, Xinwei
    IEEE TRANSACTIONS ON SMART GRID, 2019, 10 (02) : 2158 - 2169
  • [27] Cloud Job Scheduling Control Scheme Based on Gaussian Process Regression and Reinforcement Learning
    Peng, Zhiping
    Cui, Delong
    Xiong, Jianbin
    Xu, Bo
    Ma, Yuanjia
    Lin, Weiwei
    2016 IEEE 4TH INTERNATIONAL CONFERENCE ON FUTURE INTERNET OF THINGS AND CLOUD (FICLOUD 2016), 2016, : 278 - 286
  • [28] Coordinated Volt/VAR Control in Distribution Networks Considering Demand Response via Safe Deep Reinforcement Learning
    Hua, Dong
    Peng, Fei
    Liu, Suisheng
    Lin, Qinglin
    Fan, Jiahui
    Li, Qian
    ENERGIES, 2025, 18 (02)
  • [29] Reinforcement learning-based composite differential evolution for integrated demand response scheme in industrial microgrids
    Mughees, Neelam
    Jaffery, Mujtaba Hussain
    Mughees, Anam
    Ansari, Ejaz Ahmad
    Mughees, Abdullah
    APPLIED ENERGY, 2023, 342
  • [30] An IoT based Smart Irrigation Management System using Reinforcement Learning modeled through a Markov Decision Process
    Campoverde, Luis Miguel Samaniego
    Tropea, Mauro
    De Rango, Floriano
    PROCEEDINGS OF THE 2021 IEEE/ACM 25TH INTERNATIONAL SYMPOSIUM ON DISTRIBUTED SIMULATION AND REAL TIME APPLICATIONS (DS-RT 2021), 2021,