Automatic voltage control considering demand response: Approximatively completed observed Markov decision process-based reinforcement learning scheme

被引:0
|
作者
Gu, Yaru [1 ]
Huang, Xueliang [1 ]
机构
[1] Southeast Univ, Sch Elect Engn, Nanjing, Peoples R China
关键词
Automatic voltage control; Partially observable system; Uncertainty; Differential increment incentive mechanism; Reinforcement learning algorithm; MODEL;
D O I
10.1016/j.ijepes.2024.110156
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
To fully utilize the voltage regulation capacity of flexible load and distributed generations (DGs), we propose a novel Approximatively Completed Observed Markov Decision Process-based (ACOMDP-based) Reinforcement Learning (RL) (namely, ACMRL) scheme for a multi-objective Automatic Voltage Control (AVC) problem considering Differential Increment Incentive Mechanism (DIIM)-based Incentive-Based Demand Response (IBDR). Firstly, we propose a DIIM to motivate high-flexibility consumers to achieve maximum potential in realtime voltage control while ensuring the best economy. Secondly, we characterize the multi-objective AVC problem as an ACOMDP model, transformed from the Partially Observable Markov Decision Process (POMDP) model, by introducing a novel hidden system state vector that incorporates the belief state, and the high confidence probability vector. The belief state and the high-confidence probability vector describe the probability distribution extracted from the historical observed state, portraying the precise state and the uncertainty existing in the state update process. Then, the ACOMDP block is inputted into the RL block, which adopts a modified underlying network architecture with the Asynchronous Advantage Actor-Critic (MA3C) algorithm embedded with the Shared Modular Policies(SMP) module. The MA3C-based RL block, characterized by enhanced communication efficiency, enables expedited generation of optimal decision-making actions even in the face of substantial uncertainty. Case studies are conducted in a practical district in Suzhou, China, and simulation results validate the superior performance of the proposed methodology.
引用
收藏
页数:13
相关论文
共 49 条
  • [11] A Reinforcement-Learning-Based Secure Demand Response Scheme for Smart Grid System
    Kumari, Aparna
    Tanwar, Sudeep
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (03) : 2180 - 2191
  • [12] Automatic Voltage Control of Differential Power Grids Based on Transfer Learning and Deep Reinforcement Learning
    Wang, Tianjing
    Tang, Yong
    CSEE JOURNAL OF POWER AND ENERGY SYSTEMS, 2023, 9 (03): : 937 - 948
  • [13] A Markov decision process-based policy characterization approach for a stochastic inventory control problem with unreliable sourcing
    Ahiska, S. Sebnem
    Appaji, Samyuktha R.
    King, Russell E.
    Warsing, Donald P., Jr.
    INTERNATIONAL JOURNAL OF PRODUCTION ECONOMICS, 2013, 144 (02) : 485 - 496
  • [14] Reinforcement learning model for incentive-based integrated demand response considering demand-side coupling
    Hu, Rong
    Zhou, Kaile
    Yin, Hui
    ENERGY, 2024, 308
  • [15] A Reinforcement Learning-Based Markov-Decision Process (MDP) Implementation for SRAM FPGAs
    Ruan, Aiwu
    Shi, Aokai
    Qin, Liang
    Xu, Shiyang
    Zhao, Yifan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2020, 67 (10) : 2124 - 2128
  • [16] eNB Selection for Machine Type Communications Using Reinforcement Learning Based Markov Decision Process
    Liu, Yu-Jui
    Cheng, Shin-Ming
    Hsueh, Yu-Lin
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2017, 66 (12) : 11330 - 11338
  • [17] A Markov Decision Process Model for a Reinforcement Learning-based Autonomous Pedestrian Crossing Protocol
    El Hamdani, Sara
    Loudari, Salaheddine
    Novotny, Stanislav
    Bouchner, Petr
    Benamar, Nabil
    2021 3RD IEEE MIDDLE EAST AND NORTH AFRICA COMMUNICATIONS CONFERENCE (MENACOMM), 2021, : 147 - 151
  • [18] A reinforcement learning method based on an immune network adapted to a semi-Markov decision process
    Kogawa N.
    Obayashi M.
    Kobayashi K.
    Kuremoto T.
    Artificial Life and Robotics, 2009, 13 (2) : 538 - 542
  • [19] Poster Abstract: Data Efficient HVAC Control using Gaussian Process-based Reinforcement Learning
    An, Zhiyu
    Ding, Xianzhong
    Du, Wan
    PROCEEDINGS OF THE 21ST ACM CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS, SENSYS 2023, 2023, : 538 - 539
  • [20] Deep reinforcement learning based dynamic pricing for demand response considering market and supply constraints
    Fraija, Alejandro
    Henao, Nilson
    Agbossou, Kodjo
    Kelouwani, Sousso
    Fournier, Michael
    Nagarsheth, Shaival Hemant
    SMART ENERGY, 2024, 14