Interpretable Deep Reinforcement Learning for Optimizing Heterogeneous Energy Storage Systems

被引:2
|
作者
Xiong, Luolin [1 ,2 ]
Tang, Yang [1 ,2 ]
Liu, Chensheng [1 ,2 ]
Mao, Shuai [3 ]
Meng, Ke [4 ]
Dong, Zhaoyang [5 ]
Qian, Feng [1 ,2 ]
机构
[1] East China Univ Sci & Technol, Key Lab Smart Mfg Energy Chem Proc, Minist Educ, Shanghai 200237, Peoples R China
[2] East China Univ Sci & Technol, Engn Res Ctr Proc Syst Engn, Minist Educ, Shanghai 200237, Peoples R China
[3] Nantong Univ, Dept Elect Engn, Nantong 226019, Peoples R China
[4] Univ New South Wales, Sch Elect Engn & Telecommun, Sydney, NSW 2052, Australia
[5] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
基金
中国国家自然科学基金;
关键词
Heterogeneous energy storage systems; deep reinforcement learning; pre-hoc interpretability;
D O I
10.1109/TCSI.2023.3340026
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Energy storage systems (ESS) are pivotal component in the energy market, serving as both energy suppliers and consumers. ESS operators can reap benefits from energy arbitrage by optimizing operations of storage equipment. To further enhance ESS flexibility within the energy market and improve renewable energy utilization, a heterogeneous photovoltaic-ESS (PV-ESS) is proposed, which leverages the unique characteristics of battery energy storage (BES) and hydrogen energy storage (HES). For scheduling tasks of the heterogeneous PV-ESS, a practical cost function plays a crucial role in guiding operator's strategies to maximize benefits. We develop a comprehensive cost function that takes into account degradation, capital, and operation/maintenance costs to reflect real-world scenarios. Moreover, while numerous methods excel in optimizing ESS energy arbitrage, they often rely on black-box models with opaque decision-making processes, limiting practical applicability. To overcome this limitation and enable explainable scheduling strategies, a prototype-based policy network with inherent interpretability is introduced. This network employs human-designed prototypes to guide decision-making by comparing similarities between prototypical situations and encountered situations, which allows for naturally explained scheduling strategies. Comparative results across four distinct cases demonstrate the effectiveness and practicality of our proposed pre-hoc interpretable optimization method when contrasted with black-box models.
引用
收藏
页码:910 / 921
页数:12
相关论文
共 50 条
  • [1] On Optimizing Operational Efficiency in Storage Systems via Deep Reinforcement Learning
    Srinivasa, Sunil
    Kathalagiri, Girish
    Varanasi, Julu Subramanyam
    Quintela, Luis Carlos
    Charafeddine, Mohamad
    Lee, Chi-Hoon
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2018, PT III, 2019, 11053 : 238 - 253
  • [2] Reinforcement Learning Techniques in Optimizing Energy Systems
    Stavrev, Stefan
    Ginchev, Dimitar
    [J]. ELECTRONICS, 2024, 13 (08)
  • [3] Deep Reinforcement Learning for Hybrid Energy Storage Systems: Balancing Lead and Hydrogen Storage
    Desportes, Louis
    Fijalkow, Inbar
    Andry, Pierre
    [J]. ENERGIES, 2021, 14 (15)
  • [4] Optimizing Automated Trading Systems with Deep Reinforcement Learning
    Tran, Minh
    Pham-Hi, Duc
    Bui, Marc
    [J]. ALGORITHMS, 2023, 16 (01)
  • [5] Deep Reinforcement Learning for Online Scheduling of Photovoltaic Systems with Battery Energy Storage Systems
    Li Y.
    Wu J.
    Pan Y.
    [J]. Intelligent and Converged Networks, 2024, 5 (01): : 28 - 41
  • [6] Towards Interpretable Deep Reinforcement Learning Models via Inverse Reinforcement Learning
    Xie, Yuansheng
    Vosoughi, Soroush
    Hassanpour, Saeed
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 5067 - 5074
  • [7] Energy-aware Scheduling of Jobs in Heterogeneous Cluster Systems Using Deep Reinforcement Learning
    Esmaili, Amirhossein
    Pedram, Massoud
    [J]. PROCEEDINGS OF THE TWENTYFIRST INTERNATIONAL SYMPOSIUM ON QUALITY ELECTRONIC DESIGN (ISQED 2020), 2020, : 426 - 431
  • [8] Optimizing Secrecy Energy Efficiency in RIS-assisted MISO systems using Deep Reinforcement Learning
    Razaq, Mian Muaz
    Song, Huanhuan
    Peng, Limei
    Ho, Pin-Han
    [J]. COMPUTER COMMUNICATIONS, 2024, 217 : 126 - 133
  • [9] Deep Reinforcement Learning based Relay Selection for SWIPT Systems with Data Buffer and Energy Storage
    Quan, Jianping
    Xu, Peng
    Luo, Chenghong
    Huang, Chong
    Chen, Gaojie
    [J]. 2022 IEEE 96TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-FALL), 2022,
  • [10] Optimal planning of hybrid energy storage systems using curtailed renewable energy through deep reinforcement learning
    Kang, Dongju
    Kang, Doeun
    Hwangbo, Sumin
    Niaz, Haider
    Lee, Won Bo
    Liu, J. Jay
    Na, Jonggeol
    [J]. ENERGY, 2023, 284