TY - GEN
T1 - Safe Optimal Control of Battery Energy Storage Systems via Hierarchical Deep Reinforcement Learning
AU - Selim, Alaa
AU - Mo, Huadong
AU - Pota, Hemanshu
AU - Dong, Daoyi
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Effective control of Battery Energy Storage Systems (BESSs) and household appliances is crucial for transitioning toward a sustainable and robust power grid. This paper presents a Hierarchical Reinforcement Learning (HRL) control framework, executed by Deep Reinforcement Learning (DRL) agent to achieve effective control of BESSs. The proposed HRL approach compartmentalizes the control problem into overarching strategic decisions and detailed executable actions. At the higher-level, we secure a set of actions that guarantee the safety of the BESS operations. Building upon this base, the next tier of our approach is dedicated to achieving optimal performance within the confines of this established safety set. We employ two cutting-edge HRL architectures to benchmark against our proposed HRL method. Simulation results indicate that our HRL model outperforms traditional DRL, other HRL techniques, and classical optimization methods like Quantum Delta-Potential-Well-based Particle Swarm Optimization. Our approach achieves the highest reward for the defined BESS challenge and reduces computational time to just 10.7 percent compared to the best-performing DRL agent. These results underscore the proposed HRL's promise as a scalable, efficient controller for both BESS and household utilities.
AB - Effective control of Battery Energy Storage Systems (BESSs) and household appliances is crucial for transitioning toward a sustainable and robust power grid. This paper presents a Hierarchical Reinforcement Learning (HRL) control framework, executed by Deep Reinforcement Learning (DRL) agent to achieve effective control of BESSs. The proposed HRL approach compartmentalizes the control problem into overarching strategic decisions and detailed executable actions. At the higher-level, we secure a set of actions that guarantee the safety of the BESS operations. Building upon this base, the next tier of our approach is dedicated to achieving optimal performance within the confines of this established safety set. We employ two cutting-edge HRL architectures to benchmark against our proposed HRL method. Simulation results indicate that our HRL model outperforms traditional DRL, other HRL techniques, and classical optimization methods like Quantum Delta-Potential-Well-based Particle Swarm Optimization. Our approach achieves the highest reward for the defined BESS challenge and reduces computational time to just 10.7 percent compared to the best-performing DRL agent. These results underscore the proposed HRL's promise as a scalable, efficient controller for both BESS and household utilities.
KW - Battery Energy Storage Systems
KW - Hierarchical Reinforcement Learning
KW - Load management
KW - Off-policy Correction
KW - Option-Critic Architecture
UR - https://www.scopus.com/pages/publications/85207662583
U2 - 10.1109/SEST61601.2024.10694032
DO - 10.1109/SEST61601.2024.10694032
M3 - Conference Paper
AN - SCOPUS:85207662583
T3 - 2024 International Conference on Smart Energy Systems and Technologies: Driving the Advances for Future Electrification, SEST 2024 - Proceedings
BT - 2024 International Conference on Smart Energy Systems and Technologies
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2024 International Conference on Smart Energy Systems and Technologies, SEST 2024
Y2 - 10 September 2024 through 12 September 2024
ER -