Uncertainty-Aware Transient Stability-Constrained Preventive Redispatch: A Distributional Reinforcement Learning Approach

IEEE Transactions on Power Systems - Tập 40 Số 2 - Trang 1295-1308 - 2025
Zhengcheng Wang1, Fei Teng2, Yanzhen Zhou1, Qinglai Guo1, Hongbin Sun1
1State Key Laboratory of Power Systems, Department of Electrical Engineering, Tsinghua University, Beijing, China
2Department of Electrical and Electronic Engineering, Imperial College London, London, U.K.

Tóm tắt

Transient stability-constrained preventive redispatch plays a crucial role in ensuring power system security and stability. Since redispatch strategies need to simultaneously satisfy complex transient constraints and the economic need, model-based formulation and optimization become extremely challenging. In addition, the increasing uncertainty and variability introduced by renewable sources start to drive the system stability consideration from deterministic to probabilistic, which further exaggerates the complexity. In this paper, a Graph neural network guided Distributional Deep Reinforcement Learning (GD2RL) method is proposed, for the first time, to solve the uncertainty-aware transient stability-constrained preventive redispatch problem. First, a graph neural network-based transient simulator is trained by supervised learning to efficiently generate post-contingency rotor angle curves with the steady-state and contingency as inputs, which serves as a feature extractor for operating states and a surrogate time-domain simulator during the environment interaction for reinforcement learning. Distributional deep reinforcement learning with explicit uncertainty distribution of system operational conditions is then applied to generate the redispatch strategy to balance the user-specified probabilistic stability performance and economy preferences. The full distribution of the post-redispatch transient stability index is directly provided as the output. Case studies on the modified New England 39-bus system validate the proposed method.

Từ khóa

#Uncertainty #transient stability #preventive redispatch #distributional reinforcement learning #graph neural network

Tài liệu tham khảo

10.1109/TPWRS.2003.818708

10.1109/TPWRS.2004.825981

10.1049/iet-gtd.2017.0345

10.1109/59.867137

10.1109/TPWRS.2003.814856

10.1109/TPWRS.2003.818708

Zeng, 2023, A distributed deep reinforcement learning-based approach for fast preventive control considering transient stability constraints, CSEE J. Power Energy Syst., 9, 197

10.1049/iet-gtd.2014.0263

10.1016/j.epsr.2017.10.007

10.17775/CSEEJPES.2020.04780

10.1109/TII.2021.3072594

10.1109/TPWRS.2020.2983477

10.35833/MPCE.2020.000608

10.17775/CSEEJPES.2022.05030

10.1109/PESGM40551.2019.8973866

10.1109/TPWRS.2017.2699678

10.1109/TSTE.2016.2520481

10.1109/TPWRS.2023.3270800

10.1007/978-1-4615-4319-0

Barth-Maron, 2018, Distributed distributional deterministic policy gradients

Gilmer, 2017, Neural message passing for quantum chemistry, Proc. Int. Conf. Mach. Learn., 1263

Wang, 2021, Transient stability assessment of power system considering topological change: A message passing neural network-based approach, Proc. CSEE, 41, 2341

10.1038/nature14539

10.1109/tnn.1998.712192

10.1126/science.153.3731.34

Bellemare, 2017, A distributional perspective on reinforcement learning, Proc. 34th Int. Conf. Mach. Learn., 449

Singh, Sample-based distributional policy gradient, Proc. 4th Annu. Learn. Dyn. Control Conf., 676

Fujimoto, Addressing function approximation error in actor-critic methods, Proc. Mach. Learn. Res., 1587

10.1109/TPWRS.2023.3248293

Kingma, 2014, Adam: A method for stochastic optimization

10.1109/ICNN.1995.488968