Abstract
This paper proposes a safe reinforcement learning (RL)-based transient stability emergency control (TSEC) method for islanded microgrids. RL requires extensive interaction with the environment to learn control strategies, hence, a data-driven approach is used as a substitute for time-consuming time-domain simulation calculations. Deep sigma point processes (DSPP), which is a Gaussian process model, is utilized to predict the normal distribution of transient stability of microgrids and to construct a transient stability chance constraint. Reward-constrained policy optimization (RCPO) can simultaneously achieve objective prediction, policy learning, and constraint cost coefficient update across multiple timescales. RCPO interacts with the DSPP-based microgrid environment through a multi-process parallel manner, greatly increasing the training speed. Case studies on a real islanded microgrid demonstrate that the proposed method can efficiently and quickly obtain the optimal emergency control strategy while adhering to all hard constraints.
| Original language | American English |
|---|---|
| Pages (from-to) | 3432-3444 |
| Number of pages | 13 |
| Journal | IEEE Transactions on Smart Grid |
| Volume | 16 |
| Issue number | 4 |
| DOIs | |
| State | Published - 2025 |
NLR Publication Number
- NREL/JA-5D00-93140
Keywords
- chance constraint
- microgrid
- PVs
- safe reinforcement learning
- stability control
- topology change