kth.sePublications KTH
Operational message
There are currently operational disruptions. Troubleshooting is in progress.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Safe Deep Reinforcement Learning for Renewable Energy Integrated Power System
KTH, School of Electrical Engineering and Computer Science (EECS), Electrical Engineering, Electric Power and Energy Systems.ORCID iD: 0000-0003-0746-0221
KTH, School of Electrical Engineering and Computer Science (EECS), Electrical Engineering, Electric Power and Energy Systems.ORCID iD: 0000-0002-2793-9048
2025 (English)In: Digitalisation of Local Energy Systems / [ed] Weiqi Hua; Xiao-Ping Zhang; David C.H. Wallom, Springer Nature , 2025, Vol. Part F989, p. 313-336Chapter in book (Other academic)
Abstract [en]

Deep reinforcement learning (DRL) is a promising solution for the coordination of power systems with high penetration of inverter-based renewable energy sources (RESs). Yet, when adopting the DRL-based control method, the safe and optimal operation of the system cannot be guaranteed at the same time, as the conventional DRL agent is not designed to solve the hard constraint problem. To address this challenge, a deep neural network (DNN) assisted projection-based DRL method for safe control of distribution grids is proposed in this section. First, a finite iteration projection algorithm is proposed to guarantee hard constraints by converting a non-convex optimization problem into a finite iteration problem. Next, a DNN-assisted projection method is proposed to accelerate the calculation of projection and achieve the practical implementation of hard constraints in the DRL problem. Finally, a DNN Projection embedded twin-delayed deep deterministic policy gradient (DPe-TD3) method is proposed to achieve optimal operation of distribution grids with guaranteed 100% safety of the distribution grid. The safety of the DRL training is guaranteed via the embedded Projection DNN in TD3 with participation in the gradient return process, which could smoothly and effectively project the DRL agent actions into the feasible area, thus guaranteeing the safety of data-driven control and the optimal operation at the same time. The case studies and comparisons are conducted in the IEEE 33 bus system to show the effectiveness of the proposed method.

Place, publisher, year, edition, pages
Springer Nature , 2025. Vol. Part F989, p. 313-336
National Category
Control Engineering
Identifiers
URN: urn:nbn:se:kth:diva-372595DOI: 10.1007/978-3-031-77833-9_11Scopus ID: 2-s2.0-105019185039OAI: oai:DiVA.org:kth-372595DiVA, id: diva2:2013111
Note

Part of ISBN 9783031778322, 9783031778339

QC 20251111

Available from: 2025-11-11 Created: 2025-11-11 Last updated: 2025-11-11Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Zhang, MengfanXu, Qianwen

Search in DiVA

By author/editor
Zhang, MengfanXu, Qianwen
By organisation
Electric Power and Energy Systems
Control Engineering

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 13 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf