kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Safe multi-agent deep reinforcement learning for real-time decentralized control of inverter based renewable energy resources considering communication delay
State Grid Economic and Technological Research Institute Co., Ltd., Beijing 102209, China.
KTH, School of Electrical Engineering and Computer Science (EECS), Electrical Engineering, Electric Power and Energy Systems.ORCID iD: 0000-0003-0746-0221
School of Electrical & Electronic Engineering, North China Electric Power University, Beijing 102206, China.
KTH, School of Electrical Engineering and Computer Science (EECS), Electrical Engineering, Electric Power and Energy Systems.ORCID iD: 0000-0002-2793-9048
2023 (English)In: Applied Energy, ISSN 0306-2619, E-ISSN 1872-9118, Vol. 349, article id 121648Article in journal (Refereed) Published
Abstract [en]

The increasing penetration of distributed renewable energy resources brings a great challenge for real-time voltage security of distribution grids. The paper proposes a safe multi-agent deep reinforcement learning (MADRL) algorithm for real-time control of inverter-based Volt-Var control (VVC) in distribution grids considering communication delay to minimize the network power loss, while maintaining the nodal voltages in a safe range. The multi-agent VVC is modeled as a constrained Markov game, which is solved by the MADRL algorithm. In the training stage, the safety projection is added to the combined policy to analytically solve an action correction formulation to promote more efficient and safe exploration. In the real-time decision-making stage, a state synchronization block is designed to impute the data under the latest timestamp as the input of the agents deployed in a distributed manner, to avoid instability caused by communication delay. The simulation results show that the proposed algorithm performs well in safe exploration, and also achieves better performance under communication delay.

Place, publisher, year, edition, pages
Elsevier BV , 2023. Vol. 349, article id 121648
Keywords [en]
Communication delay, Decentralized control, Distribution grids, Inverter based renewable energy resources, Multi-agent reinforcement learning, Safe exploration, Voltage control
National Category
Computer Sciences Control Engineering
Identifiers
URN: urn:nbn:se:kth:diva-335312DOI: 10.1016/j.apenergy.2023.121648ISI: 001053289400001Scopus ID: 2-s2.0-85166643906OAI: oai:DiVA.org:kth-335312DiVA, id: diva2:1794248
Note

QC 20230905

Available from: 2023-09-05 Created: 2023-09-05 Last updated: 2023-09-11Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Zhang, MengfanXu, Qianwen

Search in DiVA

By author/editor
Zhang, MengfanXu, Qianwen
By organisation
Electric Power and Energy Systems
In the same journal
Applied Energy
Computer SciencesControl Engineering

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 35 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf