kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
A model-based deep reinforcement learning approach to the nonblocking coordination of modular supervisors of discrete event systems
School of Electro-Mechanical Engineering, Xidian University, Xi'an 710071, China.ORCID iD: 0000-0002-5988-0335
KTH, School of Industrial Engineering and Management (ITM), Machine Design (Dept.).ORCID iD: 0000-0003-4535-3849
KTH, School of Industrial Engineering and Management (ITM), Machine Design (Dept.).ORCID iD: 0000-0001-5703-5923
School of Electro-Mechanical Engineering, Xidian University, Xi'an 710071, China; Institute of Systems Engineering, Macau University of Science and Technology, Taipa, Macau.ORCID iD: 0000-0003-1547-5503
2023 (English)In: Information Sciences, ISSN 0020-0255, E-ISSN 1872-6291, Vol. 630, p. 305-321Article in journal (Refereed) Published
Abstract [en]

Modular supervisory control may lead to conflicts among the modular supervisors for large-scale discrete event systems. The existing methods for ensuring nonblocking control of modular supervisors either exploit favorable structures in the system model to guarantee the nonblocking property of modular supervisors or employ hierarchical model abstraction methods for reducing the computational complexity of designing a nonblocking coordinator. The nonblocking modular control problem is, in general, NP-hard. This study integrates supervisory control theory and a model-based deep reinforcement learning method to synthesize a nonblocking coordinator for the modular supervisors. The deep reinforcement learning method significantly reduces the computational complexity by avoiding the computation of synchronization of multiple modular supervisors and the plant models. The supervisory control function is approximated by the deep neural network instead of a large-sized finite automaton. Furthermore, the proposed model-based deep reinforcement learning method is more efficient than the standard deep Q network algorithm.

Place, publisher, year, edition, pages
Elsevier BV , 2023. Vol. 630, p. 305-321
Keywords [en]
Deep reinforcement learning; Discrete event system; Local modular control; Supervisory control theory
National Category
Control Engineering Embedded Systems Computer Systems
Research subject
Industrial Information and Control Systems
Identifiers
URN: urn:nbn:se:kth:diva-324252DOI: 10.1016/j.ins.2023.02.033ISI: 000944058400001Scopus ID: 2-s2.0-85148323580OAI: oai:DiVA.org:kth-324252DiVA, id: diva2:1739162
Funder
XPRES - Initiative for excellence in production research
Note

QC 20230404

Available from: 2023-02-23 Created: 2023-02-23 Last updated: 2025-02-21Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Tan, KaigeFeng, Lei

Search in DiVA

By author/editor
Yang, JunjunTan, KaigeFeng, LeiLi, Zhiwu
By organisation
Machine Design (Dept.)
In the same journal
Information Sciences
Control EngineeringEmbedded SystemsComputer Systems

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 119 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf