kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Logistics-involved service composition in a dynamic cloud manufacturing environment: A DDPG-based approach
Xidian Univ, Sch Mechanoelect Engn, Xian 710071, Peoples R China..
Changan Univ, Sch Elect & Control, Xian 710064, Peoples R China..
Beijing Inst Elect Syst Engn, State Key Lab Complex Prod Intelligent Mfg Syst T, Beijing 100854, Peoples R China..
Changan Univ, Sch Elect & Control, Xian 710064, Peoples R China..
Show others and affiliations
2022 (English)In: Robotics and Computer-Integrated Manufacturing, ISSN 0736-5845, E-ISSN 1879-2537, Vol. 76, p. 102323-, article id 102323Article in journal (Refereed) Published
Abstract [en]

Service composition as an important technique for combining multiple services to construct a value-added service is a major research issue in cloud manufacturing. Highly dynamic environments present great challenges to cloud manufacturing service composition (CMfg-SC). Most of previous studies employ heuristic algorithms to solve service composition issues in cloud manufacturing, which, however, are designed for specific problems and lack adaptability necessary to dynamic environment. Hence, CMfg-SC calls for new adaptive approaches. Recent advances in deep reinforcement learning (DRL) provide a new means for solving this issue. Based on DRL, we propose a Deep Deterministic Policy Gradient (DDPG)-based service composition approach to cloud manufacturing, with which optimal service composition solutions can be learned through repeated training. Performance of DDPG in solving CMfg-SC in both static and dynamic environments is examined. Results obtained with another DRL algorithm -Deep Q-Networks (DQN) and the traditional Ant Colony Optimization (ACO) are also presented. Comparison indicates that DDPG has better adaptability, robustness, and extensibility to dynamic environments than ACO, although ACO converges faster and its steady QoS value of the service composition solution is higher than that of DDPG by 0.997%. DDPG outperforms DQN in convergence speed and stability, and the QoS value of the service composition solution of DDPG is higher than that of DQN by 3.249%.

Place, publisher, year, edition, pages
Elsevier BV , 2022. Vol. 76, p. 102323-, article id 102323
Keywords [en]
Cloud manufacturing, service composition, deep reinforcement learning, deep deterministic policy gradient algorithm, Ant Colony Optimization
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-313518DOI: 10.1016/j.rcim.2022.102323ISI: 000791334600005Scopus ID: 2-s2.0-85124380192OAI: oai:DiVA.org:kth-313518DiVA, id: diva2:1665258
Note

QC 20220607

Available from: 2022-06-07 Created: 2022-06-07 Last updated: 2022-06-25Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Wang, Lihui

Search in DiVA

By author/editor
Wang, Lihui
By organisation
Sustainable Production Systems
In the same journal
Robotics and Computer-Integrated Manufacturing
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 45 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf