kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Graph attention networks and deep q-learning for service mesh optimization: a digital twinning approach
Karlstad University, Dep. of Mathematics and Computer Science, Karlstad, Sweden.
School of Electronics, Electrical Eng. and Computer Science, Queen's University Belfast, Belfast, United Kingdom.
Deggendorf Institute of Technology, Faculty of Computer Science, Deggendorf, Germany.
KTH, School of Engineering Sciences in Chemistry, Biotechnology and Health (CBH), Biomedical Engineering and Health Systems, Health Informatics and Logistics.ORCID iD: 0000-0002-1985-3690
2024 (English)In: ICC 2024 - IEEE International Conference on Communications, Institute of Electrical and Electronics Engineers (IEEE) , 2024, p. 2913-2918Conference paper, Published paper (Refereed)
Abstract [en]

In the realm of cloud native environments, Ku-bernetes has emerged as the de facto orchestration system for containers, and the service mesh architecture, with its interconnected microservices, has become increasingly prominent. Efficient scheduling and resource allocation for these microservices play a pivotal role in achieving high performance and maintaining system reliability. In this paper, we introduce a novel approach for container scheduling within Kubernetes clusters, leveraging Graph Attention Networks (GATs) for representation learning. Our proposed method captures the intricate dependencies among containers and services by constructing a representation graph. The deep Q-learning algorithm is then employed to optimize scheduling decisions, focusing on container-to-node placements, CPU request-response allocation, and adherence to node affinity and anti-affinity rules. Our experiments demonstrate that our GATs-based method outperforms traditional scheduling strategies, leading to enhanced resource utilization, reduced service latency, and improved overall system throughput. The insights gleaned from this study pave the way for a new frontier in cloud native performance optimization and offer tangible benefits to industries adopting microservice-based architectures.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2024. p. 2913-2918
Keywords [en]
component, formatting, insert, style, styling
National Category
Computer Sciences Computer Systems
Identifiers
URN: urn:nbn:se:kth:diva-353522DOI: 10.1109/ICC51166.2024.10622616Scopus ID: 2-s2.0-85202817543OAI: oai:DiVA.org:kth-353522DiVA, id: diva2:1899197
Conference
59th Annual IEEE International Conference on Communications, ICC 2024, June 9-13, 2024, Denver, United States of America
Note

Part of ISBN: 9781728190549

QC 20240926

Available from: 2024-09-19 Created: 2024-09-19 Last updated: 2024-09-26Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Boodaghian Asl, Arsineh

Search in DiVA

By author/editor
Boodaghian Asl, Arsineh
By organisation
Health Informatics and Logistics
Computer SciencesComputer Systems

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 35 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf