kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Remote Electrical Tilt Optimization via Safe Reinforcement Learning
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Decision and Control Systems (Automatic Control). Ericsson Res, Stockholm, Sweden..ORCID iD: 0000-0002-7668-0650
KTH. Ericsson Res, Stockholm, Sweden..ORCID iD: 0000-0001-7168-3169
Ericsson Res, Stockholm, Sweden..
LM Ericsson, Network Management Res Lab, Athlone, Ireland..
Show others and affiliations
2021 (English)In: 2021 IEEE wireless communications and networking conference (WCNC), Institute of Electrical and Electronics Engineers (IEEE) , 2021Conference paper, Published paper (Refereed)
Abstract [en]

Remote Electrical Tilt (RET) optimization is an efficient method for adjusting the vertical tilt angle of Base Stations (BSs) antennas in order to optimize Key Performance Indicators (KPIs) of the network. Reinforcement Learning (RL) provides a powerful framework for RET optimization because of its self-learning capabilities and adaptivity to environmental changes. However, an RL agent may execute unsafe actions during the course of its interaction, i.e., actions resulting in undesired network performance degradation. Since the reliability of services is critical for Mobile Network Operators (MNOs), the prospect of performance degradation has prohibited the real-world deployment of RL methods for RET optimization. In this work, we model the RET optimization problem in the Safe Reinforcement Learning (SRL) framework with the goal of learning a tilt control strategy providing performance improvement guarantees with respect to a safe baseline. We leverage a recent SRL method, namely Safe Policy Improvement through Baseline Bootstrapping (SPIBB), to learn an improved policy from an offline dataset of interactions collected by the safe baseline. Our experiments show that the proposed approach is able to learn a safe and improved tilt update policy, providing a higher degree of reliability and potential for real-world network deployment.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2021.
Series
IEEE Wireless Communications and Networking Conference, ISSN 1525-3511
National Category
Computer Sciences Control Engineering
Identifiers
URN: urn:nbn:se:kth:diva-304553DOI: 10.1109/WCNC49053.2021.9417363ISI: 000704226500136Scopus ID: 2-s2.0-85106427959OAI: oai:DiVA.org:kth-304553DiVA, id: diva2:1609634
Conference
IEEE Wireless Communications and Networking Conference (WCNC), MAR 29-APR 01, 2021, Nanjing, PEOPLES R CHINA
Note

Part of proceedings: ISBN 978-1-7281-9505-6, QC 20230117

Available from: 2021-11-09 Created: 2021-11-09 Last updated: 2023-01-17Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Vannella, FilippoIakovidis, Grigorios

Search in DiVA

By author/editor
Vannella, FilippoIakovidis, Grigorios
By organisation
Decision and Control Systems (Automatic Control)KTH
Computer SciencesControl Engineering

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 76 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf