kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Model-Based Reinforcement Learning for Cavity Filter Tuning
Ericsson GAIA, Sweden.
Qualcomm, Sweden.
Ericsson GAIA, Sweden.
Ericsson GAIA, Sweden.
Show others and affiliations
2023 (English)In: Proceedings of the 5th Annual Learning for Dynamics and Control Conference, L4DC 2023, ML Research Press , 2023, p. 1297-1307Conference paper, Published paper (Refereed)
Abstract [en]

The ongoing development of telecommunication systems like 5G has led to an increase in demand of well calibrated base transceiver station (BTS) components. A pivotal component of every BTS is cavity filters, which provide a sharp frequency characteristic to select a particular band of interest and reject the rest. Unfortunately, their characteristics in combination with manufacturing tolerances make them difficult for mass production and often lead to costly manual post-production fine tuning. To address this, numerous approaches have been proposed to automate the tuning process. One particularly promising one, that has emerged in the past few years, is to use model free reinforcement learning (MFRL); however, the agents are not sample efficient. This poses a serious bottleneck, as utilising complex simulators or training with real filters is prohibitively time demanding. This work advocates for the usage of model based reinforcement learning (MBRL) and showcases how its utilisation can significantly decrease sample complexity, while maintaining similar levels of success rate. More specifically, we propose an improvement over a state-of-the-art (SoTA) MBRL algorithm, namely the Dreamer algorithm. This improvement can serve as a template for applications in other similar, high-dimensional non-image data problems. We carry experiments on two complex filter types, and show that our novel modification on the Dreamer architecture reduces sample complexity by a factor of 4 and 10, respectively. Our findings pioneer the usage of MBRL which paves the way for utilising more precise and accurate simulators which was previously prohibitively time demanding.

Place, publisher, year, edition, pages
ML Research Press , 2023. p. 1297-1307
Keywords [en]
Model Based Reinforcement Learning, Reinforcement Learning, Telecommunication
National Category
Computer Sciences Signal Processing
Identifiers
URN: urn:nbn:se:kth:diva-338028ISI: 001221742900099Scopus ID: 2-s2.0-85172912389OAI: oai:DiVA.org:kth-338028DiVA, id: diva2:1804631
Conference
5th Annual Conference on Learning for Dynamics and Control, L4DC 2023, Philadelphia, United States of America, Jun 16 2023 - Jun 15 2023
Note

QC 20231013

Available from: 2023-10-13 Created: 2023-10-13 Last updated: 2024-09-05Bibliographically approved

Open Access in DiVA

No full text in DiVA

Scopus

Authority records

Ögren, Petter

Search in DiVA

By author/editor
Ögren, Petter
By organisation
Robotics, Perception and Learning, RPL
Computer SciencesSignal Processing

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 77 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf