kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Learning Optimal Scheduling Policy for Remote State Estimation under Uncertain Channel Condition
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Decision and Control Systems (Automatic Control).ORCID iD: 0000-0001-9940-5929
Show others and affiliations
2020 (English)In: IEEE Transactions on Control of Network Systems, E-ISSN 2325-5870, Vol. 7, no 2, p. 579-591Article in journal (Refereed) Published
Abstract [en]

We consider optimal sensor scheduling with unknown communication channel statistics. We formulate two types of scheduling problems with the communication rate being a soft or hard constraint, respectively. We first present some structural results on the optimal scheduling policy using dynamic programming and assuming that the channel statistics is known. We prove that the Q-factor is monotonic and submodular, which leads to threshold-like structures in both problems. Then we develop a stochastic approximation and parameter learning frameworks to deal with the two scheduling problems with unknown channel statistics. We utilize their structures to design specialized learning algorithms. We prove the convergence of these algorithms. Performance improvement compared with the standard Q-learning algorithm is shown through numerical examples, which also discuss an alternative method based on recursive estimation of the channel quality. 

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2020. Vol. 7, no 2, p. 579-591
Keywords [en]
learning algorithm, scheduling, State estimation, threshold structure, Dynamic programming, Numerical methods, Q factor measurement, Reinforcement learning, Scheduling algorithms, Stochastic systems, Channel conditions, Communication rate, Optimal scheduling, Parameter learning, Q-learning algorithms, Recursive estimation, Remote state estimations, Stochastic approximations, Learning algorithms
National Category
Control Engineering
Identifiers
URN: urn:nbn:se:kth:diva-268453DOI: 10.1109/TCNS.2019.2959162ISI: 000549872800005Scopus ID: 2-s2.0-85076396692OAI: oai:DiVA.org:kth-268453DiVA, id: diva2:1422907
Note

QC 20200409

Available from: 2020-04-09 Created: 2020-04-09 Last updated: 2022-06-26Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Johansson, Karl H.

Search in DiVA

By author/editor
Johansson, Karl H.
By organisation
Decision and Control Systems (Automatic Control)
In the same journal
IEEE Transactions on Control of Network Systems
Control Engineering

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 31 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf