kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
State and parameter learning with PARIS particle Gibbs
CMAP, École Polytechnique, Institut Polytechnique de Paris, Palaiseau; Université de Bordeaux, CRCTB U4045, INSERM, Bordeaux, France; IHU Liryc, fondation Bordeaux Université, Pessac, France.
Samovar, Télécom SudParis, département CITI, TIPIC, Institut Polytechnique de Paris, Palaiseau.
LPSM, Sorbonne Université, UMR CNRS 8001, 4 Place Jussieu, 75005 Paris, 4 Place Jussieu.
CMAP, École Polytechnique, Institut Polytechnique de Paris, Palaiseau.
Show others and affiliations
2023 (English)In: Proceedings of the 40th International Conference on Machine Learning, ICML 2023, ML Research Press , 2023, p. 3625-3675Conference paper, Published paper (Refereed)
Abstract [en]

Non-linear state-space models, also known as general hidden Markov models (HMM), are ubiquitous in statistical machine learning, being the most classical generative models for serial data and sequences. Learning in HMM, either via Maximum Likelihood Estimation (MLE) or Markov Score Climbing (MSC) requires the estimation of the smoothing expectation of some additive functionals. Controlling the bias and the variance of this estimation is crucial to establish the convergence of learning algorithms. Our first contribution is to design a novel additive smoothing algorithm, the Parisian particle Gibbs (PPG) sampler, which can be viewed as a PARIS (Olsson & Westerborn, 2017) algorithm driven by conditional SMC moves, resulting in bias-reduced estimates of the targeted quantities. We substantiate the PPG algorithm with theoretical results, including new bounds on bias and variance as well as deviation inequalities. We then establish, in the learning context, and under standard assumptions, non-asymptotic bounds highlighting the value of bias reduction and the implicit Rao-Blackwellization of PPG. These are the first non-asymptotic results of this kind in this setting. We illustrate our theoretical results with numerical experiments supporting our claims.

Place, publisher, year, edition, pages
ML Research Press , 2023. p. 3625-3675
National Category
Control Engineering
Identifiers
URN: urn:nbn:se:kth:diva-350075Scopus ID: 2-s2.0-85172211590OAI: oai:DiVA.org:kth-350075DiVA, id: diva2:1887373
Conference
40th International Conference on Machine Learning, ICML 2023, Honolulu, United States of America, Jul 23 2023 - Jul 29 2023
Note

QC 20240807

Available from: 2024-08-07 Created: 2024-08-07 Last updated: 2024-08-07Bibliographically approved

Open Access in DiVA

No full text in DiVA

Scopus

Authority records

Olsson, Jimmy

Search in DiVA

By author/editor
Olsson, Jimmy
By organisation
Mathematics (Dept.)
Control Engineering

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 42 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf