Open this publication in new window or tab >> (English)Manuscript (preprint) (Other academic)
Abstract [en]
Large deviations for stochastic approximations is a well-studied field that yields convergence properties for many useful algorithms in statistics, machine learning and statistical physics. In this article, we prove, under certain assumptions, a large deviation principle for a stochastic approximation with state-dependent Markovian noise and with decreasing step size. Common algorithms that satisfy these conditions include stochastic gradient descent, persistent contrastive divergence and the Wang-Landau algorithm. The proof is based don't he weak convergence approach to the theory of large deviations and uses a representation formula to rewrite the problem into a stochastic control problem. The resulting rate function is an action potential over a local rate function that is the Fenchel-Legendre transform of a limiting Hamiltonian.
National Category
Probability Theory and Statistics
Research subject
Applied and Computational Mathematics, Mathematical Statistics
Identifiers
urn:nbn:se:kth:diva-337358 (URN)
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note
QC 20231002
2023-10-022023-10-022023-10-02Bibliographically approved