kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Noisy Recurrent Neural Networks
Nordita SU.
2021 (English)In: Advances in Neural Information Processing Systems, Neural information processing systems foundation , 2021, p. 5124-5137Conference paper, Published paper (Refereed)
Abstract [en]

We provide a general framework for studying recurrent neural networks (RNNs) trained by injecting noise into hidden states. Specifically, we consider RNNs that can be viewed as discretizations of stochastic differential equations driven by input data. This framework allows us to study the implicit regularization effect of general noise injection schemes by deriving an approximate explicit regularizer in the small noise regime. We find that, under reasonable assumptions, this implicit regularization promotes flatter minima; it biases towards models with more stable dynamics; and, in classification tasks, it favors models with larger classification margin. Sufficient conditions for global stability are obtained, highlighting the phenomenon of stochastic stabilization, where noise injection can improve stability during training. Our theory is supported by empirical results which demonstrate that the RNNs have improved robustness with respect to various input perturbations.

Place, publisher, year, edition, pages
Neural information processing systems foundation , 2021. p. 5124-5137
Keywords [en]
Differential equations, Stabilization, Stochastic systems, Classification tasks, Discretizations, General noise, Hidden state, Injection schemes, Input datas, Noise injection, Regularisation, Regularizer, Stochastic differential equations, Recurrent neural networks
National Category
Computational Mathematics
Identifiers
URN: urn:nbn:se:kth:diva-316183ISI: 000922928400016Scopus ID: 2-s2.0-85131541364OAI: oai:DiVA.org:kth-316183DiVA, id: diva2:1699212
Conference
35th Conference on Neural Information Processing Systems, NeurIPS 2021, 6 December - 14 December, 2021, Virtual/Online
Note

Part of proceedings ISBN 9781713845393

QC 20230921

Available from: 2022-09-27 Created: 2022-09-27 Last updated: 2023-09-21Bibliographically approved

Open Access in DiVA

No full text in DiVA

Scopus
Computational Mathematics

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 24 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf