kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Flexible, non-parametric modeling using regularized neural networks
University of Gothenburg and Chalmers University of Technology, Gothenburg, Sweden.ORCID iD: 0000-0002-5926-0830
University of Gothenburg and Chalmers University of Technology, Gothenburg, Sweden.ORCID iD: 0000-0002-2550-3494
2022 (English)In: Computational statistics (Zeitschrift), ISSN 0943-4062, E-ISSN 1613-9658, Vol. 37, no 4, p. 2029-2047Article in journal (Refereed) Published
Abstract [en]

Non-parametric, additive models are able to capture complex data dependencies in a flexible, yet interpretable way. However, choosing the format of the additive components often requires non-trivial data exploration. Here, as an alternative, we propose PrAda-net, a one-hidden-layer neural network, trained with proximal gradient descent and adaptive lasso. PrAda-net automatically adjusts the size and architecture of the neural network to reflect the complexity and structure of the data. The compact network obtained by PrAda-net can be translated to additive model components, making it suitable for non-parametric statistical modelling with automatic model selection. We demonstrate PrAda-net on simulated data, where we compare the test error performance, variable importance and variable subset identification properties of PrAda-net to other lasso-based regularization approaches for neural networks. We also apply PrAda-net to the massive U.K. black smoke data set, to demonstrate how PrAda-net can be used to model complex and heterogeneous data with spatial and temporal components. In contrast to classical, statistical non-parametric approaches, PrAda-net requires no preliminary modeling to select the functional forms of the additive components, yet still results in an interpretable model representation. 

Place, publisher, year, edition, pages
Springer Nature , 2022. Vol. 37, no 4, p. 2029-2047
National Category
Probability Theory and Statistics Computational Mathematics
Identifiers
URN: urn:nbn:se:kth:diva-345870DOI: 10.1007/s00180-021-01190-4Scopus ID: 2-s2.0-85122443959OAI: oai:DiVA.org:kth-345870DiVA, id: diva2:1854136
Funder
Swedish Research Council, 2019-03686
Note

QC 20240429

Available from: 2024-04-24 Created: 2024-04-24 Last updated: 2024-04-29Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Allerbo, OskarJörnsten, Rebecka

Search in DiVA

By author/editor
Allerbo, OskarJörnsten, Rebecka
In the same journal
Computational statistics (Zeitschrift)
Probability Theory and StatisticsComputational Mathematics

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 116 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf