Open this publication in new window or tab >>2024 (English)In: SIAM Journal on Applied Algebra and Geometry, E-ISSN 2470-6566, Vol. 8, no 2, p. 333-362Article in journal (Refereed) Published
Abstract [en]
We study the geometry of linear networks with one-dimensional convolutional layers. The function spaces of these networks can be identified with semialgebraic families of polynomials admitting sparse factorizations. We analyze the impact of the network’s architecture on the function space’s dimension, boundary, and singular points. We also describe the critical points of the network’s parameterization map. Furthermore, we study the optimization problem of training a network with the squared error loss. We prove that for architectures where all strides are larger than one and generic data, the nonzero critical points of that optimization problem are smooth interior points of the function space. This property is known to be false for dense linear networks and linear convolutional networks with stride one.
Keywords
neural network, critical points, semialgebraic set
National Category
Geometry
Research subject
Mathematics; Applied and Computational Mathematics; Computer Science
Identifiers
urn:nbn:se:kth:diva-374082 (URN)10.1137/23m1565504 (DOI)
Funder
Knut and Alice Wallenberg Foundation
2025-12-132025-12-132025-12-13