kth.sePublications KTH
Change search
Link to record
Permanent link

Direct link
Publications (8 of 8) Show all publications
Marchetti, G. L., Shahverdi, V., Mereta, S., Trager, M. & Kohn, K. (2025). Algebra Unveils Deep Learning - An Invitation to Neuroalgebraic Geometry. In: : . Paper presented at International Conference on Machine Learning (ICML).
Open this publication in new window or tab >>Algebra Unveils Deep Learning - An Invitation to Neuroalgebraic Geometry
Show others...
2025 (English)Conference paper, Published paper (Refereed)
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-366669 (URN)
Conference
International Conference on Machine Learning (ICML)
Note

QC 20250806

Available from: 2025-07-08 Created: 2025-07-08 Last updated: 2025-12-13Bibliographically approved
Shahverdi, V. (2025). Algebraic complexity and neurovariety of linear convolutional networks. Mathematica, 17(1), Article ID 2.
Open this publication in new window or tab >>Algebraic complexity and neurovariety of linear convolutional networks
2025 (English)In: Mathematica, ISSN 1844-6094, E-ISSN 2066-7752, Vol. 17, no 1, article id 2Article in journal (Refereed) Published
Abstract [en]

In this paper, we study linear convolutional networks with one-dimensional filters and arbitrary strides. The neuromanifold of such a network is a semialgebraic set, represented by a space of polynomials admitting specific factorizations. Introducing a recursive algorithm, we generate polynomial equations whose common zero locus corresponds to the Zariski closure of the corresponding neuromanifold. Furthermore, we explore the algebraic complexity of training these networks employing tools from metric algebraic geometry. Our findings reveal that the number of all complex critical points in the optimization of such a network is equal to the generic Euclidean distance degree of a Segre variety. Notably, this count significantly surpasses the number of critical points encountered in the training of a fully connected linear network with the same number of parameters.

Place, publisher, year, edition, pages
Springer Nature, 2025
Keywords
Function space description of neural networks, Linear networks, Euclidean distance degree, Critical points
National Category
Mathematical sciences
Identifiers
urn:nbn:se:kth:diva-367866 (URN)10.1007/s44426-025-00002-2 (DOI)001504294400002 ()2-s2.0-105006453573 (Scopus ID)
Note

QC 20250801

Available from: 2025-08-01 Created: 2025-08-01 Last updated: 2025-12-13Bibliographically approved
Kohn, K., Sattelberger, A.-L. & Shahverdi, V. (2025). Geometry of Linear Neural Networks: Equivariance and Invariance under Permutation Groups. SIAM Journal on Matrix Analysis and Applications, 1378-1415
Open this publication in new window or tab >>Geometry of Linear Neural Networks: Equivariance and Invariance under Permutation Groups
2025 (English)In: SIAM Journal on Matrix Analysis and Applications, ISSN 0895-4798, E-ISSN 1095-7162, p. 1378-1415Article in journal (Refereed) Published
National Category
Algebra and Logic
Identifiers
urn:nbn:se:kth:diva-358453 (URN)
Note

QC 20250117

Available from: 2025-01-17 Created: 2025-01-17 Last updated: 2025-12-13Bibliographically approved
Shahverdi, V., Marchetti, G. L. & Kohn, K. (2025). On the Geometry and Optimization of Polynomial Convolutional Networks. In: Proceedings of the 28th International Conference on Artificial Intelligence and Statistics, AISTATS 2025: . Paper presented at The 28th International Conference on Artificial Intelligence and Statistics (AISTATS), Thailand, May 3rd - May 5th, 2025 (pp. 604-612). ML Research Press, 258
Open this publication in new window or tab >>On the Geometry and Optimization of Polynomial Convolutional Networks
2025 (English)In: Proceedings of the 28th International Conference on Artificial Intelligence and Statistics, AISTATS 2025, ML Research Press , 2025, Vol. 258, p. 604-612Conference paper, Published paper (Refereed)
Abstract [en]

We study convolutional neural networks with monomial activation functions. Specifically, we prove that their parameterization map is regular and is an isomorphism almost everywhere, up to rescaling the filters. By leveraging on tools from algebraic geometry, we explore the geometric properties of the image in function space of this map - typically referred to as neuromanifold. In particular, we compute the dimension and the degree of the neuromanifold, which measure the expressivity of the model, and describe its singularities. Moreover, for a generic large dataset, we derive an explicit formula that quantifies the number of critical points arising in the optimization of a regression loss.

Place, publisher, year, edition, pages
ML Research Press, 2025
National Category
Computer Sciences Computer graphics and computer vision
Identifiers
urn:nbn:se:kth:diva-358449 (URN)2-s2.0-105014321299 (Scopus ID)
Conference
The 28th International Conference on Artificial Intelligence and Statistics (AISTATS), Thailand, May 3rd - May 5th, 2025
Note

QC 20250117

Available from: 2025-01-17 Created: 2025-01-17 Last updated: 2025-12-13Bibliographically approved
Kohn, K., Montúfar, G., Shahverdi, V. & Trager, M. (2024). Function Space and Critical Points of Linear Convolutional Networks. SIAM Journal on Applied Algebra and Geometry, 8(2), 333-362
Open this publication in new window or tab >>Function Space and Critical Points of Linear Convolutional Networks
2024 (English)In: SIAM Journal on Applied Algebra and Geometry, E-ISSN 2470-6566, Vol. 8, no 2, p. 333-362Article in journal (Refereed) Published
Abstract [en]

We study the geometry of linear networks with one-dimensional convolutional layers. The function spaces of these networks can be identified with semialgebraic families of polynomials admitting sparse factorizations. We analyze the impact of the network’s architecture on the function space’s dimension, boundary, and singular points. We also describe the critical points of the network’s parameterization map. Furthermore, we study the optimization problem of training a network with the squared error loss. We prove that for architectures where all strides are larger than one and generic data, the nonzero critical points of that optimization problem are smooth interior points of the function space. This property is known to be false for dense linear networks and linear convolutional networks with stride one.

Keywords
neural network, critical points, semialgebraic set
National Category
Geometry
Research subject
Mathematics; Applied and Computational Mathematics; Computer Science
Identifiers
urn:nbn:se:kth:diva-374082 (URN)10.1137/23m1565504 (DOI)
Funder
Knut and Alice Wallenberg Foundation
Available from: 2025-12-13 Created: 2025-12-13 Last updated: 2025-12-13
Arfaeezarandi, S. F. & Shahverdi, V. (2023). A new approach to character-free proof for Frobenius theorem. AUT Journal of Mathematics and Computing, 4(1), 99-103
Open this publication in new window or tab >>A new approach to character-free proof for Frobenius theorem
2023 (English)In: AUT Journal of Mathematics and Computing, ISSN 2783-2449, Vol. 4, no 1, p. 99-103Article in journal (Refereed) Published
Abstract [en]

Let G be a Frobenius group. Using character theory, it is proved that the Frobenius kernel of G is a normal subgroup of G, which is well-known as a Frobenius theorem. There is no known character-free proof for Frobenius theorem. In this note, we prove it, by assuming that Frobenius groups are non-simple. Also, we prove that whether K is a subgroup of G or not, Sylow 2-subgroups of G are either cyclic or generalized quaternion group. Also by assuming some additional arithmetical hypothesis on G we prove Frobenius theorem. We should mention that our proof is character-free.

Place, publisher, year, edition, pages
Amirkabir University of Technology, 2023
Keywords
Finite group, Frobenius group, Frobenius theorem
National Category
Discrete Mathematics
Identifiers
urn:nbn:se:kth:diva-360602 (URN)10.22060/ajmc.2022.21305.1085 (DOI)2-s2.0-85217943853 (Scopus ID)
Note

QC 20250228

Available from: 2025-02-26 Created: 2025-02-26 Last updated: 2025-02-28Bibliographically approved
Shahverdi, V., Marchetti, G. L. & Kohn, K.Learning on a Razor’s Edge: the Singularity Bias of Polynomial Neural Networks: arXiv: 2505.11846.
Open this publication in new window or tab >>Learning on a Razor’s Edge: the Singularity Bias of Polynomial Neural Networks: arXiv: 2505.11846
(English)Manuscript (preprint) (Other academic)
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-366674 (URN)
Note

QC 20250804

Available from: 2025-07-08 Created: 2025-07-08 Last updated: 2025-12-13Bibliographically approved
Shahverdi, V., Ström, E. & Andén, J.Moment Constraints and Phase Recovery for Multireference Alignment.
Open this publication in new window or tab >>Moment Constraints and Phase Recovery for Multireference Alignment
(English)Manuscript (preprint) (Other academic)
Abstract [en]

Multireference alignment (MRA) refers to the problem of recovering a signal from noisy samples subject to random circular shifts. Expectation--maximization (EM) and variational approaches use statistical modeling to achieve high accuracy at the cost of solving computationally expensive optimization problems. The method of moments, instead, achieves fast reconstructions by utilizing the power spectrum and bispectrum to determine the signal up to shift. Our approach combines the two philosophies by viewing the power spectrum as a manifold on which to constrain the signal. We then maximize the data likelihood function on this manifold with a gradient-based approach to estimate the true signal. Algorithmically, our method involves iterating between template alignment and projections onto the manifold. The method offers increased speed compared to EM and demonstrates improved accuracy over bispectrum-based methods.

Keywords
Multireference alignment, signal reconstruction, nonlinear optimization
National Category
Signal Processing
Research subject
Applied and Computational Mathematics
Identifiers
urn:nbn:se:kth:diva-374083 (URN)arXiv:2409.04868 (ISRN)10.48550/arXiv.2409.04868 (DOI)
Available from: 2025-12-13 Created: 2025-12-13 Last updated: 2025-12-13
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0009-0005-2619-9198

Search in DiVA

Show all publications