Endre søk
Link to record
Permanent link

Direct link
Marchetti, Giovanni LucaORCID iD iconorcid.org/0009-0004-8248-229X
Publikasjoner (10 av 20) Visa alla publikasjoner
Marchetti, G. L., Shahverdi, V., Mereta, S., Trager, M. & Kohn, K. (2025). Algebra Unveils Deep Learning - An Invitation to Neuroalgebraic Geometry. In: : . Paper presented at International Conference on Machine Learning (ICML).
Åpne denne publikasjonen i ny fane eller vindu >>Algebra Unveils Deep Learning - An Invitation to Neuroalgebraic Geometry
Vise andre…
2025 (engelsk)Konferansepaper, Publicerat paper (Fagfellevurdert)
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-366669 (URN)
Konferanse
International Conference on Machine Learning (ICML)
Merknad

QC 20250806

Tilgjengelig fra: 2025-07-08 Laget: 2025-07-08 Sist oppdatert: 2025-12-13bibliografisk kontrollert
Henry, N. W., Marchetti, G. L. & Kohn, K. (2025). Geometry of Lightning Self-Attention: Identifiability and Dimension. In: : . Paper presented at International Conference on Learning Representations (ICLR).
Åpne denne publikasjonen i ny fane eller vindu >>Geometry of Lightning Self-Attention: Identifiability and Dimension
2025 (engelsk)Konferansepaper, Publicerat paper (Fagfellevurdert)
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-358451 (URN)
Konferanse
International Conference on Learning Representations (ICLR)
Merknad

QC 20250117

Tilgjengelig fra: 2025-01-17 Laget: 2025-01-17 Sist oppdatert: 2025-08-06bibliografisk kontrollert
García-Castellanos, A., Medbouhi, A. A., Marchetti, G. L., Bekkers, E. J. & Kragic, D. (2025). HyperSteiner: Computing Heuristic Hyperbolic Steiner Minimal Trees. In: SIAM Symposium on Algorithm Engineering and Experiments, ALENEX 2025: . Paper presented at 2025 SIAM Symposium on Algorithm Engineering and Experiments, ALENEX 2025, New Orleans, United States of America, Januari 12-13, 2025 (pp. 194-208). Society for Industrial & Applied Mathematics (SIAM)
Åpne denne publikasjonen i ny fane eller vindu >>HyperSteiner: Computing Heuristic Hyperbolic Steiner Minimal Trees
Vise andre…
2025 (engelsk)Inngår i: SIAM Symposium on Algorithm Engineering and Experiments, ALENEX 2025, Society for Industrial & Applied Mathematics (SIAM) , 2025, s. 194-208Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

We propose HyperSteiner – an efficient heuristic algorithm for computing Steiner minimal trees in the hyperbolic space. HyperSteiner extends the Euclidean Smith-Lee-Liebman algorithm, which is grounded in a divide-and-conquer approach involving the Delaunay triangulation. The central idea is rephrasing Steiner tree problems with three terminals as a system of equations in the Klein-Beltrami model. Motivated by the fact that hyperbolic geometry is well-suited for representing hierarchies, we explore applications to hierarchy discovery in data. Results show that HyperSteiner infers more realistic hierarchies than the Minimum Spanning Tree and is more scalable to large datasets than Neighbor Joining.

sted, utgiver, år, opplag, sider
Society for Industrial & Applied Mathematics (SIAM), 2025
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-359645 (URN)10.1137/1.9781611978339.16 (DOI)2-s2.0-85216422778 (Scopus ID)
Konferanse
2025 SIAM Symposium on Algorithm Engineering and Experiments, ALENEX 2025, New Orleans, United States of America, Januari 12-13, 2025
Merknad

Part of ISBN 9798331311995

QC 20250207

Tilgjengelig fra: 2025-02-06 Laget: 2025-02-06 Sist oppdatert: 2025-02-07bibliografisk kontrollert
Marchetti, G. L., Cesa, G., Kumar, P. & Behboodi, A. (2025). Neural Lattice Reduction: A Self-Supervised Geometric Deep Learning Approach. Transactions on Machine Learning Research, 2025-February, 1-22
Åpne denne publikasjonen i ny fane eller vindu >>Neural Lattice Reduction: A Self-Supervised Geometric Deep Learning Approach
2025 (engelsk)Inngår i: Transactions on Machine Learning Research, E-ISSN 2835-8856, Vol. 2025-February, s. 1-22Artikkel i tidsskrift (Fagfellevurdert) Epub ahead of print
Abstract [en]

Lattice reduction is a combinatorial optimization problem aimed at finding the most orthogonal basis in a given lattice. The Lenstra–Lenstra–Lovász (LLL) algorithm is the best algorithm in the literature for solving this problem. In light of recent research on algorithm discovery, in this work, we would like to answer this question: is it possible to parametrize the algorithm space for lattice reduction problem with neural networks and find an algorithm without supervised data? Our strategy is to use equivariant and invariant parametrizations and train in a self-supervised way. We design a deep neural model outputting factorized unimodular matrices and train it in a self-supervised manner by penalizing non-orthogonal lattice bases. We incorporate the symmetries of lattice reduction into the model by making it invariant to isometries and scaling of the ambient space and equivariant with respect to the hyperocrahedral group permuting and flipping the lattice basis elements. We show that this approach yields an algorithm with comparable complexity and performance to the LLL algorithm on a set of benchmarks. Additionally, motivated by certain applications for wire-less communication, we extend our method to a convolutional architecture which performs joint reduction of spatially-correlated lattices arranged in a grid, thereby amortizing its cost over multiple lattices.

sted, utgiver, år, opplag, sider
Transactions on Machine Learning Research, 2025
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-361196 (URN)2-s2.0-85219517527 (Scopus ID)
Merknad

QC 20250313

Tilgjengelig fra: 2025-03-12 Laget: 2025-03-12 Sist oppdatert: 2025-03-13bibliografisk kontrollert
Shahverdi, V., Marchetti, G. L. & Kohn, K. (2025). On the Geometry and Optimization of Polynomial Convolutional Networks. In: Proceedings of the 28th International Conference on Artificial Intelligence and Statistics, AISTATS 2025: . Paper presented at The 28th International Conference on Artificial Intelligence and Statistics (AISTATS), Thailand, May 3rd - May 5th, 2025 (pp. 604-612). ML Research Press, 258
Åpne denne publikasjonen i ny fane eller vindu >>On the Geometry and Optimization of Polynomial Convolutional Networks
2025 (engelsk)Inngår i: Proceedings of the 28th International Conference on Artificial Intelligence and Statistics, AISTATS 2025, ML Research Press , 2025, Vol. 258, s. 604-612Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

We study convolutional neural networks with monomial activation functions. Specifically, we prove that their parameterization map is regular and is an isomorphism almost everywhere, up to rescaling the filters. By leveraging on tools from algebraic geometry, we explore the geometric properties of the image in function space of this map - typically referred to as neuromanifold. In particular, we compute the dimension and the degree of the neuromanifold, which measure the expressivity of the model, and describe its singularities. Moreover, for a generic large dataset, we derive an explicit formula that quantifies the number of critical points arising in the optimization of a regression loss.

sted, utgiver, år, opplag, sider
ML Research Press, 2025
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-358449 (URN)2-s2.0-105014321299 (Scopus ID)
Konferanse
The 28th International Conference on Artificial Intelligence and Statistics (AISTATS), Thailand, May 3rd - May 5th, 2025
Merknad

QC 20250117

Tilgjengelig fra: 2025-01-17 Laget: 2025-01-17 Sist oppdatert: 2025-12-13bibliografisk kontrollert
Taleb, F., Medbouhi, A. A., Marchetti, G. L. & Kragic Jensfelt, D. (2025). Towards Discovering the Hierarchy of the Olfactory Perceptual Space via Hyperbolic Embeddings. In: Science Communications Worldwide (Ed.), 22nd annual Computational and Systems Neuroscience (COSYNE) conference, Montreal and Mont Tremblant, Quebec, Canada, March 27 - April 1, 2025.: . Paper presented at Computational and Systems Neuroscience (COSYNE), Montreal 27-30, Canada.
Åpne denne publikasjonen i ny fane eller vindu >>Towards Discovering the Hierarchy of the Olfactory Perceptual Space via Hyperbolic Embeddings
2025 (engelsk)Inngår i: 22nd annual Computational and Systems Neuroscience (COSYNE) conference, Montreal and Mont Tremblant, Quebec, Canada, March 27 - April 1, 2025. / [ed] Science Communications Worldwide, 2025Konferansepaper, Poster (with or without abstract) (Fagfellevurdert)
Abstract [en]

Human olfactory perception is understudied in the whole spectrum of neuroscience, from computational to system neuroscience. In this study, we explore the hierarchy underlying human olfactory perception by embedding perceptual data in the hyperbolic space. Previous research emphasizes the significance of hyperbolic geometry in gaining insights into the neural encoding of natural odorants. This is due to the exponential growth of the hyperbolic space, that makes it appropriate to encode hierarchical data. We employ a contrastive learning approach over the Poincare ball in order to embed olfactory perceptual data in a hyperbolic space. The results indicate the emergence of a hierarchical representation in the hyperbolic space, which could have implications for understanding the structure of the olfactory perceptual space in the brain. Our finding suggests that the human brain may encode olfactory perceptions in a hierarchical manner, where higher odor perceptual certainty correlates with deeper levels in the hierarchical representation.

Emneord
hyperbolic geometry, olfaction, representation
HSV kategori
Forskningsprogram
Datalogi
Identifikatorer
urn:nbn:se:kth:diva-363971 (URN)10.57736/0d78-6155 (DOI)
Konferanse
Computational and Systems Neuroscience (COSYNE), Montreal 27-30, Canada
Merknad

QC 20250602

Tilgjengelig fra: 2025-06-01 Laget: 2025-06-01 Sist oppdatert: 2025-06-02bibliografisk kontrollert
Marchetti, G. L., Hillar, C. J., Kragic, D. & Sanborn, S. (2024). Harmonics of Learning: Universal Fourier Features Emerge in Invariant Networks. In: Proceedings of 37th Conference on Learning Theory, COLT 2024: . Paper presented at 37th Annual Conference on Learning Theory, COLT 2024, Edmonton, Canada, Jun 30 2024 - Jul 3 2024 (pp. 3775-3797). ML Research Press
Åpne denne publikasjonen i ny fane eller vindu >>Harmonics of Learning: Universal Fourier Features Emerge in Invariant Networks
2024 (engelsk)Inngår i: Proceedings of 37th Conference on Learning Theory, COLT 2024, ML Research Press , 2024, s. 3775-3797Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

In this work, we formally prove that, under certain conditions, if a neural network is invariant to a finite group then its weights recover the Fourier transform on that group. This provides a mathematical explanation for the emergence of Fourier features - a ubiquitous phenomenon in both biological and artificial learning systems. The results hold even for non-commutative groups, in which case the Fourier transform encodes all the irreducible unitary group representations. Our findings have consequences for the problem of symmetry discovery. Specifically, we demonstrate that the algebraic structure of an unknown group can be recovered from the weights of a network that is at least approximately invariant within certain bounds. Overall, this work contributes to a foundation for an algebraic learning theory of invariant neural network representations.

sted, utgiver, år, opplag, sider
ML Research Press, 2024
Emneord
group representations, harmonic analysis, Invariant neural networks
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-353960 (URN)2-s2.0-85203678110 (Scopus ID)
Konferanse
37th Annual Conference on Learning Theory, COLT 2024, Edmonton, Canada, Jun 30 2024 - Jul 3 2024
Merknad

QC 20240926

Tilgjengelig fra: 2024-09-25 Laget: 2024-09-25 Sist oppdatert: 2024-10-03bibliografisk kontrollert
Medbouhi, A. A., Marchetti, G. L., Polianskii, V., Kravberg, A., Poklukar, P., Varava, A. & Kragic, D. (2024). Hyperbolic Delaunay Geometric Alignment. In: Bifet, A Krilavicius, T Davis, J Kull, M Ntoutsi, E Zliobaite, I (Ed.), Machine learning and knowledge discovery in databases: Research track, pt iii, ECML PKDD 2024. Paper presented at Joint European Conference on Machine Learning and Knowledge Discovery in Databases (ECML PKDD), SEP 09-13, 2024, Vilnius, Lithuania (pp. 111-126). Springer Nature
Åpne denne publikasjonen i ny fane eller vindu >>Hyperbolic Delaunay Geometric Alignment
Vise andre…
2024 (engelsk)Inngår i: Machine learning and knowledge discovery in databases: Research track, pt iii, ECML PKDD 2024 / [ed] Bifet, A Krilavicius, T Davis, J Kull, M Ntoutsi, E Zliobaite, I, Springer Nature , 2024, s. 111-126Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

Hyperbolic machine learning is an emerging field aimed at representing data with a hierarchical structure. However, there is a lack of tools for evaluation and analysis of the resulting hyperbolic data representations. To this end, we propose Hyperbolic Delaunay Geometric Alignment (HyperDGA) - a similarity score for comparing datasets in a hyperbolic space. The core idea is counting the edges of the hyperbolic Delaunay graph connecting datapoints across the given sets. We provide an empirical investigation on synthetic and real-life biological data and demonstrate that HyperDGA outperforms the hyperbolic version of classical distances between sets. Furthermore, we showcase the potential of HyperDGA for evaluating latent representations inferred by a Hyperbolic Variational Auto-Encoder.

sted, utgiver, år, opplag, sider
Springer Nature, 2024
Serie
Lecture Notes in Artificial Intelligence, ISSN 2945-9133 ; 14943
Emneord
Hyperbolic Geometry, Hierarchical Data, Evaluation
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-355149 (URN)10.1007/978-3-031-70352-2_7 (DOI)001308375900007 ()
Konferanse
Joint European Conference on Machine Learning and Knowledge Discovery in Databases (ECML PKDD), SEP 09-13, 2024, Vilnius, Lithuania
Merknad

Part of ISBN: 978-3-031-70351-5, 978-3-031-70352-2

QC 20241025

Tilgjengelig fra: 2024-10-25 Laget: 2024-10-25 Sist oppdatert: 2024-10-25bibliografisk kontrollert
Marchetti, G. L. (2024). On Symmetries and Metrics in Geometric Inference. (Doctoral dissertation). KTH Royal Institute of Technology
Åpne denne publikasjonen i ny fane eller vindu >>On Symmetries and Metrics in Geometric Inference
2024 (engelsk)Doktoravhandling, med artikler (Annet vitenskapelig)
Abstract [en]

Spaces of data naturally carry intrinsic geometry. Statistics and machine learning can leverage on this rich structure in order to achieve efficiency and semantic generalization. Extracting geometry from data is therefore a fundamental challenge which by itself defines a statistical, computational and unsupervised learning problem. To this end, symmetries and metrics are two fundamental objects which are ubiquitous in continuous and discrete geometry. Both are suitable for data-driven approaches since symmetries arise as interactions and are thus collectable in practice while metrics can be induced locally from the ambient space. In this thesis, we address the question of extracting geometry from data by leveraging on symmetries and metrics. Additionally, we explore methods for statistical inference exploiting the extracted geometric structure. On the metric side, we focus on Voronoi tessellations and Delaunay triangulations, which are classical tools in computational geometry. Based on them, we propose novel non-parametric methods for machine learning and statistics, focusing on theoretical and computational aspects. These methods include an active version of the nearest neighbor regressor as well as two high-dimensional density estimators. All of them possess convergence guarantees due to the adaptiveness of Voronoi cells. On the symmetry side, we focus on representation learning in the context of data acted upon by a group. Specifically, we propose a method for learning equivariant representations which are guaranteed to be isomorphic to the data space, even in the presence of symmetries stabilizing data. We additionally explore applications of such representations in a robotics context, where symmetries correspond to actions performed by an agent. Lastly, we provide a theoretical analysis of invariant neural networks and show how the group-theoretical Fourier transform emerges in their weights. This addresses the problem of symmetry discovery in a self-supervised manner.  

Abstract [sv]

Datamängder innehar en naturlig inneboende geometri. Statistik och maskininlärning kan dra nytta av denna rika struktur för att uppnå effektivitet och semantisk generalisering. Att extrahera geometri ifrån data är därför en grundläggande utmaning som i sig definierar ett statistiskt, beräkningsmässigt och oövervakat inlärningsproblem. För detta ändamål är symmetrier och metriker två grundläggande objekt som är allestädes närvarande i kontinuerlig och diskret geometri. Båda är lämpliga för datadrivna tillvägagångssätt eftersom symmetrier uppstår som interaktioner och är därmed i praktiken samlingsbara medan metriker kan induceras lokalt ifrån det omgivande rummet. I denna avhandling adresserar vi frågan om att extrahera geometri ifrån data genom att utnyttja symmetrier och metriker. Dessutom utforskar vi metoder för statistisk inferens som utnyttjar den extraherade geometriska strukturen. På den metriska sidan fokuserar vi på Voronoi-tessellationer och Delaunay-trianguleringar, som är klassiska verktyg inom beräkningsgeometri. Baserat på dem föreslår vi nya icke-parametriska metoder för maskininlärning och statistik, med fokus på teoretiska och beräkningsmässiga aspekter. Dessa metoder inkluderar en aktiv version av närmaste grann-regressorn samt två högdimensionella täthetsskattare. Alla dessa besitter konvergensgarantier på grund av Voronoi-cellernas anpassningsbarhet. På symmetrisidan fokuserar vi på representationsinlärning i sammanhanget av data som påverkas av en grupp. Specifikt föreslår vi en metod för att lära sig ekvivarianta representationer som garanteras vara isomorfa till datarummet, även i närvaro av symmetrier som stabiliserar data. Vi utforskar även tillämpningar av sådana representationer i ett robotiksammanhang, där symmetrier motsvarar handlingar utförda av en agent. Slutligen tillhandahåller vi en teoretisk analys av invarianta neuronnät och visar hur den gruppteoretiska Fouriertransformen framträder i deras vikter. Detta adresserar problemet med att upptäcka symmetrier på ett självövervakat sätt.

sted, utgiver, år, opplag, sider
KTH Royal Institute of Technology, 2024. s. 61
Serie
TRITA-EECS-AVL ; 2024:26
Emneord
Machine Learning, Computational Geometry, Voronoi, Delaunay, Symmetry, Equivariance
HSV kategori
Forskningsprogram
Datalogi
Identifikatorer
urn:nbn:se:kth:diva-344129 (URN)978-91-8040-864-6 (ISBN)
Disputas
2024-04-09, https://kth-se.zoom.us/j/61437033234?pwd=dnBpMnYyaDVWWC95RHNTakNXWkNRQT09, F3 (Flodis) Lindstedtsvägen 26, Stockholm, 09:00 (engelsk)
Opponent
Veileder
Merknad

QC 20240304

Tilgjengelig fra: 2024-03-04 Laget: 2024-03-02 Sist oppdatert: 2025-05-23bibliografisk kontrollert
Garcia-Castellanos, A., Marchetti, G. L., Kragic Jensfelt, D. & Scolamiero, M. (2024). Relative Representations: Topological and Geometric Perspectives. In: Marco Fumero; Clementine Domine; Zorah Lähner; Donato Crisostomi; Luca Moschella; Kimberly Stachenfeld (Ed.), Proceedings of UniReps: 2nd Edition of the Workshop on Unifying Representations in Neural Models: . Paper presented at 2nd Edition of the Workshop on Unifying Representations in Neural Models, UniReps 2024, Vancouver, Canada, December 14, 2024. ML Research Press
Åpne denne publikasjonen i ny fane eller vindu >>Relative Representations: Topological and Geometric Perspectives
2024 (engelsk)Inngår i: Proceedings of UniReps: 2nd Edition of the Workshop on Unifying Representations in Neural Models / [ed] Marco Fumero; Clementine Domine; Zorah Lähner; Donato Crisostomi; Luca Moschella; Kimberly Stachenfeld, ML Research Press , 2024Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

Relative representations are an established approach to zero-shot model stitching, consisting of a non-trainable transformation of the latent space of a deep neural network. Based on insights of topological and geometric nature, we propose two improvements to relative representations. First, we introduce a normalization procedure in the relative transformation, resulting in invariance to non-isotropic rescalings and permutations. The latter coincides with the symmetries in parameter space induced by common activation functions. Second, we propose to deploy topological densification when fine-tuning relative representations, a topological regularization loss encouraging clustering within classes. We provide an empirical investigation on a natural language task, where both the proposed variations yield improved performance on zero-shot model stitching.

sted, utgiver, år, opplag, sider
ML Research Press, 2024
Serie
Proceedings of Machine Learning Research ; 285
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-370458 (URN)2-s2.0-105014754343 (Scopus ID)
Konferanse
2nd Edition of the Workshop on Unifying Representations in Neural Models, UniReps 2024, Vancouver, Canada, December 14, 2024
Merknad

QC 20250929

Tilgjengelig fra: 2025-09-29 Laget: 2025-09-29 Sist oppdatert: 2025-09-29bibliografisk kontrollert
Organisasjoner
Identifikatorer
ORCID-id: ORCID iD iconorcid.org/0009-0004-8248-229X