kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
A Coding-Theoretic Analysis of Hyperspherical Prototypical Learning Geometry
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Information Science and Engineering.ORCID iD: 0000-0002-7423-7970
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Information Science and Engineering.ORCID iD: 0000-0002-0862-1333
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Information Science and Engineering.ORCID iD: 0000-0001-9307-484X
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Information Science and Engineering.ORCID iD: 0000-0002-7926-5081
2024 (English)In: Proceedings of the Geometry-Grounded Representation Learning and Generative Modeling Workshop, GRaM 2024 at ICML 2024, ML Research Press , 2024, p. 78-91Conference paper, Published paper (Refereed)
Abstract [en]

Hyperspherical Prototypical Learning (HPL) is a supervised approach to representation learning that designs class prototypes on the unit hypersphere. The prototypes bias the representations to class separation in a scale invariant and known geometry. Previous approaches to HPL have either of the following shortcomings: (i) they follow an unprincipled optimisation procedure; or (ii) they are theoretically sound, but are constrained to only one possible latent dimension. In this paper, we address both shortcomings. To address (i), we present a principled optimisation procedure whose solution we show is optimal. To address (ii), we construct well-separated prototypes in a wide range of dimensions using linear block codes. Additionally, we give a full characterisation of the optimal prototype placement in terms of achievable and converse bounds, showing that our proposed methods are near-optimal.

Place, publisher, year, edition, pages
ML Research Press , 2024. p. 78-91
National Category
Probability Theory and Statistics Computational Mathematics
Identifiers
URN: urn:nbn:se:kth:diva-359860Scopus ID: 2-s2.0-85216611518OAI: oai:DiVA.org:kth-359860DiVA, id: diva2:1937169
Conference
1st Geometry-Grounded Representation Learning and Generative Modeling Workshop, GRaM 2024 at the 41st International Conference on Machine Learning, ICML 2024, Vienna, Austria, July 29, 2024
Note

QC 20250213

Available from: 2025-02-12 Created: 2025-02-12 Last updated: 2025-02-13Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

ScopusFulltext

Authority records

Lindström, MartinRodríguez Gálvez, BorjaThobaben, RagnarSkoglund, Mikael

Search in DiVA

By author/editor
Lindström, MartinRodríguez Gálvez, BorjaThobaben, RagnarSkoglund, Mikael
By organisation
Information Science and Engineering
Probability Theory and StatisticsComputational Mathematics

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 96 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf