kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Similarity Learning for Spectral Clustering
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS. Qamcom Research and Technology AB, Stockholm, Sweden.ORCID iD: 0000-0003-4984-029X
Uppsala University, Uppsala, Sweden.ORCID iD: 0000-0002-2891-5435
Qamcom Research and Technology AB, Stockholm, Sweden.ORCID iD: 0000-0001-6780-7755
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS.ORCID iD: 0000-0003-4516-7317
2025 (English)In: Discovery Science - 28th International Conference, DS 2025, Proceedings, Springer Nature , 2025, p. 207-221Conference paper, Published paper (Refereed)
Abstract [en]

Spectral clustering is a widely adopted method capable of identifying complicated cluster boundaries. However, traditional spectral clustering requires the definition of a predefined similarity metric for constructing the Laplacian matrix, a requirement that limits flexibility and adaptability. Instead of predefining this metric upfront as a fixed parametric function, we introduce a novel approach that learns the optimal parameters of a similarity function through parameter optimization. This optimizes a similarity function to assign high similarity values to data pairs with shared discriminative features and low values to those without such features. Previous methods that adapt similarity measures typically treat their parameters as hyperparameters or rely on non-convex optimization strategies. However, these approaches are not well-suited for unsupervised scenarios, as they depend heavily on initial conditions and require labeled data for validation, which is unavailable in such settings. In contrast, our method employs convex optimization to learn the parameters of the similarity metrics directly, rather than treating them as hyperparameters. This enables robust and reliable unsupervised learning, making our approach particularly well-suited for spectral clustering. We validate the effectiveness and adaptability of our method on several benchmark datasets, demonstrating superior performance compared to existing techniques.

Place, publisher, year, edition, pages
Springer Nature , 2025. p. 207-221
Keywords [en]
Similarity Learning, Spectral Clustering
National Category
Computer Sciences Computer graphics and computer vision
Identifiers
URN: urn:nbn:se:kth:diva-372798DOI: 10.1007/978-3-032-05461-6_14Scopus ID: 2-s2.0-105020024626OAI: oai:DiVA.org:kth-372798DiVA, id: diva2:2014576
Conference
28th International Conference on Discovery Science, DS 2025, Ljubljana, Slovenia, September 23-25, 2025
Note

Part of ISBN 9783032054609

QC 20251118

Available from: 2025-11-18 Created: 2025-11-18 Last updated: 2025-11-18Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Komini, VangjushGirdzijauskas, Sarunas

Search in DiVA

By author/editor
Komini, VangjushKoriakina, NadezhdaRoy, DebadityaGirdzijauskas, Sarunas
By organisation
Software and Computer systems, SCS
Computer SciencesComputer graphics and computer vision

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 19 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf