kth.sePublications KTH
Operational message
There are currently operational disruptions. Troubleshooting is in progress.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Indirectly Parameterized Concrete Autoencoders
KTH, School of Engineering Sciences in Chemistry, Biotechnology and Health (CBH), Gene Technology. Science for Life Laboratory, Solna, Sweden.
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL. KTH, Centres, SeRC - Swedish e-Science Research Centre.
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0003-4535-2520
Show others and affiliations
2024 (English)In: International Conference on Machine Learning, ICML 2024, ML Research Press , 2024, p. 38237-38252Conference paper, Published paper (Refereed)
Abstract [en]

Feature selection is a crucial task in settings where data is high-dimensional or acquiring the full set of features is costly. Recent developments in neural network-based embedded feature selection show promising results across a wide range of applications. Concrete Autoencoders (CAEs), considered state-of-the-art in embedded feature selection, may struggle to achieve stable joint optimization, hurting their training time and generalization. In this work, we identify that this instability is correlated with the CAE learning duplicate selections. To remedy this, we propose a simple and effective improvement: Indirectly Parameterized CAEs (IP-CAEs). IP-CAEs learn an embedding and a mapping from it to the Gumbel-Softmax distributions' parameters. Despite being simple to implement, IP-CAE exhibits significant and consistent improvements over CAE in both generalization and training time across several datasets for reconstruction and classification. Unlike CAE, IP-CAE effectively leverages non-linear relationships and does not require retraining the jointly optimized decoder. Furthermore, our approach is, in principle, generalizable to Gumbel-Softmax distributions beyond feature selection.

Place, publisher, year, edition, pages
ML Research Press , 2024. p. 38237-38252
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-353956Scopus ID: 2-s2.0-85203808876OAI: oai:DiVA.org:kth-353956DiVA, id: diva2:1901032
Conference
41st International Conference on Machine Learning, ICML 2024, Vienna, Austria, Jul 21 2024 - Jul 27 2024
Note

QC 20240926

Available from: 2024-09-25 Created: 2024-09-25 Last updated: 2024-09-26Bibliographically approved

Open Access in DiVA

No full text in DiVA

Scopus

Authority records

Nilsson, AlfredWijk, KlasGutha, Sai bharath chandraEnglesson, ErikHotti, AlexandraSaccardi, CarloKviman, OskarLagergren, JensVinuesa, RicardoAzizpour, Hossein

Search in DiVA

By author/editor
Nilsson, AlfredWijk, KlasGutha, Sai bharath chandraEnglesson, ErikHotti, AlexandraSaccardi, CarloKviman, OskarLagergren, JensVinuesa, RicardoAzizpour, Hossein
By organisation
Gene TechnologyRobotics, Perception and Learning, RPLSeRC - Swedish e-Science Research CentreScience for Life Laboratory, SciLifeLabComputational Science and Technology (CST)Fluid Mechanics
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 199 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf