kth.sePublikationer KTH
Ändra sökning
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Model-Based Feature Selection for Neural Networks: A Mixed-Integer Programming Approach
KTH, Skolan för teknikvetenskap (SCI), Matematik (Inst.), Optimeringslära och systemteori.
Department of Computing, Imperial College London, London, UK.
KTH, Skolan för teknikvetenskap (SCI), Matematik (Inst.), Optimeringslära och systemteori.ORCID-id: 0000-0003-0299-5745
2023 (Engelska)Ingår i: Learning and Intelligent Optimization: 17th International Conference, LION 17, Revised Selected Papers, Springer Nature , 2023, s. 223-238Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

In this work, we develop a novel input feature selection framework for ReLU-based deep neural networks (DNNs), which builds upon a mixed-integer optimization approach. While the method is generally applicable to various classification tasks, we focus on finding input features for image classification for clarity of presentation. The idea is to use a trained DNN, or an ensemble of trained DNNs, to identify the salient input features. The input feature selection is formulated as a sequence of mixed-integer linear programming (MILP) problems that find sets of sparse inputs that maximize the classification confidence of each category. These “inverse” problems are regularized by the number of inputs selected for each category and by distribution constraints. Numerical results on the well-known MNIST and FashionMNIST datasets show that the proposed input feature selection allows us to drastically reduce the size of the input to ∼ 15% while maintaining a good classification accuracy. This allows us to design DNNs with significantly fewer connections, reducing computational effort and producing DNNs that are more robust towards adversarial attacks.

Ort, förlag, år, upplaga, sidor
Springer Nature , 2023. s. 223-238
Nyckelord [en]
Deep neural networks, Feature selection, Mixed-integer programming, Model reduction, Sparse DNNs
Nationell ämneskategori
Datorgrafik och datorseende
Identifikatorer
URN: urn:nbn:se:kth:diva-339676DOI: 10.1007/978-3-031-44505-7_16ISI: 001532132100016Scopus ID: 2-s2.0-85175971445OAI: oai:DiVA.org:kth-339676DiVA, id: diva2:1812485
Konferens
17th International Conference on Learning and Intelligent Optimization, LION-17 2023, Nice, France, Jun 4 2023 - Jun 8 2023
Anmärkning

Part of ISBN 9783031445040

QC 20231116

Tillgänglig från: 2023-11-16 Skapad: 2023-11-16 Senast uppdaterad: 2025-12-08Bibliografiskt granskad

Open Access i DiVA

Fulltext saknas i DiVA

Övriga länkar

Förlagets fulltextScopus

Person

Zhao, ShudianKronqvist, Jan

Sök vidare i DiVA

Av författaren/redaktören
Zhao, ShudianKronqvist, Jan
Av organisationen
Optimeringslära och systemteori
Datorgrafik och datorseende

Sök vidare utanför DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetricpoäng

doi
urn-nbn
Totalt: 74 träffar
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf