kth.sePublications KTH
Change search
Link to record
Permanent link

Direct link
Publications (4 of 4) Show all publications
Gyllberg, L., Zhao, S. & Kronqvist, J. (2026). Convex Quadratic Programming-Based Predictors: An Algorithmic Framework and a Study of Possibilities and Computational Challenges. In: Learning and Intelligent Optimization - 19th International Conference, LION 19 2025, Proceedings: . Paper presented at 19th International Conference on Learning and Intelligent Optimization, LION 2025, Prague, Czechia, June 15-19, 2025 (pp. 191-202). Springer Science and Business Media Deutschland GmbH
Open this publication in new window or tab >>Convex Quadratic Programming-Based Predictors: An Algorithmic Framework and a Study of Possibilities and Computational Challenges
2026 (English)In: Learning and Intelligent Optimization - 19th International Conference, LION 19 2025, Proceedings, Springer Science and Business Media Deutschland GmbH , 2026, p. 191-202Conference paper, Published paper (Refereed)
Abstract [en]

We present a class of predictive models for forecasting time-series data, referred to as convex quadratic programming-based (CQPB) predictors. The predictions are computed from the minimizer of a convex quadratic problem, where previous observations are integrated as parameters. The remaining parameters, including constraints and objective coefficients, are trainable parameters. This work investigates the predictive capabilities of CQPB predictors and the computational challenges in their training. We analyze their properties and prove that this class of predictors includes classical autoregressive (AR) models, thus forming a generalization of AR models. The training problem is formulated as a bilevel optimization problem. To solve these training problems efficiently, we propose a two-stage heuristic algorithm based on the block coordinate descent approach. The results highlight the potential of CQPB predictors. Although training is challenging, our approach efficiently computes good solutions for moderate-size datasets.

Place, publisher, year, edition, pages
Springer Science and Business Media Deutschland GmbH, 2026
Keywords
Bilevel optimization, Inverse optimization, Optimization-based predictive models, Time-series prediction
National Category
Control Engineering
Identifiers
urn:nbn:se:kth:diva-376717 (URN)10.1007/978-3-032-09192-5_13 (DOI)2-s2.0-105028355546 (Scopus ID)
Conference
19th International Conference on Learning and Intelligent Optimization, LION 2025, Prague, Czechia, June 15-19, 2025
Note

Part of ISBN 9783032091918

QC 20260218

Available from: 2026-02-18 Created: 2026-02-18 Last updated: 2026-02-18Bibliographically approved
Bautista, L. A., Hrga, T., Povh, J. & Zhao, S. (2025). Ground truth clustering is not the optimum clustering. Scientific Reports, 15(1), Article ID 9223.
Open this publication in new window or tab >>Ground truth clustering is not the optimum clustering
2025 (English)In: Scientific Reports, E-ISSN 2045-2322, Vol. 15, no 1, article id 9223Article in journal (Refereed) Published
Abstract [en]

Data clustering is a fundamental yet challenging task in data science. The minimum sum-of-squares clustering (MSSC) problem aims to partition data points into k clusters to minimize the sum of squared distances between the points and their cluster centers (centroids). Despite being NP-hard, solvers exist that can compute optimal solutions for small to medium-sized datasets. One such solver is SOS-SDP, a branch-and-bound algorithm based on semidefinite programming. We used it to obtain optimal MSSC solutions (optimum clusterings) for various k across multiple datasets with known ground truth clusterings. We evaluated the alignment between the optimum and ground truth clusterings using six extrinsic measures and assessed their quality using three intrinsic measures. The results reveal that the optimum clusterings often differ significantly from the ground truth clusterings. Additionally, the optimum clusterings frequently outperform the ground truth clusterings, according to the intrinsic measures that we used. However, when ground truth clusters are well-separated convex shapes, such as ellipsoids, the optimum and ground truth clusterings closely align.

Place, publisher, year, edition, pages
Springer Nature, 2025
Keywords
Extrinsic measures, Ground truth clustering, Intrinsic measures, Minimum sum-of-squares clustering
National Category
Computer graphics and computer vision
Identifiers
urn:nbn:se:kth:diva-362013 (URN)10.1038/s41598-025-90865-9 (DOI)001446949700011 ()40097499 (PubMedID)2-s2.0-105000375355 (Scopus ID)
Note

QC 20250425

Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-04-25Bibliographically approved
Kronqvist, J., Li, B., Rolfes, J. & Zhao, S. (2024). Alternating Mixed-Integer Programming and Neural Network Training for Approximating Stochastic Two-Stage Problems. In: Machine Learning, Optimization, and Data Science - 9th International Conference, LOD 2023, Revised Selected Papers: . Paper presented at 9th International Conference on Machine Learning, Optimization, and Data Science, LOD 2023, Grasmere, United Kingdom of Great Britain and Northern Ireland, Sep 22 2023 - Sep 26 2023 (pp. 124-139). Springer Nature, 14506
Open this publication in new window or tab >>Alternating Mixed-Integer Programming and Neural Network Training for Approximating Stochastic Two-Stage Problems
2024 (English)In: Machine Learning, Optimization, and Data Science - 9th International Conference, LOD 2023, Revised Selected Papers, Springer Nature , 2024, Vol. 14506, p. 124-139Conference paper, Published paper (Refereed)
Abstract [en]

The presented work addresses two-stage stochastic programs (2SPs), a broadly applicable model to capture optimization problems subject to uncertain parameters with adjustable decision variables. In case the adjustable or second-stage variables contain discrete decisions, the corresponding 2SPs are known to be NP-complete. The standard approach of forming a single-stage deterministic equivalent problem can be computationally challenging even for small instances, as the number of variables and constraints scales with the number of scenarios. To avoid forming a potentially huge MILP problem, we build upon an approach of approximating the expected value of the second-stage problem by a neural network (NN) and encoding the resulting NN into the first-stage problem. The proposed algorithm alternates between optimizing the first-stage variables and retraining the NN. We demonstrate the value of our approach with the example of computing operating points in power systems by showing that the alternating approach provides improved first-stage decisions and a tighter approximation between the expected objective and its neural network approximation.

Place, publisher, year, edition, pages
Springer Nature, 2024
Series
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), ISSN 0302-9743 ; 14506
Keywords
Neural Network, Power Systems, Stochastic Optimization
National Category
Computational Mathematics
Identifiers
urn:nbn:se:kth:diva-344367 (URN)10.1007/978-3-031-53966-4_10 (DOI)001217090300010 ()2-s2.0-85186266492 (Scopus ID)
Conference
9th International Conference on Machine Learning, Optimization, and Data Science, LOD 2023, Grasmere, United Kingdom of Great Britain and Northern Ireland, Sep 22 2023 - Sep 26 2023
Note

QC 20240314

 Part of ISBN 9783031539657

Available from: 2024-03-13 Created: 2024-03-13 Last updated: 2024-06-14Bibliographically approved
Zhao, S., Tsay, C. & Kronqvist, J. (2023). Model-Based Feature Selection for Neural Networks: A Mixed-Integer Programming Approach. In: Learning and Intelligent Optimization: 17th International Conference, LION 17, Revised Selected Papers. Paper presented at 17th International Conference on Learning and Intelligent Optimization, LION-17 2023, Nice, France, Jun 4 2023 - Jun 8 2023 (pp. 223-238). Springer Nature
Open this publication in new window or tab >>Model-Based Feature Selection for Neural Networks: A Mixed-Integer Programming Approach
2023 (English)In: Learning and Intelligent Optimization: 17th International Conference, LION 17, Revised Selected Papers, Springer Nature , 2023, p. 223-238Conference paper, Published paper (Refereed)
Abstract [en]

In this work, we develop a novel input feature selection framework for ReLU-based deep neural networks (DNNs), which builds upon a mixed-integer optimization approach. While the method is generally applicable to various classification tasks, we focus on finding input features for image classification for clarity of presentation. The idea is to use a trained DNN, or an ensemble of trained DNNs, to identify the salient input features. The input feature selection is formulated as a sequence of mixed-integer linear programming (MILP) problems that find sets of sparse inputs that maximize the classification confidence of each category. These “inverse” problems are regularized by the number of inputs selected for each category and by distribution constraints. Numerical results on the well-known MNIST and FashionMNIST datasets show that the proposed input feature selection allows us to drastically reduce the size of the input to ∼ 15% while maintaining a good classification accuracy. This allows us to design DNNs with significantly fewer connections, reducing computational effort and producing DNNs that are more robust towards adversarial attacks.

Place, publisher, year, edition, pages
Springer Nature, 2023
Keywords
Deep neural networks, Feature selection, Mixed-integer programming, Model reduction, Sparse DNNs
National Category
Computer graphics and computer vision
Identifiers
urn:nbn:se:kth:diva-339676 (URN)10.1007/978-3-031-44505-7_16 (DOI)001532132100016 ()2-s2.0-85175971445 (Scopus ID)
Conference
17th International Conference on Learning and Intelligent Optimization, LION-17 2023, Nice, France, Jun 4 2023 - Jun 8 2023
Note

Part of ISBN 9783031445040

QC 20231116

Available from: 2023-11-16 Created: 2023-11-16 Last updated: 2025-12-08Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0001-6352-0968

Search in DiVA

Show all publications