kth.sePublications KTH
Change search
Link to record
Permanent link

Direct link
Publications (10 of 11) Show all publications
Hakkinen, I., Melekhov, I., Englesson, E., Azizpour, H. & Kannala, J. (2025). Medical Image Segmentation with SAM-Generated Annotations. In: DelBue, A Canton, C Pont-Tuset, J Tommasi, T (Ed.), Computer Vision-Eccv 2024 Workshops, Pt Xxii: . Paper presented at 18th European Conference on Computer Vision (ECCV), Sep 29- 04, 2024, Milan, Italy (pp. 51-62). Springer Nature, 15644
Open this publication in new window or tab >>Medical Image Segmentation with SAM-Generated Annotations
Show others...
2025 (English)In: Computer Vision-Eccv 2024 Workshops, Pt Xxii / [ed] DelBue, A Canton, C Pont-Tuset, J Tommasi, T, Springer Nature , 2025, Vol. 15644, p. 51-62Conference paper, Published paper (Refereed)
Abstract [en]

The field of medical image segmentation is hindered by the scarcity of large, publicly available annotated datasets. Not all datasets are made public for privacy reasons, and creating annotations for a large dataset is time-consuming and expensive, as it requires specialized expertise to accurately identify regions of interest (ROIs) within the images. To address these challenges, we evaluate the performance of the Segment Anything Model (SAM) as an annotation tool for medical data by using it to produce so-called "pseudo labels" on the Medical Segmentation Decathlon (MSD) computed tomography (CT) tasks. The pseudo labels are then used in place of ground truth labels to train a UNet model in a weakly-supervised manner. We experiment with different prompt types on SAM and find that the bounding box prompt is a simple yet effective method for generating pseudo labels. This method allows us to develop a weakly-supervised model that performs comparably to a fully supervised model.

Place, publisher, year, edition, pages
Springer Nature, 2025
Series
Lecture Notes in Computer Science, ISSN 0302-9743
Keywords
Foundation Model, Segment Anything Model, Medical Image Segmentation, Data Annotation
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-374359 (URN)10.1007/978-3-031-92089-9_4 (DOI)001544995100004 ()2-s2.0-105006891572 (Scopus ID)
Conference
18th European Conference on Computer Vision (ECCV), Sep 29- 04, 2024, Milan, Italy
Note

Part of ISBN 978-3-031-92088-2; 978-3-031-92089-9

QC 20251218

Available from: 2025-12-18 Created: 2025-12-18 Last updated: 2025-12-18Bibliographically approved
Mehrpanah, A., Englesson, E. & Azizpour, H. (2025). On Spectral Properties of Gradient-Based Explanation Methods. In: Computer Vision – ECCV 2024 - 18th European Conference, Proceedings: . Paper presented at 18th European Conference on Computer Vision, ECCV 2024, Milan, Italy, Sep 29 2024 - Oct 4 2024 (pp. 282-299). Springer Nature
Open this publication in new window or tab >>On Spectral Properties of Gradient-Based Explanation Methods
2025 (English)In: Computer Vision – ECCV 2024 - 18th European Conference, Proceedings, Springer Nature , 2025, p. 282-299Conference paper, Published paper (Refereed)
Abstract [en]

Understanding the behavior of deep networks is crucial to increase our confidence in their results. Despite an extensive body of work for explaining their predictions, researchers have faced reliability issues, which can be attributed to insufficient formalism. In our research, we adopt novel probabilistic and spectral perspectives to formally analyze explanation methods. Our study reveals a pervasive spectral bias stemming from the use of gradient, and sheds light on some common design choices that have been discovered experimentally, in particular, the use of squared gradient and input perturbation. We further characterize how the choice of perturbation hyperparameters in explanation methods, such as SmoothGrad, can lead to inconsistent explanations and introduce two remedies based on our proposed formalism: (i) a mechanism to determine a standard perturbation scale, and (ii) an aggregation method which we call SpectralLens. Finally, we substantiate our theoretical results through quantitative evaluations.

Place, publisher, year, edition, pages
Springer Nature, 2025
Keywords
Deep Neural Networks, Explainability, Gradient-based Explanation Methods, Probabilistic Machine Learning, Probabilistic Pixel Attribution Techniques, Spectral Analysis
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-357698 (URN)10.1007/978-3-031-73021-4_17 (DOI)001416940200017 ()2-s2.0-85210488897 (Scopus ID)
Conference
18th European Conference on Computer Vision, ECCV 2024, Milan, Italy, Sep 29 2024 - Oct 4 2024
Note

Part of ISBN 978-303173020-7

QC 20241213

Available from: 2024-12-12 Created: 2024-12-12 Last updated: 2025-03-17Bibliographically approved
Nilsson, A., Wijk, K., Gutha, S. b., Englesson, E., Hotti, A., Saccardi, C., . . . Azizpour, H. (2024). Indirectly Parameterized Concrete Autoencoders. In: International Conference on Machine Learning, ICML 2024: . Paper presented at 41st International Conference on Machine Learning, ICML 2024, Vienna, Austria, Jul 21 2024 - Jul 27 2024 (pp. 38237-38252). ML Research Press
Open this publication in new window or tab >>Indirectly Parameterized Concrete Autoencoders
Show others...
2024 (English)In: International Conference on Machine Learning, ICML 2024, ML Research Press , 2024, p. 38237-38252Conference paper, Published paper (Refereed)
Abstract [en]

Feature selection is a crucial task in settings where data is high-dimensional or acquiring the full set of features is costly. Recent developments in neural network-based embedded feature selection show promising results across a wide range of applications. Concrete Autoencoders (CAEs), considered state-of-the-art in embedded feature selection, may struggle to achieve stable joint optimization, hurting their training time and generalization. In this work, we identify that this instability is correlated with the CAE learning duplicate selections. To remedy this, we propose a simple and effective improvement: Indirectly Parameterized CAEs (IP-CAEs). IP-CAEs learn an embedding and a mapping from it to the Gumbel-Softmax distributions' parameters. Despite being simple to implement, IP-CAE exhibits significant and consistent improvements over CAE in both generalization and training time across several datasets for reconstruction and classification. Unlike CAE, IP-CAE effectively leverages non-linear relationships and does not require retraining the jointly optimized decoder. Furthermore, our approach is, in principle, generalizable to Gumbel-Softmax distributions beyond feature selection.

Place, publisher, year, edition, pages
ML Research Press, 2024
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-353956 (URN)2-s2.0-85203808876 (Scopus ID)
Conference
41st International Conference on Machine Learning, ICML 2024, Vienna, Austria, Jul 21 2024 - Jul 27 2024
Note

QC 20240926

Available from: 2024-09-25 Created: 2024-09-25 Last updated: 2024-09-26Bibliographically approved
Englesson, E. (2024). On Label Noise in Image Classification: An Aleatoric Uncertainty Perspective. (Doctoral dissertation). Stockholm: KTH Royal Institute of Technology
Open this publication in new window or tab >>On Label Noise in Image Classification: An Aleatoric Uncertainty Perspective
2024 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Deep neural networks and large-scale datasets have revolutionized the field of machine learning. However, these large networks are susceptible to overfitting to label noise, resulting in generalization degradation. In response, the thesis closely examines the problem both from an empirical and theoretical perspective. We empirically analyse the input smoothness of networks as they overfit to label noise, and we theoretically explore the connection to aleatoric uncertainty. These analyses improve our understanding of the problem and have led to our novel methods aimed at enhancing robustness against label noise in classification.

Abstract [sv]

Djupa neurala nätverk och storskaliga dataset har revolutionerat maskininlärningsområdet. Dock är dessa stora nätverk känsliga för överanpassning till felmarkerade etiketter, vilket leder till försämrad generalisering. Som svar på detta undersöker avhandlingen noggrant problemet både från en empirisk och teoretisk synvinkel. Vi analyserar empiriskt nätverkens känslighet försmå ändringar i indatan när de över anpassar till felmarkerade etiketter, och vi utforskar teoretiskt kopplingen till aleatorisk osäkerhet. Dessa analyser förbättrar vår förståelse av problemet och har lett till våra nya metoder med syfte att vara robusta mot felmarkerade etiketter i klassificering.

Place, publisher, year, edition, pages
Stockholm: KTH Royal Institute of Technology, 2024. p. xi, 68
Series
TRITA-EECS-AVL ; 2024:45
Keywords
Label noise, aleatoric uncertainty, noisy labels, robustness, etikettbrus, osäkerhet, felmarkerade etiketter, robusthet
National Category
Computer Sciences
Research subject
Computer Science
Identifiers
urn:nbn:se:kth:diva-346453 (URN)978-91-8040-925-4 (ISBN)
Public defence
2024-06-03, https://kth-se.zoom.us/w/61097277235, F3 (Flodis), Lindstedsvägen 26 & 28, Stockholm, 09:00 (English)
Opponent
Supervisors
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

QC 20240516

Available from: 2024-05-16 Created: 2024-05-16 Last updated: 2025-12-03Bibliographically approved
Englesson, E. & Azizpour, H. (2024). Robust Classification via Regression for Learning with Noisy Labels. In: Proceedings ICLR 2024 - The Twelfth International Conference on Learning Representations: . Paper presented at ICLR 2024 - The Twelfth International Conference on Learning Representations, Messe Wien Exhibition and Congress Center, Vienna, Austria, May 7-11t, 2024 .
Open this publication in new window or tab >>Robust Classification via Regression for Learning with Noisy Labels
2024 (English)In: Proceedings ICLR 2024 - The Twelfth International Conference on Learning Representations, 2024Conference paper, Poster (with or without abstract) (Refereed)
Abstract [en]

Deep neural networks and large-scale datasets have revolutionized the field of machine learning. However, these large networks are susceptible to overfitting to label noise, resulting in reduced generalization. To address this challenge, two promising approaches have emerged: i) loss reweighting, which reduces the influence of noisy examples on the training loss, and ii) label correction that replaces noisy labels with estimated true labels. These directions have been pursued separately or combined as independent methods, lacking a unified approach. In this work, we present a unified method that seamlessly combines loss reweighting and label correction to enhance robustness against label noise in classification tasks. Specifically, by leveraging ideas from compositional data analysis in statistics, we frame the problem as a regression task, where loss reweighting and label correction can naturally be achieved with a shifted Gaussian label noise model. Our unified approach achieves strong performance compared to recent baselines on several noisy labelled datasets. We believe this work is a promising step towards robust deep learning in the presence of label noise. Our code is available at: https://github.com/ErikEnglesson/SGN.

Keywords
label noise, noisy labels, robustness, Gaussian noise, classification, log-ratio transform, compositional data analysis
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-346452 (URN)
Conference
ICLR 2024 - The Twelfth International Conference on Learning Representations, Messe Wien Exhibition and Congress Center, Vienna, Austria, May 7-11t, 2024 
Note

QC 20240515

Available from: 2024-05-15 Created: 2024-05-15 Last updated: 2024-05-16Bibliographically approved
Englesson, E. & Azizpour, H. (2024). Robust classification via regression for learning with noisy labels. In: 12th International Conference on Learning Representations, ICLR 2024: . Paper presented at 12th International Conference on Learning Representations, ICLR 2024, Hybrid, Vienna, Austria, May 7-11, 2024. International Conference on Learning Representations, ICLR
Open this publication in new window or tab >>Robust classification via regression for learning with noisy labels
2024 (English)In: 12th International Conference on Learning Representations, ICLR 2024, International Conference on Learning Representations, ICLR , 2024Conference paper, Published paper (Refereed)
Abstract [en]

Deep neural networks and large-scale datasets have revolutionized the field of machine learning. However, these large networks are susceptible to overfitting to label noise, resulting in reduced generalization. To address this challenge, two promising approaches have emerged: i) loss reweighting, which reduces the influence of noisy examples on the training loss, and ii) label correction that replaces noisy labels with estimated true labels. These directions have been pursued separately or combined as independent methods, lacking a unified approach. In this work, we present a unified method that seamlessly combines loss reweighting and label correction to enhance robustness against label noise in classification tasks. Specifically, by leveraging ideas from compositional data analysis in statistics, we frame the problem as a regression task, where loss reweighting and label correction can naturally be achieved with a shifted Gaussian label noise model. Our unified approach achieves strong performance compared to recent baselines on several noisy labelled datasets. We believe this work is a promising step towards robust deep learning in the presence of label noise. Our code is available at: github.com/ErikEnglesson/SGN.

Place, publisher, year, edition, pages
International Conference on Learning Representations, ICLR, 2024
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-367431 (URN)2-s2.0-85190096539 (Scopus ID)
Conference
12th International Conference on Learning Representations, ICLR 2024, Hybrid, Vienna, Austria, May 7-11, 2024
Note

QC 20250718

Available from: 2025-07-18 Created: 2025-07-18 Last updated: 2025-07-18Bibliographically approved
Gamba, M., Englesson, E., Björkman, M. & Azizpour, H. (2023). Deep Double Descent via Smooth Interpolation. Transactions on Machine Learning Research, 2023(4)
Open this publication in new window or tab >>Deep Double Descent via Smooth Interpolation
2023 (English)In: Transactions on Machine Learning Research, E-ISSN 2835-8856, Vol. 2023, no 4Article in journal (Refereed) Published
Abstract [en]

The ability of overparameterized deep networks to interpolate noisy data, while at the same time showing good generalization performance, has been recently characterized in terms of the double descent curve for the test error. Common intuition from polynomial regression suggests that overparameterized networks are able to sharply interpolate noisy data, without considerably deviating from the ground-truth signal, thus preserving generalization ability. At present, a precise characterization of the relationship between interpolation and generalization for deep networks is missing. In this work, we quantify sharpness of fit of the training data interpolated by neural network functions, by studying the loss landscape w.r.t. to the input variable locally to each training point, over volumes around cleanly- and noisily-labelled training samples, as we systematically increase the number of model parameters and training epochs. Our findings show that loss sharpness in the input space follows both model- and epoch-wise double descent, with worse peaks observed around noisy labels. While small interpolating models sharply fit both clean and noisy data, large interpolating models express a smooth loss landscape, where noisy targets are predicted over large volumes around training data points, in contrast to existing intuition.

Place, publisher, year, edition, pages
Transactions on Machine Learning Research (TMLR), 2023
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-346450 (URN)2-s2.0-86000152632 (Scopus ID)
Note

QC 20250320

Available from: 2024-05-15 Created: 2024-05-15 Last updated: 2025-03-20Bibliographically approved
Englesson, E., Mehrpanah, A. & Azizpour, H. (2023). Logistic-Normal Likelihoods for Heteroscedastic Label Noise. Transactions on Machine Learning Research, 2023(8)
Open this publication in new window or tab >>Logistic-Normal Likelihoods for Heteroscedastic Label Noise
2023 (English)In: Transactions on Machine Learning Research, E-ISSN 2835-8856, Vol. 2023, no 8Article in journal (Refereed) Published
Abstract [en]

A natural way of estimating heteroscedastic label noise in regression is to model the observed (potentially noisy) target as a sample from a normal distribution, whose parameters can be learned by minimizing the negative log-likelihood. This formulation has desirable loss attenuation properties, as it reduces the contribution of high-error examples. Intuitively, this behavior can improve robustness against label noise by reducing overfitting. We propose an extension of this simple and probabilistic approach to classification that has the same desirable loss attenuation properties. Furthermore, we discuss and address some practical challenges of this extension. We evaluate the effectiveness of the method by measuring its robustness against label noise in classification. We perform enlightening experiments exploring the inner workings of the method, including sensitivity to hyperparameters, ablation studies, and other insightful analyses.

Place, publisher, year, edition, pages
Transactions on Machine Learning Research (TMLR), 2023
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-346451 (URN)2-s2.0-86000109470 (Scopus ID)
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

QC 20250325

Available from: 2024-05-15 Created: 2024-05-15 Last updated: 2025-03-25Bibliographically approved
Englesson, E. & Azizpour, H. (2021). Consistency Regularization Can Improve Robustness to Label Noise. In: : . Paper presented at International Conference on Machine Learning (ICML) Workshops, 2021 Workshop on Uncertainty and Robustness in Deep Learning.
Open this publication in new window or tab >>Consistency Regularization Can Improve Robustness to Label Noise
2021 (English)Conference paper, Poster (with or without abstract) (Refereed)
Abstract [en]

Consistency regularization is a commonly-used technique for semi-supervised and self-supervised learning. It is an auxiliary objective function that encourages the prediction of the network to be similar in the vicinity of the observed training samples. Hendrycks et al. (2020) have recently shown such regularization naturally brings test-time robustness to corrupted data and helps with calibration. This paper empirically studies the relevance of consistency regularization for training-time robustness to noisy labels. First, we make two interesting and useful observations regarding the consistency of networks trained with the standard cross entropy loss on noisy datasets which are: (i) networks trained on noisy data have lower consistency than those trained on clean data, and (ii) the consistency reduces more significantly around noisy-labelled training data points than correctly-labelled ones. Then, we show that a simple loss function that encourages consistency improves the robustness of the models to label noise on both synthetic (CIFAR-10, CIFAR-100) and real-world (WebVision) noise as well as different noise rates and types and achieves state-of-the-art results.

National Category
Computer graphics and computer vision
Identifiers
urn:nbn:se:kth:diva-305932 (URN)
Conference
International Conference on Machine Learning (ICML) Workshops, 2021 Workshop on Uncertainty and Robustness in Deep Learning
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

QC 20211220

Available from: 2021-12-09 Created: 2021-12-09 Last updated: 2025-02-07Bibliographically approved
Englesson, E. & Azizpour, H. (2021). Generalized Jensen-Shannon Divergence Loss for Learning with Noisy Labels. In: Proceedings 35th Conference on Neural Information Processing Systems (NeurIPS 2021).: . Paper presented at 35th Conference on Neural Information Processing Systems, NeurIPS 2021, Virtual, Online, Dec 14 2021 - Dec 6 2021. NIPS
Open this publication in new window or tab >>Generalized Jensen-Shannon Divergence Loss for Learning with Noisy Labels
2021 (English)In: Proceedings 35th Conference on Neural Information Processing Systems (NeurIPS 2021)., NIPS , 2021Conference paper, Published paper (Refereed)
Abstract [en]

Prior works have found it beneficial to combine provably noise-robust loss functions e.g. mean absolute error (MAE) with standard categorical loss function e.g. crossentropy (CE) to improve their learnability. Here, we propose to use Jensen-Shannondivergence as a noise-robust loss function and show that it interestingly interpolatebetween CE and MAE with a controllable mixing parameter. Furthermore, wemake a crucial observation that CE exhibits lower consistency around noisy datapoints. Based on this observation, we adopt a generalized version of the Jensen-Shannon divergence for multiple distributions to encourage consistency arounddata points. Using this loss function, we show state-of-the-art results on bothsynthetic (CIFAR), and real-world (e.g. WebVision) noise with varying noise rates.

Place, publisher, year, edition, pages
NIPS, 2021
National Category
Computer graphics and computer vision
Identifiers
urn:nbn:se:kth:diva-305931 (URN)000922928202022 ()2-s2.0-85124784124 (Scopus ID)
Conference
35th Conference on Neural Information Processing Systems, NeurIPS 2021, Virtual, Online, Dec 14 2021 - Dec 6 2021
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

Part of proceedings ISBN 978-171384539-3 

QC 20211220

Available from: 2021-12-09 Created: 2021-12-09 Last updated: 2025-02-07Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0003-4535-2520

Search in DiVA

Show all publications