kth.sePublikationer KTH
Ändra sökning
Länk till posten
Permanent länk

Direktlänk
Publikationer (10 of 11) Visa alla publikationer
Hakkinen, I., Melekhov, I., Englesson, E., Azizpour, H. & Kannala, J. (2025). Medical Image Segmentation with SAM-Generated Annotations. In: DelBue, A Canton, C Pont-Tuset, J Tommasi, T (Ed.), Computer Vision-Eccv 2024 Workshops, Pt Xxii: . Paper presented at 18th European Conference on Computer Vision (ECCV), Sep 29- 04, 2024, Milan, Italy (pp. 51-62). Springer Nature, 15644
Öppna denna publikation i ny flik eller fönster >>Medical Image Segmentation with SAM-Generated Annotations
Visa övriga...
2025 (Engelska)Ingår i: Computer Vision-Eccv 2024 Workshops, Pt Xxii / [ed] DelBue, A Canton, C Pont-Tuset, J Tommasi, T, Springer Nature , 2025, Vol. 15644, s. 51-62Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

The field of medical image segmentation is hindered by the scarcity of large, publicly available annotated datasets. Not all datasets are made public for privacy reasons, and creating annotations for a large dataset is time-consuming and expensive, as it requires specialized expertise to accurately identify regions of interest (ROIs) within the images. To address these challenges, we evaluate the performance of the Segment Anything Model (SAM) as an annotation tool for medical data by using it to produce so-called "pseudo labels" on the Medical Segmentation Decathlon (MSD) computed tomography (CT) tasks. The pseudo labels are then used in place of ground truth labels to train a UNet model in a weakly-supervised manner. We experiment with different prompt types on SAM and find that the bounding box prompt is a simple yet effective method for generating pseudo labels. This method allows us to develop a weakly-supervised model that performs comparably to a fully supervised model.

Ort, förlag, år, upplaga, sidor
Springer Nature, 2025
Serie
Lecture Notes in Computer Science, ISSN 0302-9743
Nyckelord
Foundation Model, Segment Anything Model, Medical Image Segmentation, Data Annotation
Nationell ämneskategori
Datavetenskap (datalogi)
Identifikatorer
urn:nbn:se:kth:diva-374359 (URN)10.1007/978-3-031-92089-9_4 (DOI)001544995100004 ()2-s2.0-105006891572 (Scopus ID)
Konferens
18th European Conference on Computer Vision (ECCV), Sep 29- 04, 2024, Milan, Italy
Anmärkning

Part of ISBN 978-3-031-92088-2; 978-3-031-92089-9

QC 20251218

Tillgänglig från: 2025-12-18 Skapad: 2025-12-18 Senast uppdaterad: 2025-12-18Bibliografiskt granskad
Mehrpanah, A., Englesson, E. & Azizpour, H. (2025). On Spectral Properties of Gradient-Based Explanation Methods. In: Computer Vision – ECCV 2024 - 18th European Conference, Proceedings: . Paper presented at 18th European Conference on Computer Vision, ECCV 2024, Milan, Italy, Sep 29 2024 - Oct 4 2024 (pp. 282-299). Springer Nature
Öppna denna publikation i ny flik eller fönster >>On Spectral Properties of Gradient-Based Explanation Methods
2025 (Engelska)Ingår i: Computer Vision – ECCV 2024 - 18th European Conference, Proceedings, Springer Nature , 2025, s. 282-299Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

Understanding the behavior of deep networks is crucial to increase our confidence in their results. Despite an extensive body of work for explaining their predictions, researchers have faced reliability issues, which can be attributed to insufficient formalism. In our research, we adopt novel probabilistic and spectral perspectives to formally analyze explanation methods. Our study reveals a pervasive spectral bias stemming from the use of gradient, and sheds light on some common design choices that have been discovered experimentally, in particular, the use of squared gradient and input perturbation. We further characterize how the choice of perturbation hyperparameters in explanation methods, such as SmoothGrad, can lead to inconsistent explanations and introduce two remedies based on our proposed formalism: (i) a mechanism to determine a standard perturbation scale, and (ii) an aggregation method which we call SpectralLens. Finally, we substantiate our theoretical results through quantitative evaluations.

Ort, förlag, år, upplaga, sidor
Springer Nature, 2025
Nyckelord
Deep Neural Networks, Explainability, Gradient-based Explanation Methods, Probabilistic Machine Learning, Probabilistic Pixel Attribution Techniques, Spectral Analysis
Nationell ämneskategori
Datavetenskap (datalogi)
Identifikatorer
urn:nbn:se:kth:diva-357698 (URN)10.1007/978-3-031-73021-4_17 (DOI)001416940200017 ()2-s2.0-85210488897 (Scopus ID)
Konferens
18th European Conference on Computer Vision, ECCV 2024, Milan, Italy, Sep 29 2024 - Oct 4 2024
Anmärkning

Part of ISBN 978-303173020-7

QC 20241213

Tillgänglig från: 2024-12-12 Skapad: 2024-12-12 Senast uppdaterad: 2025-03-17Bibliografiskt granskad
Nilsson, A., Wijk, K., Gutha, S. b., Englesson, E., Hotti, A., Saccardi, C., . . . Azizpour, H. (2024). Indirectly Parameterized Concrete Autoencoders. In: International Conference on Machine Learning, ICML 2024: . Paper presented at 41st International Conference on Machine Learning, ICML 2024, Vienna, Austria, Jul 21 2024 - Jul 27 2024 (pp. 38237-38252). ML Research Press
Öppna denna publikation i ny flik eller fönster >>Indirectly Parameterized Concrete Autoencoders
Visa övriga...
2024 (Engelska)Ingår i: International Conference on Machine Learning, ICML 2024, ML Research Press , 2024, s. 38237-38252Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

Feature selection is a crucial task in settings where data is high-dimensional or acquiring the full set of features is costly. Recent developments in neural network-based embedded feature selection show promising results across a wide range of applications. Concrete Autoencoders (CAEs), considered state-of-the-art in embedded feature selection, may struggle to achieve stable joint optimization, hurting their training time and generalization. In this work, we identify that this instability is correlated with the CAE learning duplicate selections. To remedy this, we propose a simple and effective improvement: Indirectly Parameterized CAEs (IP-CAEs). IP-CAEs learn an embedding and a mapping from it to the Gumbel-Softmax distributions' parameters. Despite being simple to implement, IP-CAE exhibits significant and consistent improvements over CAE in both generalization and training time across several datasets for reconstruction and classification. Unlike CAE, IP-CAE effectively leverages non-linear relationships and does not require retraining the jointly optimized decoder. Furthermore, our approach is, in principle, generalizable to Gumbel-Softmax distributions beyond feature selection.

Ort, förlag, år, upplaga, sidor
ML Research Press, 2024
Nationell ämneskategori
Datavetenskap (datalogi)
Identifikatorer
urn:nbn:se:kth:diva-353956 (URN)2-s2.0-85203808876 (Scopus ID)
Konferens
41st International Conference on Machine Learning, ICML 2024, Vienna, Austria, Jul 21 2024 - Jul 27 2024
Anmärkning

QC 20240926

Tillgänglig från: 2024-09-25 Skapad: 2024-09-25 Senast uppdaterad: 2024-09-26Bibliografiskt granskad
Englesson, E. (2024). On Label Noise in Image Classification: An Aleatoric Uncertainty Perspective. (Doctoral dissertation). Stockholm: KTH Royal Institute of Technology
Öppna denna publikation i ny flik eller fönster >>On Label Noise in Image Classification: An Aleatoric Uncertainty Perspective
2024 (Engelska)Doktorsavhandling, sammanläggning (Övrigt vetenskapligt)
Abstract [en]

Deep neural networks and large-scale datasets have revolutionized the field of machine learning. However, these large networks are susceptible to overfitting to label noise, resulting in generalization degradation. In response, the thesis closely examines the problem both from an empirical and theoretical perspective. We empirically analyse the input smoothness of networks as they overfit to label noise, and we theoretically explore the connection to aleatoric uncertainty. These analyses improve our understanding of the problem and have led to our novel methods aimed at enhancing robustness against label noise in classification.

Abstract [sv]

Djupa neurala nätverk och storskaliga dataset har revolutionerat maskininlärningsområdet. Dock är dessa stora nätverk känsliga för överanpassning till felmarkerade etiketter, vilket leder till försämrad generalisering. Som svar på detta undersöker avhandlingen noggrant problemet både från en empirisk och teoretisk synvinkel. Vi analyserar empiriskt nätverkens känslighet försmå ändringar i indatan när de över anpassar till felmarkerade etiketter, och vi utforskar teoretiskt kopplingen till aleatorisk osäkerhet. Dessa analyser förbättrar vår förståelse av problemet och har lett till våra nya metoder med syfte att vara robusta mot felmarkerade etiketter i klassificering.

Ort, förlag, år, upplaga, sidor
Stockholm: KTH Royal Institute of Technology, 2024. s. xi, 68
Serie
TRITA-EECS-AVL ; 2024:45
Nyckelord
Label noise, aleatoric uncertainty, noisy labels, robustness, etikettbrus, osäkerhet, felmarkerade etiketter, robusthet
Nationell ämneskategori
Datavetenskap (datalogi)
Forskningsämne
Datalogi
Identifikatorer
urn:nbn:se:kth:diva-346453 (URN)978-91-8040-925-4 (ISBN)
Disputation
2024-06-03, https://kth-se.zoom.us/w/61097277235, F3 (Flodis), Lindstedsvägen 26 & 28, Stockholm, 09:00 (Engelska)
Opponent
Handledare
Forskningsfinansiär
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Anmärkning

QC 20240516

Tillgänglig från: 2024-05-16 Skapad: 2024-05-16 Senast uppdaterad: 2025-12-03Bibliografiskt granskad
Englesson, E. & Azizpour, H. (2024). Robust Classification via Regression for Learning with Noisy Labels. In: Proceedings ICLR 2024 - The Twelfth International Conference on Learning Representations: . Paper presented at ICLR 2024 - The Twelfth International Conference on Learning Representations, Messe Wien Exhibition and Congress Center, Vienna, Austria, May 7-11t, 2024 .
Öppna denna publikation i ny flik eller fönster >>Robust Classification via Regression for Learning with Noisy Labels
2024 (Engelska)Ingår i: Proceedings ICLR 2024 - The Twelfth International Conference on Learning Representations, 2024Konferensbidrag, Poster (med eller utan abstract) (Refereegranskat)
Abstract [en]

Deep neural networks and large-scale datasets have revolutionized the field of machine learning. However, these large networks are susceptible to overfitting to label noise, resulting in reduced generalization. To address this challenge, two promising approaches have emerged: i) loss reweighting, which reduces the influence of noisy examples on the training loss, and ii) label correction that replaces noisy labels with estimated true labels. These directions have been pursued separately or combined as independent methods, lacking a unified approach. In this work, we present a unified method that seamlessly combines loss reweighting and label correction to enhance robustness against label noise in classification tasks. Specifically, by leveraging ideas from compositional data analysis in statistics, we frame the problem as a regression task, where loss reweighting and label correction can naturally be achieved with a shifted Gaussian label noise model. Our unified approach achieves strong performance compared to recent baselines on several noisy labelled datasets. We believe this work is a promising step towards robust deep learning in the presence of label noise. Our code is available at: https://github.com/ErikEnglesson/SGN.

Nyckelord
label noise, noisy labels, robustness, Gaussian noise, classification, log-ratio transform, compositional data analysis
Nationell ämneskategori
Datavetenskap (datalogi)
Identifikatorer
urn:nbn:se:kth:diva-346452 (URN)
Konferens
ICLR 2024 - The Twelfth International Conference on Learning Representations, Messe Wien Exhibition and Congress Center, Vienna, Austria, May 7-11t, 2024 
Anmärkning

QC 20240515

Tillgänglig från: 2024-05-15 Skapad: 2024-05-15 Senast uppdaterad: 2024-05-16Bibliografiskt granskad
Englesson, E. & Azizpour, H. (2024). Robust classification via regression for learning with noisy labels. In: 12th International Conference on Learning Representations, ICLR 2024: . Paper presented at 12th International Conference on Learning Representations, ICLR 2024, Hybrid, Vienna, Austria, May 7-11, 2024. International Conference on Learning Representations, ICLR
Öppna denna publikation i ny flik eller fönster >>Robust classification via regression for learning with noisy labels
2024 (Engelska)Ingår i: 12th International Conference on Learning Representations, ICLR 2024, International Conference on Learning Representations, ICLR , 2024Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

Deep neural networks and large-scale datasets have revolutionized the field of machine learning. However, these large networks are susceptible to overfitting to label noise, resulting in reduced generalization. To address this challenge, two promising approaches have emerged: i) loss reweighting, which reduces the influence of noisy examples on the training loss, and ii) label correction that replaces noisy labels with estimated true labels. These directions have been pursued separately or combined as independent methods, lacking a unified approach. In this work, we present a unified method that seamlessly combines loss reweighting and label correction to enhance robustness against label noise in classification tasks. Specifically, by leveraging ideas from compositional data analysis in statistics, we frame the problem as a regression task, where loss reweighting and label correction can naturally be achieved with a shifted Gaussian label noise model. Our unified approach achieves strong performance compared to recent baselines on several noisy labelled datasets. We believe this work is a promising step towards robust deep learning in the presence of label noise. Our code is available at: github.com/ErikEnglesson/SGN.

Ort, förlag, år, upplaga, sidor
International Conference on Learning Representations, ICLR, 2024
Nationell ämneskategori
Datavetenskap (datalogi)
Identifikatorer
urn:nbn:se:kth:diva-367431 (URN)2-s2.0-85190096539 (Scopus ID)
Konferens
12th International Conference on Learning Representations, ICLR 2024, Hybrid, Vienna, Austria, May 7-11, 2024
Anmärkning

QC 20250718

Tillgänglig från: 2025-07-18 Skapad: 2025-07-18 Senast uppdaterad: 2025-07-18Bibliografiskt granskad
Gamba, M., Englesson, E., Björkman, M. & Azizpour, H. (2023). Deep Double Descent via Smooth Interpolation. Transactions on Machine Learning Research, 2023(4)
Öppna denna publikation i ny flik eller fönster >>Deep Double Descent via Smooth Interpolation
2023 (Engelska)Ingår i: Transactions on Machine Learning Research, E-ISSN 2835-8856, Vol. 2023, nr 4Artikel i tidskrift (Refereegranskat) Published
Abstract [en]

The ability of overparameterized deep networks to interpolate noisy data, while at the same time showing good generalization performance, has been recently characterized in terms of the double descent curve for the test error. Common intuition from polynomial regression suggests that overparameterized networks are able to sharply interpolate noisy data, without considerably deviating from the ground-truth signal, thus preserving generalization ability. At present, a precise characterization of the relationship between interpolation and generalization for deep networks is missing. In this work, we quantify sharpness of fit of the training data interpolated by neural network functions, by studying the loss landscape w.r.t. to the input variable locally to each training point, over volumes around cleanly- and noisily-labelled training samples, as we systematically increase the number of model parameters and training epochs. Our findings show that loss sharpness in the input space follows both model- and epoch-wise double descent, with worse peaks observed around noisy labels. While small interpolating models sharply fit both clean and noisy data, large interpolating models express a smooth loss landscape, where noisy targets are predicted over large volumes around training data points, in contrast to existing intuition.

Ort, förlag, år, upplaga, sidor
Transactions on Machine Learning Research (TMLR), 2023
Nationell ämneskategori
Datavetenskap (datalogi)
Identifikatorer
urn:nbn:se:kth:diva-346450 (URN)2-s2.0-86000152632 (Scopus ID)
Anmärkning

QC 20250320

Tillgänglig från: 2024-05-15 Skapad: 2024-05-15 Senast uppdaterad: 2025-03-20Bibliografiskt granskad
Englesson, E., Mehrpanah, A. & Azizpour, H. (2023). Logistic-Normal Likelihoods for Heteroscedastic Label Noise. Transactions on Machine Learning Research, 2023(8)
Öppna denna publikation i ny flik eller fönster >>Logistic-Normal Likelihoods for Heteroscedastic Label Noise
2023 (Engelska)Ingår i: Transactions on Machine Learning Research, E-ISSN 2835-8856, Vol. 2023, nr 8Artikel i tidskrift (Refereegranskat) Published
Abstract [en]

A natural way of estimating heteroscedastic label noise in regression is to model the observed (potentially noisy) target as a sample from a normal distribution, whose parameters can be learned by minimizing the negative log-likelihood. This formulation has desirable loss attenuation properties, as it reduces the contribution of high-error examples. Intuitively, this behavior can improve robustness against label noise by reducing overfitting. We propose an extension of this simple and probabilistic approach to classification that has the same desirable loss attenuation properties. Furthermore, we discuss and address some practical challenges of this extension. We evaluate the effectiveness of the method by measuring its robustness against label noise in classification. We perform enlightening experiments exploring the inner workings of the method, including sensitivity to hyperparameters, ablation studies, and other insightful analyses.

Ort, förlag, år, upplaga, sidor
Transactions on Machine Learning Research (TMLR), 2023
Nationell ämneskategori
Datavetenskap (datalogi)
Identifikatorer
urn:nbn:se:kth:diva-346451 (URN)2-s2.0-86000109470 (Scopus ID)
Forskningsfinansiär
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Anmärkning

QC 20250325

Tillgänglig från: 2024-05-15 Skapad: 2024-05-15 Senast uppdaterad: 2025-03-25Bibliografiskt granskad
Englesson, E. & Azizpour, H. (2021). Consistency Regularization Can Improve Robustness to Label Noise. In: : . Paper presented at International Conference on Machine Learning (ICML) Workshops, 2021 Workshop on Uncertainty and Robustness in Deep Learning.
Öppna denna publikation i ny flik eller fönster >>Consistency Regularization Can Improve Robustness to Label Noise
2021 (Engelska)Konferensbidrag, Poster (med eller utan abstract) (Refereegranskat)
Abstract [en]

Consistency regularization is a commonly-used technique for semi-supervised and self-supervised learning. It is an auxiliary objective function that encourages the prediction of the network to be similar in the vicinity of the observed training samples. Hendrycks et al. (2020) have recently shown such regularization naturally brings test-time robustness to corrupted data and helps with calibration. This paper empirically studies the relevance of consistency regularization for training-time robustness to noisy labels. First, we make two interesting and useful observations regarding the consistency of networks trained with the standard cross entropy loss on noisy datasets which are: (i) networks trained on noisy data have lower consistency than those trained on clean data, and (ii) the consistency reduces more significantly around noisy-labelled training data points than correctly-labelled ones. Then, we show that a simple loss function that encourages consistency improves the robustness of the models to label noise on both synthetic (CIFAR-10, CIFAR-100) and real-world (WebVision) noise as well as different noise rates and types and achieves state-of-the-art results.

Nationell ämneskategori
Datorgrafik och datorseende
Identifikatorer
urn:nbn:se:kth:diva-305932 (URN)
Konferens
International Conference on Machine Learning (ICML) Workshops, 2021 Workshop on Uncertainty and Robustness in Deep Learning
Forskningsfinansiär
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Anmärkning

QC 20211220

Tillgänglig från: 2021-12-09 Skapad: 2021-12-09 Senast uppdaterad: 2025-02-07Bibliografiskt granskad
Englesson, E. & Azizpour, H. (2021). Generalized Jensen-Shannon Divergence Loss for Learning with Noisy Labels. In: Proceedings 35th Conference on Neural Information Processing Systems (NeurIPS 2021).: . Paper presented at 35th Conference on Neural Information Processing Systems, NeurIPS 2021, Virtual, Online, Dec 14 2021 - Dec 6 2021. NIPS
Öppna denna publikation i ny flik eller fönster >>Generalized Jensen-Shannon Divergence Loss for Learning with Noisy Labels
2021 (Engelska)Ingår i: Proceedings 35th Conference on Neural Information Processing Systems (NeurIPS 2021)., NIPS , 2021Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

Prior works have found it beneficial to combine provably noise-robust loss functions e.g. mean absolute error (MAE) with standard categorical loss function e.g. crossentropy (CE) to improve their learnability. Here, we propose to use Jensen-Shannondivergence as a noise-robust loss function and show that it interestingly interpolatebetween CE and MAE with a controllable mixing parameter. Furthermore, wemake a crucial observation that CE exhibits lower consistency around noisy datapoints. Based on this observation, we adopt a generalized version of the Jensen-Shannon divergence for multiple distributions to encourage consistency arounddata points. Using this loss function, we show state-of-the-art results on bothsynthetic (CIFAR), and real-world (e.g. WebVision) noise with varying noise rates.

Ort, förlag, år, upplaga, sidor
NIPS, 2021
Nationell ämneskategori
Datorgrafik och datorseende
Identifikatorer
urn:nbn:se:kth:diva-305931 (URN)000922928202022 ()2-s2.0-85124784124 (Scopus ID)
Konferens
35th Conference on Neural Information Processing Systems, NeurIPS 2021, Virtual, Online, Dec 14 2021 - Dec 6 2021
Forskningsfinansiär
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Anmärkning

Part of proceedings ISBN 978-171384539-3 

QC 20211220

Tillgänglig från: 2021-12-09 Skapad: 2021-12-09 Senast uppdaterad: 2025-02-07Bibliografiskt granskad
Organisationer
Identifikatorer
ORCID-id: ORCID iD iconorcid.org/0000-0003-4535-2520

Sök vidare i DiVA

Visa alla publikationer