Medical Image Segmentation with SAM-Generated AnnotationsShow others and affiliations
2025 (English)In: Computer Vision-Eccv 2024 Workshops, Pt Xxii / [ed] DelBue, A Canton, C Pont-Tuset, J Tommasi, T, Springer Nature , 2025, Vol. 15644, p. 51-62Conference paper, Published paper (Refereed)
Abstract [en]
The field of medical image segmentation is hindered by the scarcity of large, publicly available annotated datasets. Not all datasets are made public for privacy reasons, and creating annotations for a large dataset is time-consuming and expensive, as it requires specialized expertise to accurately identify regions of interest (ROIs) within the images. To address these challenges, we evaluate the performance of the Segment Anything Model (SAM) as an annotation tool for medical data by using it to produce so-called "pseudo labels" on the Medical Segmentation Decathlon (MSD) computed tomography (CT) tasks. The pseudo labels are then used in place of ground truth labels to train a UNet model in a weakly-supervised manner. We experiment with different prompt types on SAM and find that the bounding box prompt is a simple yet effective method for generating pseudo labels. This method allows us to develop a weakly-supervised model that performs comparably to a fully supervised model.
Place, publisher, year, edition, pages
Springer Nature , 2025. Vol. 15644, p. 51-62
Series
Lecture Notes in Computer Science, ISSN 0302-9743
Keywords [en]
Foundation Model, Segment Anything Model, Medical Image Segmentation, Data Annotation
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-374359DOI: 10.1007/978-3-031-92089-9_4ISI: 001544995100004Scopus ID: 2-s2.0-105006891572OAI: oai:DiVA.org:kth-374359DiVA, id: diva2:2022925
Conference
18th European Conference on Computer Vision (ECCV), Sep 29- 04, 2024, Milan, Italy
Note
Part of ISBN 978-3-031-92088-2; 978-3-031-92089-9
QC 20251218
2025-12-182025-12-182025-12-18Bibliographically approved