kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Domain Randomization for Object Detection in Manufacturing Applications Using Synthetic Data: A Comprehensive Study
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL. Scania CV AB, Södertälje, Sweden; KTH Royal Institute of Technology, Stockholm, Sweden.ORCID iD: 0000-0002-4180-3809
Scania CV AB, Södertälje, Sweden.
Scania CV AB, Södertälje, Sweden.
Scania CV AB, Södertälje, Sweden.
Show others and affiliations
2025 (English)In: 2025 IEEE International Conference on Robotics and Automation, ICRA 2025, Institute of Electrical and Electronics Engineers (IEEE) , 2025, p. 16715-16721Conference paper, Published paper (Refereed)
Abstract [en]

This paper addresses key aspects of domain randomization in generating synthetic data for manufacturing object detection applications. To this end, we present a comprehensive data generation pipeline that reflects different factors: object characteristics, background, illumination, camera settings, and post-processing. We also introduce the Synthetic Industrial Parts Object Detection dataset (SIP15-OD) consisting of 15 objects from three industrial use cases under varying environments as a test bed for the study, while also employing an industrial dataset publicly available for robotic applications. In our experiments, we present more abundant results and insights into the feasibility as well as challenges of sim-toreal object detection. In particular, we identified material properties, rendering methods, post-processing, and distractors as important factors. Our method, leveraging these, achieves top performance on the public dataset with Yolov8 models trained exclusively on synthetic data; mAP@50 scores of 96.4% for the robotics dataset, and 94.1%, 99.5%, and 95.3% across three of the SIP15-OD use cases, respectively. The results showcase the effectiveness of the proposed domain randomization, potentially covering the distribution close to real data for the applications.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2025. p. 16715-16721
National Category
Computer graphics and computer vision Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-371386DOI: 10.1109/ICRA55743.2025.11128647Scopus ID: 2-s2.0-105016571384OAI: oai:DiVA.org:kth-371386DiVA, id: diva2:2005180
Conference
2025 IEEE International Conference on Robotics and Automation, ICRA 2025, Atlanta, United States of America, May 19 2025 - May 23 2025
Note

Part of ISBN 9798331541392

QC 20251009

Available from: 2025-10-09 Created: 2025-10-09 Last updated: 2025-11-12Bibliographically approved
In thesis
1. Towards Automated Parts Recognition in Manufacturing with Synthetic Data
Open this publication in new window or tab >>Towards Automated Parts Recognition in Manufacturing with Synthetic Data
2025 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

This thesis advances the understanding and application of synthetic data for manufacturing parts recognition. Vision-based inspection systems in manufacturing rely heavily on real image data, which are costly to collect, annotate, and adapt across products and environments. To address these challenges, this work presents a systematic investigation of how synthetic data can be effectively generated, evaluated, and applied for robust and scalable performance. The research introduces a series of new industrial benchmark datasets covering multiple manufacturing use cases and factory environments: SIP-17, SIP15-OD, and SIP2A-OD, to enable unified evaluation of sim-to-real transfer in classification and detection tasks. Building on these datasets, a domain randomization pipeline is developed to systematically explore the effects of rendering parameters, material variability, and illumination on model generalization. To further automate data generation, the thesis proposes Synthetic Active Learning (SAL), a closed-loop framework that identifies model weaknesses and adaptively refines synthetic data generation without requiring real samples or manual tuning. Experiments across the benchmark datasets show that the proposed method improves model robustness compared to existing approaches while reducing manual labeling requirements. Collectively, these contributions provide new insights into how synthetic data can be systematically leveraged to build data-efficient, automated, and reliable vision systems for manufacturing, aiming to support future development of intelligent and flexible production systems.

Abstract [sv]

 Denna avhandling bidrar till förståelsen och tillämpningen av syntetisk data för igenkänning av tillverkningskomponenter. Visionsbaserade inspektionssystem inom industrin är starkt beroende av verkliga bilddata, vilka är kostsamma att samla in, annotera och anpassa mellan olika produkter och miljöer. För att hantera dessa utmaningar presenterar arbetet en systematisk undersökning av hur syntetisk data effektivt kan genereras, utvärderas och tillämpas för robust och skalbar prestanda. Forskningen introducerar en serie nya industriella benchmark-datamängder som täcker flera användningsfall och fabriksmiljöer: SIP-17, SIP15-OD ochSIP2A-OD, för att möjliggöra en enhetlig utvärdering av sim-to-real-överföring inom klassificerings- och detektionsuppgifter. Med utgångspunkt i dessa datamängder utvecklas en domain randomizationpipeline som systematiskt undersöker effekterna av renderingsparametrar, materialvariation och belysning på modellens generaliseringsförmåga. För att ytterligare automatisera datagenereringen föreslås Synthetic Active Learning (SAL), ett slutet ramverk som identifierar modellens svagheter och adaptivt förfinar den syntetiska datagenereringen utan att kräva verkliga prover eller manuell justering. Experiment över benchmark-datamängderna visar att de föreslagna metoderna förbättrar modellens robusthet jämfört med befintliga tillvägagångssätt, samtidigt som behovet av manuell annotering minskar. Sammantaget ger dessa bidrag nya insikter i hur syntetisk data systematiskt kan utnyttjas för att bygga dataeffektiva, automatiserade och tillförlitliga visionssystem för tillverkningsindustrin, med målet att stödja utvecklingen av framtida intelligenta och flexibla produktionssystem.

Place, publisher, year, edition, pages
Stockholm: KTH Royal Institute of Technology, 2025. p. xix, 75
Series
TRITA-EECS-AVL ; 2025:105
Keywords
Synthetic Data, Sim-to-real Transfer, Object Detection, Domain Randomization, Ac tive Learning, Dataset Generation, Automation, Computer Vision, Industrial Inspec tion, Syntetisk Data, Sim-to-real-överföring, Objektigenkänning, Domain Randomization, Aktivt Lärande, Datagenerering, Automation, Datorseende, Industriell Inspektion
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-372644 (URN)978-91-8106-466-7 (ISBN)
Public defence
2025-12-09, https://kth-se.zoom.us/j/62913476523, Kollegiesalen, Brinellvägen 8, Stockholm, 13:00 (English)
Opponent
Supervisors
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

QC 20251112

Available from: 2025-11-12 Created: 2025-11-11 Last updated: 2025-11-24Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Zhu, XiaomengBjörkman, MårtenMaki, Atsuto

Search in DiVA

By author/editor
Zhu, XiaomengBjörkman, MårtenMaki, Atsuto
By organisation
Robotics, Perception and Learning, RPL
Computer graphics and computer visionComputer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 166 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf