kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Towards Automated Parts Recognition in Manufacturing with Synthetic Data
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0002-4180-3809
2025 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

This thesis advances the understanding and application of synthetic data for manufacturing parts recognition. Vision-based inspection systems in manufacturing rely heavily on real image data, which are costly to collect, annotate, and adapt across products and environments. To address these challenges, this work presents a systematic investigation of how synthetic data can be effectively generated, evaluated, and applied for robust and scalable performance. The research introduces a series of new industrial benchmark datasets covering multiple manufacturing use cases and factory environments: SIP-17, SIP15-OD, and SIP2A-OD, to enable unified evaluation of sim-to-real transfer in classification and detection tasks. Building on these datasets, a domain randomization pipeline is developed to systematically explore the effects of rendering parameters, material variability, and illumination on model generalization. To further automate data generation, the thesis proposes Synthetic Active Learning (SAL), a closed-loop framework that identifies model weaknesses and adaptively refines synthetic data generation without requiring real samples or manual tuning. Experiments across the benchmark datasets show that the proposed method improves model robustness compared to existing approaches while reducing manual labeling requirements. Collectively, these contributions provide new insights into how synthetic data can be systematically leveraged to build data-efficient, automated, and reliable vision systems for manufacturing, aiming to support future development of intelligent and flexible production systems.

Abstract [sv]

 Denna avhandling bidrar till förståelsen och tillämpningen av syntetisk data för igenkänning av tillverkningskomponenter. Visionsbaserade inspektionssystem inom industrin är starkt beroende av verkliga bilddata, vilka är kostsamma att samla in, annotera och anpassa mellan olika produkter och miljöer. För att hantera dessa utmaningar presenterar arbetet en systematisk undersökning av hur syntetisk data effektivt kan genereras, utvärderas och tillämpas för robust och skalbar prestanda. Forskningen introducerar en serie nya industriella benchmark-datamängder som täcker flera användningsfall och fabriksmiljöer: SIP-17, SIP15-OD ochSIP2A-OD, för att möjliggöra en enhetlig utvärdering av sim-to-real-överföring inom klassificerings- och detektionsuppgifter. Med utgångspunkt i dessa datamängder utvecklas en domain randomizationpipeline som systematiskt undersöker effekterna av renderingsparametrar, materialvariation och belysning på modellens generaliseringsförmåga. För att ytterligare automatisera datagenereringen föreslås Synthetic Active Learning (SAL), ett slutet ramverk som identifierar modellens svagheter och adaptivt förfinar den syntetiska datagenereringen utan att kräva verkliga prover eller manuell justering. Experiment över benchmark-datamängderna visar att de föreslagna metoderna förbättrar modellens robusthet jämfört med befintliga tillvägagångssätt, samtidigt som behovet av manuell annotering minskar. Sammantaget ger dessa bidrag nya insikter i hur syntetisk data systematiskt kan utnyttjas för att bygga dataeffektiva, automatiserade och tillförlitliga visionssystem för tillverkningsindustrin, med målet att stödja utvecklingen av framtida intelligenta och flexibla produktionssystem.

Place, publisher, year, edition, pages
Stockholm: KTH Royal Institute of Technology, 2025. , p. xix, 75
Series
TRITA-EECS-AVL ; 2025:105
Keywords [en]
Synthetic Data, Sim-to-real Transfer, Object Detection, Domain Randomization, Ac tive Learning, Dataset Generation, Automation, Computer Vision, Industrial Inspec tion
Keywords [sv]
Syntetisk Data, Sim-to-real-överföring, Objektigenkänning, Domain Randomization, Aktivt Lärande, Datagenerering, Automation, Datorseende, Industriell Inspektion
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-372644ISBN: 978-91-8106-466-7 (print)OAI: oai:DiVA.org:kth-372644DiVA, id: diva2:2013003
Public defence
2025-12-09, https://kth-se.zoom.us/j/62913476523, Kollegiesalen, Brinellvägen 8, Stockholm, 13:00 (English)
Opponent
Supervisors
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

QC 20251112

Available from: 2025-11-12 Created: 2025-11-11 Last updated: 2025-11-24Bibliographically approved
List of papers
1. Towards sim-to-real industrial parts classification with synthetic dataset
Open this publication in new window or tab >>Towards sim-to-real industrial parts classification with synthetic dataset
Show others...
2023 (English)In: Proceedings: 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2023, Institute of Electrical and Electronics Engineers (IEEE) , 2023, p. 4454-4463Conference paper, Published paper (Refereed)
Abstract [en]

This paper is about effectively utilizing synthetic data for training deep neural networks for industrial parts classification, in particular, by taking into account the domain gap against real-world images. To this end, we introduce a synthetic dataset that may serve as a preliminary testbed for the Sim-to-Real challenge; it contains 17 objects of six industrial use cases, including isolated and assembled parts. A few subsets of objects exhibit large similarities in shape and albedo for reflecting challenging cases of industrial parts. All the sample images come with and without random backgrounds and post-processing for evaluating the importance of domain randomization. We call it Synthetic Industrial Parts dataset (SIP-17). We study the usefulness of SIP-17 through benchmarking the performance of five state-of-the-art deep network models, supervised and self-supervised, trained only on the synthetic data while testing them on real data. By analyzing the results, we deduce some insights on the feasibility and challenges of using synthetic data for industrial parts classification and for further developing larger-scale synthetic datasets. Our dataset † and code ‡ are publicly available.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2023
National Category
Computer graphics and computer vision
Identifiers
urn:nbn:se:kth:diva-337847 (URN)10.1109/CVPRW59228.2023.00468 (DOI)2-s2.0-85170821045 (Scopus ID)
Conference
2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2023, Vancouver, Canada, Jun 18 2023 - Jun 22 2023
Note

Part of ISBN 9798350302493

QC 20231010

Available from: 2023-10-10 Created: 2023-10-10 Last updated: 2025-11-12Bibliographically approved
2. Domain Randomization for Object Detection in Manufacturing Applications Using Synthetic Data: A Comprehensive Study
Open this publication in new window or tab >>Domain Randomization for Object Detection in Manufacturing Applications Using Synthetic Data: A Comprehensive Study
Show others...
2025 (English)In: 2025 IEEE International Conference on Robotics and Automation, ICRA 2025, Institute of Electrical and Electronics Engineers (IEEE) , 2025, p. 16715-16721Conference paper, Published paper (Refereed)
Abstract [en]

This paper addresses key aspects of domain randomization in generating synthetic data for manufacturing object detection applications. To this end, we present a comprehensive data generation pipeline that reflects different factors: object characteristics, background, illumination, camera settings, and post-processing. We also introduce the Synthetic Industrial Parts Object Detection dataset (SIP15-OD) consisting of 15 objects from three industrial use cases under varying environments as a test bed for the study, while also employing an industrial dataset publicly available for robotic applications. In our experiments, we present more abundant results and insights into the feasibility as well as challenges of sim-toreal object detection. In particular, we identified material properties, rendering methods, post-processing, and distractors as important factors. Our method, leveraging these, achieves top performance on the public dataset with Yolov8 models trained exclusively on synthetic data; mAP@50 scores of 96.4% for the robotics dataset, and 94.1%, 99.5%, and 95.3% across three of the SIP15-OD use cases, respectively. The results showcase the effectiveness of the proposed domain randomization, potentially covering the distribution close to real data for the applications.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
National Category
Computer graphics and computer vision Computer Sciences
Identifiers
urn:nbn:se:kth:diva-371386 (URN)10.1109/ICRA55743.2025.11128647 (DOI)2-s2.0-105016571384 (Scopus ID)
Conference
2025 IEEE International Conference on Robotics and Automation, ICRA 2025, Atlanta, United States of America, May 19 2025 - May 23 2025
Note

Part of ISBN 9798331541392

QC 20251009

Available from: 2025-10-09 Created: 2025-10-09 Last updated: 2025-11-12Bibliographically approved
3. Towards Automated Assembly Quality Inspection with Synthetic Data and Domain Randomization
Open this publication in new window or tab >>Towards Automated Assembly Quality Inspection with Synthetic Data and Domain Randomization
Show others...
2025 (English)In: Proceedings: IEEE/CVF International Conference on Computer Vision Workshop, ICCVW, 2025, 2025, p. 1395-1403Conference paper, Published paper (Refereed)
Abstract [en]

Assembly quality inspection plays a vital role in manufacturing, where correct part placement and alignment directly affect product reliability. While deep learning–based object detection offers a promising solution for automatic assembly quality inspection, it is hindered by data scarcity. Training on synthetic data with Domain Randomization (DR) helps address this challenge, yet existing DR methods focus on generating individual objects and do not capture the relational structure needed for assembly inspection. In this paper, we identify two key factors for effective synthetic data generation in assembly inspection: preserving spatial relationships between components and providing part-level textures and annotations. We propose an Assembly-Specific Generation Scheme that incorporates these factors into a state-of-the-art DR pipeline. To evaluate its impact, we introduce SIP2A-OD, a new object detection dataset comprising two real-world assembly use cases collected under varied manufacturing conditions. We train a YOLOv12 model on synthetic data generated by our pipeline and test it on real data from the SIP2A-OD dataset. Compared to the baseline pipeline designed for individual object detection, our method improves mAP@50 by more than 15% in both use cases. These results demonstrate the effectiveness of our scheme and its potential for broader applications in industrial assembly inspection without the need for manual data collection or annotation.

National Category
Engineering and Technology
Identifiers
urn:nbn:se:kth:diva-372634 (URN)
Conference
IEEE/CVF International Conference on Computer Vision Workshop 2025, Honolulu, Hawaii, USA, October 19-25, 2025
Note

QC 20251112

Available from: 2025-11-11 Created: 2025-11-11 Last updated: 2025-11-12Bibliographically approved
4. Designing Synthetic Active Learning for Model Refinement in Manufacturing Parts Detection
Open this publication in new window or tab >>Designing Synthetic Active Learning for Model Refinement in Manufacturing Parts Detection
Show others...
(English)Manuscript (preprint) (Other academic)
National Category
Computer Vision and Learning Systems
Identifiers
urn:nbn:se:kth:diva-372636 (URN)
Note

QC 20251112

Available from: 2025-11-11 Created: 2025-11-11 Last updated: 2025-11-12Bibliographically approved
5. Automated assembly quality inspection by deep learning with 2D and 3D synthetic CAD data
Open this publication in new window or tab >>Automated assembly quality inspection by deep learning with 2D and 3D synthetic CAD data
Show others...
2025 (English)In: Journal of Intelligent Manufacturing, ISSN 0956-5515, E-ISSN 1572-8145, Vol. 36, no 4, p. 2567-2582, article id e222Article in journal (Refereed) Published
Abstract [en]

In the manufacturing industry, automatic quality inspections can lead to improved product quality and productivity. Deep learning-based computer vision technologies, with their superior performance in many applications, can be a possible solution for automatic quality inspections. However, collecting a large amount of annotated training data for deep learning is expensive and time-consuming, especially for processes involving various products and human activities such as assembly. To address this challenge, we propose a method for automated assembly quality inspection using synthetic data generated from computer-aided design (CAD) models. The method involves two steps: automatic data generation and model implementation. In the first step, we generate synthetic data in two formats: two-dimensional (2D) images and three-dimensional (3D) point clouds. In the second step, we apply different state-of-the-art deep learning approaches to the data for quality inspection, including unsupervised domain adaptation, i.e., a method of adapting models across different data distributions, and transfer learning, which transfers knowledge between related tasks. We evaluate the methods in a case study of pedal car front-wheel assembly quality inspection to identify the possible optimal approach for assembly quality inspection. Our results show that the method using Transfer Learning on 2D synthetic images achieves superior performance compared with others. Specifically, it attained 95% accuracy through fine-tuning with only five annotated real images per class. With promising results, our method may be suggested for other similar quality inspection use cases. By utilizing synthetic CAD data, our method reduces the need for manual data collection and annotation. Furthermore, our method performs well on test data with different backgrounds, making it suitable for different manufacturing environments.

Place, publisher, year, edition, pages
Springer Nature, 2025
Keywords
Assembly quality inspection, Computer vision, Point cloud, Synthetic data, Transfer learning, Unsupervised domain adaptation
National Category
Computer Sciences Production Engineering, Human Work Science and Ergonomics
Identifiers
urn:nbn:se:kth:diva-363099 (URN)10.1007/s10845-024-02375-6 (DOI)001205028300001 ()2-s2.0-105002924620 (Scopus ID)
Note

QC 20250506

Available from: 2025-05-06 Created: 2025-05-06 Last updated: 2025-11-12Bibliographically approved

Open Access in DiVA

fulltext(22365 kB)98 downloads
File information
File name FULLTEXT01.pdfFile size 22365 kBChecksum SHA-512
293aa691bbd54a7e0a28170f90e6423421d16e6396de814645439c8edca38d3cd9990f43ad54ab3584559d6e9bf45c7e5e86ce760f7eb69bd874740f72005b73
Type fulltextMimetype application/pdf

Authority records

Zhu, Xiaomeng

Search in DiVA

By author/editor
Zhu, Xiaomeng
By organisation
Robotics, Perception and Learning, RPL
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 1453 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf