PerfCam: Digital Twinning for Production Lines Using 3D Gaussian Splatting and Vision ModelsShow others and affiliations
2025 (English)In: IEEE Access, E-ISSN 2169-3536, Vol. 13, p. 85390-85406
Article in journal (Refereed) Published
Abstract [en]
We introduce PerfCam, an open source Proof-of-Concept (PoC) digital twinning framework that combines camera and sensory data with 3D Gaussian Splatting and computer vision models for digital twinning, object tracking, and Key Performance Indicators (KPIs) extraction in industrial production lines. By utilizing 3D reconstruction and Convolutional Neural Networks (CNNs), PerfCam offers a semi-automated approach to object tracking and spatial mapping, enabling highly accurate digital twins that capture real-time KPIs such as availability, performance, Overall Equipment Effectiveness (OEE), and rate of conveyor belts in the production line. We validate the effectiveness of PerfCam through a practical deployment within realistic test production lines in the pharmaceutical industry and contribute an openly published dataset to support further research and development in the field. The results demonstrate PerfCam’s ability to deliver actionable insights through its precise digital twin capabilities, underscoring its value as an effective tool for developing usable digital twins in smart manufacturing environments and extracting operational analytics.
Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2025. Vol. 13, p. 85390-85406
Keywords [en]
Production Line, Visual Model, Digital Twin, Convolutional Neural Network, Computer Vision, Sensor Data, 3D Reconstruction
National Category
Computer Sciences
Research subject
Computer Science; Industrial Engineering and Management
Identifiers
URN: urn:nbn:se:kth:diva-363250DOI: 10.1109/access.2025.3567702ISI: 001492129400039Scopus ID: 2-s2.0-105004694919OAI: oai:DiVA.org:kth-363250DiVA, id: diva2:1957350
Projects
SMART Pharmaceutical Manufacturing
Funder
AstraZeneca, KTH-RPROJ-0146472
Note
QC 20260127
2025-05-092025-05-092026-01-27Bibliographically approved