Endre søk
Link to record
Permanent link

Direct link
Publikasjoner (10 av 104) Visa alla publikasjoner
Zhu, X., Mårtensson, P., Hanson, L., Björkman, M. & Maki, A. (2025). Automated assembly quality inspection by deep learning with 2D and 3D synthetic CAD data. Journal of Intelligent Manufacturing, 36(4), 2567-2582, Article ID e222.
Åpne denne publikasjonen i ny fane eller vindu >>Automated assembly quality inspection by deep learning with 2D and 3D synthetic CAD data
Vise andre…
2025 (engelsk)Inngår i: Journal of Intelligent Manufacturing, ISSN 0956-5515, E-ISSN 1572-8145, Vol. 36, nr 4, s. 2567-2582, artikkel-id e222Artikkel i tidsskrift (Fagfellevurdert) Published
Abstract [en]

In the manufacturing industry, automatic quality inspections can lead to improved product quality and productivity. Deep learning-based computer vision technologies, with their superior performance in many applications, can be a possible solution for automatic quality inspections. However, collecting a large amount of annotated training data for deep learning is expensive and time-consuming, especially for processes involving various products and human activities such as assembly. To address this challenge, we propose a method for automated assembly quality inspection using synthetic data generated from computer-aided design (CAD) models. The method involves two steps: automatic data generation and model implementation. In the first step, we generate synthetic data in two formats: two-dimensional (2D) images and three-dimensional (3D) point clouds. In the second step, we apply different state-of-the-art deep learning approaches to the data for quality inspection, including unsupervised domain adaptation, i.e., a method of adapting models across different data distributions, and transfer learning, which transfers knowledge between related tasks. We evaluate the methods in a case study of pedal car front-wheel assembly quality inspection to identify the possible optimal approach for assembly quality inspection. Our results show that the method using Transfer Learning on 2D synthetic images achieves superior performance compared with others. Specifically, it attained 95% accuracy through fine-tuning with only five annotated real images per class. With promising results, our method may be suggested for other similar quality inspection use cases. By utilizing synthetic CAD data, our method reduces the need for manual data collection and annotation. Furthermore, our method performs well on test data with different backgrounds, making it suitable for different manufacturing environments.

sted, utgiver, år, opplag, sider
Springer Nature, 2025
Emneord
Assembly quality inspection, Computer vision, Point cloud, Synthetic data, Transfer learning, Unsupervised domain adaptation
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-363099 (URN)10.1007/s10845-024-02375-6 (DOI)001205028300001 ()2-s2.0-105002924620 (Scopus ID)
Merknad

QC 20250506

Tilgjengelig fra: 2025-05-06 Laget: 2025-05-06 Sist oppdatert: 2025-05-19bibliografisk kontrollert
Zhang, Y., Rajabi, N., Taleb, F., Matviienko, A., Ma, Y., Björkman, M. & Kragic, D. (2025). Mind Meets Robots: A Review of EEG-Based Brain-Robot Interaction Systems. International Journal of Human-Computer Interaction, 1-32
Åpne denne publikasjonen i ny fane eller vindu >>Mind Meets Robots: A Review of EEG-Based Brain-Robot Interaction Systems
Vise andre…
2025 (engelsk)Inngår i: International Journal of Human-Computer Interaction, ISSN 1044-7318, E-ISSN 1532-7590, s. 1-32Artikkel i tidsskrift (Fagfellevurdert) Published
Abstract [en]

Brain-robot interaction (BRI) empowers individuals to control (semi-)automated machines through brain activity, either passively or actively. In the past decade, BRI systems have advanced significantly, primarily leveraging electroencephalogram (EEG) signals. This article presents an up-to-date review of 87 curated studies published between 2018 and 2023, identifying the research landscape of EEG-based BRI systems. The review consolidates methodologies, interaction modes, application contexts, system evaluation, existing challenges, and future directions in this domain. Based on our analysis, we propose a BRI system model comprising three entities: Brain, Robot, and Interaction, depicting their internal relationships. We especially examine interaction modes between human brains and robots, an aspect not yet fully explored. Within this model, we scrutinize and classify current research, extract insights, highlight challenges, and offer recommendations for future studies. Our findings provide a structured design space for human-robot interaction (HRI), informing the development of more efficient BRI frameworks.

sted, utgiver, år, opplag, sider
Informa UK Limited, 2025
Emneord
EEG based, brain-robot interaction, interaction mode, comprehensive review
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-361866 (URN)10.1080/10447318.2025.2464915 (DOI)001446721000001 ()2-s2.0-105000309480 (Scopus ID)
Merknad

QC 20250402

Tilgjengelig fra: 2025-04-02 Laget: 2025-04-02 Sist oppdatert: 2025-04-02bibliografisk kontrollert
Tarle, M., Larsson, M., Ingeström, G., Nordström, L. & Björkman, M. (2025). Offline to Online Reinforcement Learning for Optimizing FACTS Setpoints. Paper presented at Bulk Power System Dynamics and Control - XII, June 2025, Sorrento, Italy. Sustainable Energy, Grids and Networks
Åpne denne publikasjonen i ny fane eller vindu >>Offline to Online Reinforcement Learning for Optimizing FACTS Setpoints
Vise andre…
2025 (engelsk)Inngår i: Sustainable Energy, Grids and Networks, E-ISSN 2352-4677Artikkel i tidsskrift (Fagfellevurdert) Accepted
Abstract [en]

With the growing electrification and integration of renewables, network operators face unprecedented challenges. Coordinated control of Flexible AC Transmission Systems (FACTS) setpoints using real-time optimization techniques has been proposed to substantially improve voltage and power flow control. However, optimizing the setpoints of several FACTS devices is rarely done in practice. In part, this can be derived from the challenges with model-based methods. As alternative control methods, data-driven methods based on reinforcement learning (RL) have shown great promise. However, RL has its own challenges that include data and safety during learning. Motivated by the increasing collection of data, we study an RL-based optimization of FACTS setpoints and how datasets can be leveraged for pre-training to improve safety. We demonstrate on the IEEE 14-bus and IEEE 57-bus systems that an offline to online RL algorithm can significantly reduce voltage deviations and constraint violations. The performance is compared against an RL agent learning from scratch and the original control policy that generated the dataset. Moreover, our analysis shows that dataset coverage and the amount of pre-training updates affect the performance considerably. Finally, to identify the gap to an optimal policy, the proposed approach is benchmarked against an optimal controller with perfect information.

Emneord
Decision support systems, Flexible AC Transmission Systems (FACTS), power system control, reinforcement learning
HSV kategori
Forskningsprogram
Datalogi; Elektro- och systemteknik
Identifikatorer
urn:nbn:se:kth:diva-365883 (URN)
Konferanse
Bulk Power System Dynamics and Control - XII, June 2025, Sorrento, Italy
Forskningsfinansiär
Swedish Foundation for Strategic Research, ID19-0058
Tilgjengelig fra: 2025-07-01 Laget: 2025-07-01 Sist oppdatert: 2025-07-01
Khanna, P., Naoum, A., Yadollahi, E., Björkman, M. & Smith, C. (2025). REFLEX Dataset: A Multimodal Dataset of Human Reactions to Robot Failures and Explanations. In: Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction: . Paper presented at ACM/IEEE International Conference on Human-Robot Interaction, HRI, Melbourne, Australia, March 4-6, 2025 (pp. 1032-1036). IEEE
Åpne denne publikasjonen i ny fane eller vindu >>REFLEX Dataset: A Multimodal Dataset of Human Reactions to Robot Failures and Explanations
Vise andre…
2025 (engelsk)Inngår i: Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction, IEEE , 2025, s. 1032-1036Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

This work presents REFLEX: Robotic Explanations to FaiLures and Human EXpressions, a comprehensive multimodal dataset capturing human reactions to robot failures and subsequent explanations in collaborative settings. It aims to facilitate research into human-robot interaction dynamics, addressing the need to study reactions to both initial failures and explanations, as well as the evolution of these reactions in long-term interactions. By providing rich, annotated data on human responses to different types of failures, explanation levels, and explanation varying strategies, the dataset contributes to the development of more robust, adaptive, and satisfying robotic systems capable of maintaining positive relationships with human collaborators, even during challenges like repeated failures

sted, utgiver, år, opplag, sider
IEEE, 2025
Emneord
Human Robot Interaction, Dataset, Robotic Failures, Explainable AI.
HSV kategori
Forskningsprogram
Datalogi
Identifikatorer
urn:nbn:se:kth:diva-360946 (URN)10.5555/3721488.3721616 (DOI)
Konferanse
ACM/IEEE International Conference on Human-Robot Interaction, HRI, Melbourne, Australia, March 4-6, 2025
Merknad

QC 20250310

Tilgjengelig fra: 2025-03-06 Laget: 2025-03-06 Sist oppdatert: 2025-03-10bibliografisk kontrollert
Khanna, P., Naoum, A., Yadollahi, E., Björkman, M. & Smith, C. (2025). REFLEX Dataset: A Multimodal Dataset of Human Reactions to Robot Failures and Explanations. In: HRI 2025 - Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction: . Paper presented at 20th Annual ACM/IEEE International Conference on Human-Robot Interaction, HRI 2025, Melbourne, Australia, March 4-6, 2025 (pp. 1032-1036). Institute of Electrical and Electronics Engineers (IEEE)
Åpne denne publikasjonen i ny fane eller vindu >>REFLEX Dataset: A Multimodal Dataset of Human Reactions to Robot Failures and Explanations
Vise andre…
2025 (engelsk)Inngår i: HRI 2025 - Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction, Institute of Electrical and Electronics Engineers (IEEE) , 2025, s. 1032-1036Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

This work presents REFLEX: Robotic Explanations to FaiLures and Human EXpressions, a comprehensive multimodal dataset capturing human reactions to robot failures and subsequent explanations in collaborative settings. It aims to facilitate research into human-robot interaction dynamics, addressing the need to study reactions to both initial failures and explanations, as well as the evolution of these reactions in long-term interactions. By providing rich, annotated data on human responses to different types of failures, explanation levels, and explanation varying strategies, the dataset contributes to the development of more robust, adaptive, and satisfying robotic systems capable of maintaining positive relationships with human collaborators, even during challenges like repeated failures.

sted, utgiver, år, opplag, sider
Institute of Electrical and Electronics Engineers (IEEE), 2025
Emneord
Dataset, Explainable AI, Human Robot Interaction, Robotic Failures
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-363769 (URN)10.1109/HRI61500.2025.10974185 (DOI)2-s2.0-105004877597 (Scopus ID)
Konferanse
20th Annual ACM/IEEE International Conference on Human-Robot Interaction, HRI 2025, Melbourne, Australia, March 4-6, 2025
Merknad

Part of ISBN 9798350378931

QC 20250526

Tilgjengelig fra: 2025-05-21 Laget: 2025-05-21 Sist oppdatert: 2025-05-26bibliografisk kontrollert
Tarle, M., Larsson, M., Ingeström, G., Nordström, L. & Björkman, M. (2025). Safe Reinforcement Learning to Improve FACTS Setpoint Control in Presence of Model Errors. IEEE transactions on industry applications
Åpne denne publikasjonen i ny fane eller vindu >>Safe Reinforcement Learning to Improve FACTS Setpoint Control in Presence of Model Errors
Vise andre…
2025 (engelsk)Inngår i: IEEE transactions on industry applications, ISSN 0093-9994, E-ISSN 1939-9367Artikkel i tidsskrift (Fagfellevurdert) Epub ahead of print
Abstract [en]

There is limited application of closed-loop control using model-based approaches in wide area monitoring, protection, and control. Challenges that impede model-based approaches include engineering complexity, convergence issues, and model errors. Specifically, considering the rapid growth of distributed generation and renewables in the grid, maintaining an updated model without model errors is challenging. As an alternative to model-based approaches, data-driven control architectures based on reinforcement learning (RL) have shown great promise. In this work, we confront safety concerns with data-driven approaches by studying safe RL to improve voltage and power flow control. For both a model-free RL agent and a model-based RL agent, the accumulated constraint violation is investigated in a case study on the IEEE 14-bus and IEEE 57-bus systems. To evaluate performance, agents are compared against a model-based approach subject to errors. Our findings suggest that RL could be considered for optimizing voltage and current setpoints in systems when topological model errors are present.

sted, utgiver, år, opplag, sider
Institute of Electrical and Electronics Engineers (IEEE), 2025
Emneord
Decision support systems, Flexible AC Transmission Systems (FACTS), power system control, reinforcement learning
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-365876 (URN)10.1109/tia.2025.3569502 (DOI)2-s2.0-105005212319 (Scopus ID)
Forskningsfinansiär
Swedish Foundation for Strategic Research, ID19-0058
Merknad

QC 20250701

Tilgjengelig fra: 2025-07-01 Laget: 2025-07-01 Sist oppdatert: 2025-07-01bibliografisk kontrollert
Zhou, S., Hernandez, A. c., Gomez, C., Yin, W. & Björkman, M. (2025). SmartTBD: Smart Tracking for Resource-constrained Object Detection. ACM Transactions on Embedded Computing Systems, 24(2), Article ID 24.
Åpne denne publikasjonen i ny fane eller vindu >>SmartTBD: Smart Tracking for Resource-constrained Object Detection
Vise andre…
2025 (engelsk)Inngår i: ACM Transactions on Embedded Computing Systems, ISSN 1539-9087, E-ISSN 1558-3465, Vol. 24, nr 2, artikkel-id 24Artikkel i tidsskrift (Fagfellevurdert) Published
Abstract [en]

With the growing demand for video analysis on mobile devices, object tracking has demonstrated to be a suitable assistance to object detection under the Tracking-By-Detection (TBD) paradigm for reducing computational overhead and power demands. However, performing TBD with fixed hyper-parameters leads to computational inefficiency and ignores perceptual dynamics, as fixed setups tend to run suboptimally, given the variability of scenarios. In this article, we propose SmartTBD, a scheduling strategy for TBD based on multi-objective optimization of accuracy-latency metrics. SmartTBD is a novel deep reinforcement learning based scheduling architecture that computes appropriate TBD configurations in video sequences to improve the speed and detection accuracy. This involves a challenging optimization problem due to the intrinsic relation between the video characteristics and the TBD performance. Therefore, we leverage video characteristics, frame information, and the past TBD results to drive the optimization problem. Our approach surpasses baselines with fixed TBD configurations and recent research, achieving accuracy comparable to pure detection while significantly reducing latency. Moreover, it enables performance analysis of tracking and detection in diverse scenarios. The method is proven to be generalizable and highly practical in common video analytics datasets on resource-constrained devices.

sted, utgiver, år, opplag, sider
Association for Computing Machinery (ACM), 2025
Emneord
Mobile vision, tracking-by-detection, scheduling
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-362957 (URN)10.1145/3703912 (DOI)001454951000008 ()2-s2.0-105003605284 (Scopus ID)
Merknad

QC 20250505

Tilgjengelig fra: 2025-05-05 Laget: 2025-05-05 Sist oppdatert: 2025-05-27bibliografisk kontrollert
Taleb, F., Vasco, M., Ribeiro, A. H., Björkman, M. & Kragic Jensfelt, D. (2024). Can Transformers Smell Like Humans?. In: Advances in Neural Information Processing Systems 37 - 38th Conference on Neural Information Processing Systems, NeurIPS 2024: . Paper presented at 38th Conference on Neural Information Processing Systems, NeurIPS 2024, Vancouver, Canada, Dec 9 2024 - Dec 15 2024. Neural Information Processing Systems Foundation
Åpne denne publikasjonen i ny fane eller vindu >>Can Transformers Smell Like Humans?
Vise andre…
2024 (engelsk)Inngår i: Advances in Neural Information Processing Systems 37 - 38th Conference on Neural Information Processing Systems, NeurIPS 2024, Neural Information Processing Systems Foundation , 2024Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

The human brain encodes stimuli from the environment into representations that form a sensory perception of the world. Despite recent advances in understanding visual and auditory perception, olfactory perception remains an under-explored topic in the machine learning community due to the lack of large-scale datasets annotated with labels of human olfactory perception. In this work, we ask the question of whether pre-trained transformer models of chemical structures encode representations that are aligned with human olfactory perception, i.e., can transformers smell like humans? We demonstrate that representations encoded from transformers pre-trained on general chemical structures are highly aligned with human olfactory perception. We use multiple datasets and different types of perceptual representations to show that the representations encoded by transformer models are able to predict: (i) labels associated with odorants provided by experts; (ii) continuous ratings provided by human participants with respect to pre-defined descriptors; and (iii) similarity ratings between odorants provided by human participants. Finally, we evaluate the extent to which this alignment is associated with physicochemical features of odorants known to be relevant for olfactory decoding.

sted, utgiver, år, opplag, sider
Neural Information Processing Systems Foundation, 2024
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-361995 (URN)2-s2.0-105000466521 (Scopus ID)
Konferanse
38th Conference on Neural Information Processing Systems, NeurIPS 2024, Vancouver, Canada, Dec 9 2024 - Dec 15 2024
Merknad

QC 20250408

Tilgjengelig fra: 2025-04-03 Laget: 2025-04-03 Sist oppdatert: 2025-04-08bibliografisk kontrollert
Taleb, F., Vasco, M., Rajabi, N., Björkman, M. & Kragic, D. (2024). Challenging Deep Learning Methods for EEG Signal Denoising under Data Corruption. In: 46th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2024 - Proceedings: . Paper presented at 46th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2024, Orlando, United States of America, Jul 15 2024 - Jul 19 2024. Institute of Electrical and Electronics Engineers (IEEE)
Åpne denne publikasjonen i ny fane eller vindu >>Challenging Deep Learning Methods for EEG Signal Denoising under Data Corruption
Vise andre…
2024 (engelsk)Inngår i: 46th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2024 - Proceedings, Institute of Electrical and Electronics Engineers (IEEE) , 2024Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

Capturing informative electroencephalogram (EEG) signals is a challenging task due to the presence of noise (e.g., due to human movement). In extreme cases, data recordings from specific electrodes (channels) can become corrupted and entirely devoid of information. Motivated by recent work on deep-learning-based approaches for EEG signal denoising, we present the first benchmark study on the performance of EEG signal denoising methods in the presence of corrupted channels. We design our study considering a wide variety of datasets, models, and evaluation tasks. Our results highlight the need for assessing the performance of EEG deep-learning models across a broad suite of datasets, as provided by our benchmark.

sted, utgiver, år, opplag, sider
Institute of Electrical and Electronics Engineers (IEEE), 2024
Emneord
data corruption, deep learning, EEG, signal denoising, signal noise
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-358866 (URN)10.1109/EMBC53108.2024.10782132 (DOI)40039138 (PubMedID)2-s2.0-85214969123 (Scopus ID)
Konferanse
46th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2024, Orlando, United States of America, Jul 15 2024 - Jul 19 2024
Merknad

Part of ISBN 9798350371499]

QC 20250128

Tilgjengelig fra: 2025-01-23 Laget: 2025-01-23 Sist oppdatert: 2025-05-27bibliografisk kontrollert
Longhini, A., Büsching, M., Duisterhof, B. P., Lundell, J., Ichnowski, J., Björkman, M. & Kragic, D. (2024). Cloth-Splatting: 3D Cloth State Estimation from RGB Supervision. In: Proceedings of the 8th Conference on Robot Learning, CoRL 2024: . Paper presented at 8th Annual Conference on Robot Learning, November 6-9, 2024, Munich, Germany (pp. 2845-2865). ML Research Press
Åpne denne publikasjonen i ny fane eller vindu >>Cloth-Splatting: 3D Cloth State Estimation from RGB Supervision
Vise andre…
2024 (engelsk)Inngår i: Proceedings of the 8th Conference on Robot Learning, CoRL 2024, ML Research Press , 2024, s. 2845-2865Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

We introduce Cloth-Splatting, a method for estimating 3D states of cloth from RGB images through a prediction-update framework. Cloth-Splatting leverages an action-conditioned dynamics model for predicting future states and uses 3D Gaussian Splatting to update the predicted states. Our key insight is that coupling a 3D mesh-based representation with Gaussian Splatting allows us to define a differentiable map between the cloth's state space and the image space. This enables the use of gradient-based optimization techniques to refine inaccurate state estimates using only RGB supervision. Our experiments demonstrate that Cloth-Splatting not only improves state estimation accuracy over current baselines but also reduces convergence time by ∼85 %.

sted, utgiver, år, opplag, sider
ML Research Press, 2024
Emneord
3D State Estimation, Gaussian Splatting, Vision-based Tracking, Deformable Objects
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-357192 (URN)2-s2.0-86000735293 (Scopus ID)
Konferanse
8th Annual Conference on Robot Learning, November 6-9, 2024, Munich, Germany
Merknad

QC 20250328

Tilgjengelig fra: 2024-12-04 Laget: 2024-12-04 Sist oppdatert: 2025-03-28bibliografisk kontrollert
Organisasjoner
Identifikatorer
ORCID-id: ORCID iD iconorcid.org/0000-0003-0579-3372