Endre søk
RefereraExporteraLink to record
Permanent link

Direct link
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annet format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annet språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Automatic Frustration Detection Using Thermal Imaging
KTH, Skolan för elektroteknik och datavetenskap (EECS), Intelligenta system, Robotik, perception och lärande, RPL.ORCID-id: 0000-0001-5660-5330
Univ Genoa, Genoa, Italy..
KTH, Skolan för elektroteknik och datavetenskap (EECS), Intelligenta system, Robotik, perception och lärande, RPL.
PAL Robot, Barcelona, Spain..
Vise andre og tillknytning
2022 (engelsk)Inngår i: PROCEEDINGS OF THE 2022 17TH ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION (HRI '22), Institute of Electrical and Electronics Engineers (IEEE) , 2022, s. 451-460Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

To achieve seamless interactions, robots have to be capable of reliably detecting affective states in real time. One of the possible states that humans go through while interacting with robots is frustration. Detecting frustration from RGB images can be challenging in some real-world situations; thus, we investigate in this work whether thermal imaging can be used to create a model that is capable of detecting frustration induced by cognitive load and failure. To train our model, we collected a data set from 18 participants experiencing both types of frustration induced by a robot. The model was tested using features from several modalities: thermal, RGB, Electrodermal Activity (EDA), and all three combined. When data from both frustration cases were combined and used as training input, the model reached an accuracy of 89% with just RGB features, 87% using only thermal features, 84% using EDA, and 86% when using all modalities. Furthermore, the highest accuracy for the thermal data was reached using three facial regions of interest: nose, forehead and lower lip.

sted, utgiver, år, opplag, sider
Institute of Electrical and Electronics Engineers (IEEE) , 2022. s. 451-460
Serie
ACM IEEE International Conference on Human-Robot Interaction, ISSN 2167-2121
Emneord [en]
Human-robot interaction, Thermal imaging, Frustration, cognitive load, Action units
HSV kategori
Identifikatorer
URN: urn:nbn:se:kth:diva-322478DOI: 10.1109/HRI53351.2022.9889545ISI: 000869793600050Scopus ID: 2-s2.0-85140750883OAI: oai:DiVA.org:kth-322478DiVA, id: diva2:1719935
Konferanse
17th Annual ACM/IEEE International Conference on Human-Robot Interaction (HRI), MAR 07-10, 2022, ELECTR NETWORK
Merknad

Part of proceedings: ISBN 978-1-6654-0731-1

QC 20221216

Tilgjengelig fra: 2022-12-16 Laget: 2022-12-16 Sist oppdatert: 2025-08-25bibliografisk kontrollert
Inngår i avhandling
1. Multi-Modal Affective State Detection For Dyadic Interactions Using Thermal Imaging and Context
Åpne denne publikasjonen i ny fane eller vindu >>Multi-Modal Affective State Detection For Dyadic Interactions Using Thermal Imaging and Context
2025 (engelsk)Doktoravhandling, med artikler (Annet vitenskapelig)
Abstract [en]

Until recently, most robotic systems have operated with limited emotional intelligence, primarily responding to pre-programmed cues rather than adapting to human emotional states. Thus, affect recognition in humanrobot-interaction remains a significant challenge and is twofold: robots must not only detect emotional expressions but they also need to interpret them within their social context, requiring systems that are capable of collecting information from its surrounding, analyzing it and thereafter generalizing across different interaction scenarios and cultural contexts to handle more complex situations.

This thesis tackles affect recognition using multi-modal approaches that combine thermal imaging, facial expression analysis, and contextual understanding. Thermal imaging offers unique insights into physiological responses associated with emotional states, complementing traditional vision-based approaches while maintaining non-contact operation. The integration of thermal imaging, facial expression analysis, and contextual understanding creates a comprehensive multi-modal framework that addresses the key challenges in affect recognition, such as varying lighting conditions, occlusions, and ambiguous emotional expressions. This combination provides complementary information streams that enhance robustness in real-world environments, making it an effective case study for developing context-aware emotional intelligence in robotics.

We introduce a novel context-aware transformer architecture that processes multiple data streams while maintaining temporal relationships and contextual understanding. Each modality contributes complementary information about the user’s emotional state, while the context processing ensures situation-appropriate interpretation. For instance, distinguishing between a smile indicating enjoyment during collaborative tasks versus one masking nervousness in stressful situations. This contextual awareness is crucial for appropriate robot responses in real-world deployments.

The research contributions span four areas: (1) developing robust thermal feature extraction techniques that capture subtle emotional responses (2) creating a transformer-based architecture for multi-modal fusion that effectively incorporates situational information, (3) implementing real-time processing pipelines that enable practical deployment in human-robot interaction scenarios, and (4) validating these approaches through extensive real-world interaction studies. Results show improved recognition accuracy from 77% using traditional approaches to 89% with our context-aware multi-modal system, demonstrating the ability to understand and appropriately respond to human emotions in dynamic social situations.

sted, utgiver, år, opplag, sider
Stockholm: KTH Royal Institute of Technology, 2025. s. x, 56
Serie
TRITA-EECS-AVL ; 2025:74
HSV kategori
Identifikatorer
urn:nbn:se:kth:diva-368995 (URN)9789181063431 (ISBN)
Disputas
2025-09-26, D37, Lindstedtsvägen 9, Stockholm, 13:00 (engelsk)
Opponent
Veileder
Merknad

QC 20250905

Tilgjengelig fra: 2025-09-05 Laget: 2025-08-25 Sist oppdatert: 2025-09-29bibliografisk kontrollert

Open Access i DiVA

Fulltekst mangler i DiVA

Andre lenker

Forlagets fulltekstScopus

Person

Mohamed, YoussefParreira, Maria TeresaLeite, Iolanda

Søk i DiVA

Av forfatter/redaktør
Mohamed, YoussefParreira, Maria TeresaLeite, Iolanda
Av organisasjonen

Søk utenfor DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric

doi
urn-nbn
Totalt: 206 treff
RefereraExporteraLink to record
Permanent link

Direct link
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annet format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annet språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf