Socially Aware Many-to-Machine Communication
2012 (English)Conference paper (Other academic)
This reports describes the output of the project P5: Socially Aware Many-to-Machine Communication (M2M) at the eNTERFACE’12 workshop. In this project, we designed and implemented a new front-end for handling multi-user interaction in a dialog system. We exploit the Microsoft Kinect device for capturing multimodal input and extract some features describing user and face positions. These data are then analyzed in real-time to robustly detect speech and determine both who is speaking and whether the speech is directed to the system or not. This new front-end is integrated to the SEMAINE (Sustained Emotionally colored Machine-human Interaction using Nonverbal Expression) system. Furthermore, a multimodal corpus has been created, capturing all of the system inputs in two different scenarios involving human-human and human-computer interaction.
Place, publisher, year, edition, pages
IdentifiersURN: urn:nbn:se:kth:diva-165818OAI: oai:DiVA.org:kth-165818DiVA: diva2:808849
8th International Summer Workshop on Multimodal Interfaces, Metz, France
QC 201610172015-04-292015-04-292016-10-17Bibliographically approved