Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Multimodal language grounding for improved human-robot collaboration: Exploring Spatial Semantic Representations in the Shared Space of Attention
KTH, School of Computer Science and Communication (CSC), Speech, Music and Hearing, TMH. (TMH)ORCID iD: 0000-0002-8874-6629
2017 (English)In: ICMI 2017 - Proceedings of the 19th ACM International Conference on Multimodal Interaction, ACM Digital Library, 2017, Vol. 2017, p. 660-664Conference paper, Published paper (Refereed)
Abstract [en]

There is an increased interest in arti cially intelligent technology that surrounds us and takes decisions on our behalf. This creates the need for such technology to be able to communicate with humans and understand natural language and non-verbal behaviour that may carry information about our complex physical world. Arti cial agents today still have little knowledge about the physical space that surrounds us and about the objects or concepts within our attention. We are still lacking computational methods in understanding the context of human conversation that involves objects and locations around us. Can we use multimodal cues from human perception of the real world as an example of language learning for robots? Can arti cial agents and robots learn about the physical world by observing how humans interact with it and how they refer to it and attend during their conversations? This PhD project’s focus is on combining spoken language and non-verbal behaviour extracted by multi-party dialogue in order to increase context awareness and spatial understanding for arti cial agents.

Place, publisher, year, edition, pages
ACM Digital Library, 2017. Vol. 2017, p. 660-664
Keywords [en]
human-computer interaction, human-robot collaboration, multi- modal interaction, natural language interfaces, spatial reasoning, human perception, multisensory processing
National Category
Engineering and Technology
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:kth:diva-215500DOI: 10.1145/3136755.3137038Scopus ID: 2-s2.0-85046626210ISBN: 978-1-4503-5543-8 (electronic)OAI: oai:DiVA.org:kth-215500DiVA, id: diva2:1148083
Conference
19th ACM International Conference on Multimodal Interaction, ICMI 2017, Glasgow, United Kingdom, 13 November 2017 through 17 November 2017
Projects
FACT
Funder
Swedish Research Council
Note

QC 20180528

Available from: 2017-10-09 Created: 2017-10-09 Last updated: 2018-05-28Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopushttps://dl.acm.org/citation.cfm?doid=3136755.3137038

Search in DiVA

By author/editor
Kontogiorgos, Dimosthenis
By organisation
Speech, Music and Hearing, TMH
Engineering and Technology

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 154 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf