Ändra sökning
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Multimodal language grounding for improved human-robot collaboration: Exploring Spatial Semantic Representations in the Shared Space of Attention
KTH, Skolan för datavetenskap och kommunikation (CSC), Tal, musik och hörsel, TMH. (TMH)ORCID-id: 0000-0002-8874-6629
2017 (Engelska)Ingår i: ICMI 2017 - Proceedings of the 19th ACM International Conference on Multimodal Interaction, ACM Digital Library, 2017, Vol. 2017, s. 660-664Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

There is an increased interest in arti cially intelligent technology that surrounds us and takes decisions on our behalf. This creates the need for such technology to be able to communicate with humans and understand natural language and non-verbal behaviour that may carry information about our complex physical world. Arti cial agents today still have little knowledge about the physical space that surrounds us and about the objects or concepts within our attention. We are still lacking computational methods in understanding the context of human conversation that involves objects and locations around us. Can we use multimodal cues from human perception of the real world as an example of language learning for robots? Can arti cial agents and robots learn about the physical world by observing how humans interact with it and how they refer to it and attend during their conversations? This PhD project’s focus is on combining spoken language and non-verbal behaviour extracted by multi-party dialogue in order to increase context awareness and spatial understanding for arti cial agents.

Ort, förlag, år, upplaga, sidor
ACM Digital Library, 2017. Vol. 2017, s. 660-664
Nyckelord [en]
human-computer interaction, human-robot collaboration, multi- modal interaction, natural language interfaces, spatial reasoning, human perception, multisensory processing
Nationell ämneskategori
Teknik och teknologier
Forskningsämne
Datalogi
Identifikatorer
URN: urn:nbn:se:kth:diva-215500DOI: 10.1145/3136755.3137038Scopus ID: 2-s2.0-85046626210ISBN: 978-1-4503-5543-8 (digital)OAI: oai:DiVA.org:kth-215500DiVA, id: diva2:1148083
Konferens
19th ACM International Conference on Multimodal Interaction, ICMI 2017, Glasgow, United Kingdom, 13 November 2017 through 17 November 2017
Projekt
FACT
Forskningsfinansiär
Vetenskapsrådet
Anmärkning

QC 20180528

Tillgänglig från: 2017-10-09 Skapad: 2017-10-09 Senast uppdaterad: 2018-05-28Bibliografiskt granskad

Open Access i DiVA

Fulltext saknas i DiVA

Övriga länkar

Förlagets fulltextScopushttps://dl.acm.org/citation.cfm?doid=3136755.3137038

Sök vidare i DiVA

Av författaren/redaktören
Kontogiorgos, Dimosthenis
Av organisationen
Tal, musik och hörsel, TMH
Teknik och teknologier

Sök vidare utanför DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetricpoäng

doi
isbn
urn-nbn
Totalt: 337 träffar
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf