kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
A Comparison of Visualisation Methods for Disambiguating Verbal Requests in Human-Robot Interaction
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0002-8874-6629
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0001-6671-9366
Show others and affiliations
2018 (English)In: Proceedings 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN) 2018, IEEE, 2018Conference paper, Published paper (Refereed)
Abstract [en]

Picking up objects requested by a human user is a common task in human-robot interaction. When multiple objects match the user's verbal description, the robot needs to clarify which object the user is referring to before executing the action. Previous research has focused on perceiving user's multimodal behaviour to complement verbal commands or minimising the number of follow up questions to reduce task time. In this paper, we propose a system for reference disambiguation based on visualisation and compare three methods to disambiguate natural language instructions. In a controlled experiment with a YuMi robot, we investigated realtime augmentations of the workspace in three conditions - head-mounted display, projector, and a monitor as the baseline - using objective measures such as time and accuracy, and subjective measures like engagement, immersion, and display interference. Significant differences were found in accuracy and engagement between the conditions, but no differences were found in task time. Despite the higher error rates in the head-mounted display condition, participants found that modality more engaging than the other two, but overall showed preference for the projector condition over the monitor and head-mounted display conditions.

Place, publisher, year, edition, pages
IEEE, 2018.
National Category
Human Computer Interaction
Identifiers
URN: urn:nbn:se:kth:diva-235548DOI: 10.1109/ROMAN.2018.8525554ISI: 000494315600008Scopus ID: 2-s2.0-85050581339ISBN: 978-1-5386-7981-4 (print)OAI: oai:DiVA.org:kth-235548DiVA, id: diva2:1252067
Conference
27th IEEE International Symposium on Robot and Human Interactive Communication (IEEE RO-MAN), Nanjing, PEOPLES R CHINA, AUG 27-31, 2018
Note

QC 20181207. QC 20191219. QC 20200108

Available from: 2018-09-29 Created: 2018-09-29 Last updated: 2024-03-18Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopusConference proceedings

Authority records

Sibirtseva, ElenaKontogiorgos, DimosthenisNykvist, OlovKaraoǧuz, HakanLeite, IolandaGustafson, JoakimKragic, Danica

Search in DiVA

By author/editor
Sibirtseva, ElenaKontogiorgos, DimosthenisNykvist, OlovKaraoǧuz, HakanLeite, IolandaGustafson, JoakimKragic, Danica
By organisation
Robotics, Perception and Learning, RPLSpeech, Music and Hearing, TMH
Human Computer Interaction

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 255 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf