Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Estimating Uncertainty in Task Oriented Dialogue
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Speech, Music and Hearing, TMH.ORCID iD: 0000-0002-8874-6629
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Speech, Music and Hearing, TMH.
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Speech, Music and Hearing, TMH.ORCID iD: 0000-0002-0397-6442
2019 (English)In: ICMI 2019 - Proceedings of the 2019 International Conference on Multimodal Interaction / [ed] Wen Gao, Helen Mei Ling Meng, Matthew Turk, Susan R. Fussell, ACM Digital Library, 2019, p. 414-418Conference paper, Published paper (Refereed)
Abstract [en]

Situated multimodal systems that instruct humans need to handle user uncertainties, as expressed in behaviour, and plan their actions accordingly. Speakers’ decision to reformulate or repair previous utterances depends greatly on the listeners’ signals of uncertainty. In this paper, we estimate uncertainty in a situated guided task, as leveraged in non-verbal cues expressed by the listener, and predict that the speaker will reformulate their utterance. We use a corpus where people instruct how to assemble furniture, and extract their multimodal features. While uncertainty is in cases ver- bally expressed, most instances are expressed non-verbally, which indicates the importance of multimodal approaches. In this work, we present a model for uncertainty estimation. Our findings indicate that uncertainty estimation from non- verbal cues works well, and can exceed human annotator performance when verbal features cannot be perceived.

Place, publisher, year, edition, pages
ACM Digital Library, 2019. p. 414-418
Keywords [en]
situated interaction, dialogue and discourse, grounding
National Category
Human Computer Interaction
Identifiers
URN: urn:nbn:se:kth:diva-261628DOI: 10.1145/3340555.3353722Scopus ID: 2-s2.0-85074940956ISBN: 9781450368605 (print)OAI: oai:DiVA.org:kth-261628DiVA, id: diva2:1359302
Conference
21st ACM International Conference on Multimodal Interaction, Suzhou, Jiangsu, China. October 14-18, 2019
Note

QC 20191209. QC 20200214

Available from: 2019-10-08 Created: 2019-10-08 Last updated: 2020-02-14Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopusConference websiteConference proceedings

Authority records BETA

Kontogiorgos, DimosthenisAbelho Pereira, André TiagoGustafson, Joakim

Search in DiVA

By author/editor
Kontogiorgos, DimosthenisAbelho Pereira, André TiagoGustafson, Joakim
By organisation
Speech, Music and Hearing, TMH
Human Computer Interaction

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 167 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf