kth.sePublications
Change search
Link to record
Permanent link

Direct link
Alternative names
Publications (10 of 10) Show all publications
Axelsson, A., Vaddadi, B., Bogdan, C. M. & Skantze, G. (2024). Robots in autonomous buses: Who hosts when no human is there?. In: HRI 2024 Companion - Companion of the 2024 ACM/IEEE International Conference on Human-Robot Interaction: . Paper presented at 19th Annual ACM/IEEE International Conference on Human-Robot Interaction, HRI 2024, Boulder, United States of America, Mar 11 2024 - Mar 15 2024 (pp. 1278-1280). Association for Computing Machinery (ACM)
Open this publication in new window or tab >>Robots in autonomous buses: Who hosts when no human is there?
2024 (English)In: HRI 2024 Companion - Companion of the 2024 ACM/IEEE International Conference on Human-Robot Interaction, Association for Computing Machinery (ACM) , 2024, p. 1278-1280Conference paper, Published paper (Refereed)
Abstract [en]

In mid-2023, we performed an experiment in autonomous buses in Stockholm, Sweden, to evaluate the role that social robots might have in such settings, and their effects on passengers' feeling of safety and security, given the absence of human drivers or clerks. To address the situations that may occur in autonomous public transit (APT), we compared an embodied agent to a disembodied agent. In this video publication, we showcase some of the things that worked with the interactions we created, and some problematic issues that we had not anticipated.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2024
Keywords
APT, assistant, autonomous, bus, clerk, guide, passenger, public transit, public transport, robot, self-driving, shuttle, wizard
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:kth:diva-344811 (URN)10.1145/3610978.3641115 (DOI)001255070800272 ()2-s2.0-85188117955 (Scopus ID)
Conference
19th Annual ACM/IEEE International Conference on Human-Robot Interaction, HRI 2024, Boulder, United States of America, Mar 11 2024 - Mar 15 2024
Projects
tmh_robohost
Note

QC 20240402

 Part of ISBN 9798400703232

Available from: 2024-03-28 Created: 2024-03-28 Last updated: 2024-10-24Bibliographically approved
Axelsson, A. (2023). Adaptive Robot Presenters: Modelling Grounding in Multimodal Interaction. (Doctoral dissertation). Stockholm: KTH Royal Institute of Technology
Open this publication in new window or tab >>Adaptive Robot Presenters: Modelling Grounding in Multimodal Interaction
2023 (English)Doctoral thesis, monograph (Other academic)
Abstract [en]

This thesis addresses the topic of grounding in human-robot interaction, that is, the process by which the human and robot can ensure mutual understanding. To explore this topic, the scenario of a robot holding a presentation to a human audience is used, where the robot has to process multimodal feedback from the human in order to adapt the presentation to the human's level of understanding.

First, the use of behaviour trees to model real-time interactive processes of the presentation is addressed. A system based on the behaviour tree architecture is used in a semi-automated Wizard-of-oz experiment, showing that audience members prefer an adaptive system to a non-adaptive alternative.

Next, the thesis addresses the use of knowledge graphs to represent the content of the presentation given by the robot. By building a small, local knowledge graph containing properties (edges) that represent facts about the presentation, the system can iterate over that graph and consistently find ways to refer to entities by referring to previously grounded content. A system based on this architecture is implemented, and an evaluation using simulated users is presented. The results show that crowdworkers comparing different adaptation strategies are sensitive to the types of adaptation enabled by the knowledge graph approach.

In a face-to-face presentation setting, feedback from the audience can potentially be expressed through various modalities, including speech, head movements, gaze, facial gestures and body pose. The thesis explores how such feedback can be automatically classified. A corpus of human-robot interactions is annotated, and models are trained to classify human feedback as positive, negative or neutral. A relatively high accuracy is achieved by training simple classifiers with signals found mainly in the speech and head movements.

When knowledge graphs are used as the underlying representation of the system's presentation, some consistent way of generating text, that can be turned into speech, is required. This graph-to-text problem is explored by proposing several methods, both template-based and methods based on zero-shot generation using large language models (LLMs). A novel evaluation method using a combination of factual, counter-factual and fictional graphs is proposed. 

Finally, the thesis presents and evaluates a fully automated system using all of the components above. The results show that audience members prefer the adaptive system to a non-adaptive system, matching the results from the beginning of the thesis. However, we note that clear learning results are not found, which means that the entertainment aspects of the presentation are perhaps more prominent than the learning aspects.

Abstract [sv]

Denna avhandling behandlar ämnet multimodal kommunikativ grundning (grounding) mellan robotar och människor. Detta är processen för hur en människa och en robot kan säkerställa att de har en gemensam förståelse. För att utforska detta ämne ämne, används ett scenario där en robot håller en presentation för en mänsklig publik. Roboten måste analysera multimodala signaler från människan för att anpassa presentationen till människans nivå av förståelse.

Först undersöks hur beteendeträd kan användas för att modellera realtidsaspekterna av interaktionen mellan robotpresentatören och dess publik. Ett system som baseras på beteendeträdsarkitekturen används i ett delvis automatiskt, delvis människostyrt experiment, där det visas att publikmedlemmar i labbmiljö föredrar ett system som anpassar presentationen till deras reaktioner över ett som inte anpassar sin presentation.

Efter detta, urdersöker också avhandlingen hur kunskapsgrafer kan användas för att representera innehållet som roboten presenterar. Om en liten, lokal kunskapsgraf byggs så att den innehåller relationer (kanter) som representerar fakta i presentationen, så kan roboten iterera över grafen och konsekvent hitta refererande uttryck som använder sig av kunskap som publiken redan har. Ett system som baseras på denna arkitektur implementeras, och ett experiment med simulerade interaktioner utförs och presenteras. Experimentets resultat visar att utvärderare som jämför olika anpassningsstrategier föredrar ett system som kan utföra den sortens anpassning som grafmetoden tillåter. 

Publikens reaktioner i ett presentationsscenario kan ske genom olika modaliteter, som tal, huvudrörelser, blickriktning, ansiktsuttryck och kroppsspråk. För att klassificera kommunikativ återmatning (feedback) av dessa modaliteter från presentationspubliken, utforskas hur sådana signaler kan analyseras automatiskt. En datamängd med interaktioner mellan en människa och vår robot annoteras, och statistiska modeller tränas för att klassificera mänskliga återmatningssignaler från flera olika modaliteter som positiva, negativa eller neutrala. En jämförelsevis hög klassifikationsprecision uppnås genom att träna enklare klassifikationsmodeller på relativt få klasser av signaler i tal- och huvudrörelsemodaliteterna. Detta antyder att museiscenariot med en robotpresentatör inte uppmuntrar publiken att använda komplicerade, mångtydiga kommunikativa beteenden.

När kunskapsgrafer används som presentationssystemets informationsrepresentation, behövs det konsekventa metoder för att generera text som kan omvandlas till tal, från grafdata. Graf-till-text-problemet utforskas genom att föreslå flera olika metoder, både enklare mall-baserade sådana och mer avancerade metoder baserade på stora språkmodeller (LLM:er). Genom att föreslå en ny utvärderingsmetod där sanna, fiktiva och falska grafer genereras, visar vi också att sanningshalten i vad som uttrycks påverkar kvaliteten i texten som LLM-metoderna ger från kunskapsgrafdata.

Avhandlingen använder sig slutligen av alla de ovanstående föreslagna komponenterna i ett och samma helautomatiska presentationssystem. Resultaten visar att publikmedlemmar föredrar ett system som anpassar sin presentation över ett som inte anpassar sin presentation, vilket speglar resultaten från början av avhandlingen. Vi ser också att tydliga inlärningsresultat uteblir i detta experiment, vilket kanske kan tolkas som att publikmedlemmarna i museiscenariot snarare letar efter en underhållare än efter en lärare som presentatör.

Place, publisher, year, edition, pages
Stockholm: KTH Royal Institute of Technology, 2023. p. 205
Series
TRITA-EECS-AVL ; 2023:70
Keywords
Human-robot interaction, Dialogue, Presentation, Museum, Grounding, Multimodal, Feedback, Classification, Knowledge graphs, KG, KG-to-text, WebNLG, System, Learning, Large Language Model, LLM, människa-robot-interaktion, hri, dialog, presentation, museum, grundning, multimodal, multimodalitet, återmatning, klassifikation, kunskapsgraf, kg, kg-till-text, data-tilltext, webnlg, system, inlärning, lärande. stor språkmodell, llm
National Category
Natural Language Processing Robotics and automation Computer and Information Sciences
Research subject
Speech and Music Communication
Identifiers
urn:nbn:se:kth:diva-338178 (URN)978-91-8040-728-1 (ISBN)
Public defence
2023-11-10, https://kth-se.zoom.us/j/62979383325?pwd=VnJ1a1N6azZpaGxvZVZmVkU1NE5ZUT09, F3, Lindstedtsvägen 26, Stockholm, 14:00 (English)
Opponent
Supervisors
Funder
Swedish Energy Agency, P2020-90133Swedish Foundation for Strategic Research, RIT15-0133
Note

QC 20231017

Available from: 2023-10-17 Created: 2023-10-16 Last updated: 2025-02-05Bibliographically approved
Axelsson, A. & Skantze, G. (2023). Do you follow?: A fully automated system for adaptive robot presenters. In: HRI 2023: Proceedings of the 2023 ACM/IEEE International Conference on Human-Robot Interaction. Paper presented at 18th Annual ACM/IEEE International Conference on Human-Robot Interaction, HRI 2023, Stockholm, Sweden, Mar 13 2023 - Mar 16 2023 (pp. 102-111). Association for Computing Machinery (ACM)
Open this publication in new window or tab >>Do you follow?: A fully automated system for adaptive robot presenters
2023 (English)In: HRI 2023: Proceedings of the 2023 ACM/IEEE International Conference on Human-Robot Interaction, Association for Computing Machinery (ACM) , 2023, p. 102-111Conference paper, Published paper (Refereed)
Abstract [en]

An interesting application for social robots is to act as a presenter, for example as a museum guide. In this paper, we present a fully automated system architecture for building adaptive presentations for embodied agents. The presentation is generated from a knowledge graph, which is also used to track the grounding state of information, based on multimodal feedback from the user. We introduce a novel way to use large-scale language models (GPT-3 in our case) to lexicalise arbitrary knowledge graph triples, greatly simplifying the design of this aspect of the system. We also present an evaluation where 43 participants interacted with the system. The results show that users prefer the adaptive system and consider it more human-like and flexible than a static version of the same system, but only partial results are seen in their learning of the facts presented by the robot.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2023
Keywords
adaptation, behaviour tree, feedback, knowledge graph, learning, lexicalisation, multimodal
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-333378 (URN)10.1145/3568162.3576958 (DOI)2-s2.0-85150369153 (Scopus ID)
Conference
18th Annual ACM/IEEE International Conference on Human-Robot Interaction, HRI 2023, Stockholm, Sweden, Mar 13 2023 - Mar 16 2023
Projects
tmh_feedback
Note

Part of ISBN 9781450399647

QC 20230801

Available from: 2023-08-01 Created: 2023-08-01 Last updated: 2024-10-24Bibliographically approved
Cumbal, R., Axelsson, A., Mehta, S. & Engwall, O. (2023). Stereotypical nationality representations in HRI: perspectives from international young adults. Frontiers in Robotics and AI, 10, Article ID 1264614.
Open this publication in new window or tab >>Stereotypical nationality representations in HRI: perspectives from international young adults
2023 (English)In: Frontiers in Robotics and AI, E-ISSN 2296-9144, Vol. 10, article id 1264614Article in journal (Refereed) Published
Abstract [en]

People often form immediate expectations about other people, or groups of people, based on visual appearance and characteristics of their voice and speech. These stereotypes, often inaccurate or overgeneralized, may translate to robots that carry human-like qualities. This study aims to explore if nationality-based preconceptions regarding appearance and accents can be found in people's perception of a virtual and a physical social robot. In an online survey with 80 subjects evaluating different first-language-influenced accents of English and nationality-influenced human-like faces for a virtual robot, we find that accents, in particular, lead to preconceptions on perceived competence and likeability that correspond to previous findings in social science research. In a physical interaction study with 74 participants, we then studied if the perception of competence and likeability is similar after interacting with a robot portraying one of four different nationality representations from the online survey. We find that preconceptions on national stereotypes that appeared in the online survey vanish or are overshadowed by factors related to general interaction quality. We do, however, find some effects of the robot's stereotypical alignment with the subject group, with Swedish subjects (the majority group in this study) rating the Swedish-accented robot as less competent than the international group, but, on the other hand, recalling more facts from the Swedish robot's presentation than the international group does. In an extension in which the physical robot was replaced by a virtual robot interacting in the same scenario online, we further found the same results that preconceptions are of less importance after actual interactions, hence demonstrating that the differences in the ratings of the robot between the online survey and the interaction is not due to the interaction medium. We hence conclude that attitudes towards stereotypical national representations in HRI have a weak effect, at least for the user group included in this study (primarily educated young students in an international setting).

Place, publisher, year, edition, pages
Frontiers Media SA, 2023
Keywords
accent, appearance, social robot, nationality, stereotype, impression, competence, likeability
National Category
Human Computer Interaction
Identifiers
urn:nbn:se:kth:diva-341526 (URN)10.3389/frobt.2023.1264614 (DOI)001115613500001 ()38077460 (PubMedID)2-s2.0-85178920101 (Scopus ID)
Note

QC 20231222

Available from: 2023-12-22 Created: 2023-12-22 Last updated: 2024-02-26Bibliographically approved
Axelsson, A. & Skantze, G. (2023). Using Large Language Models for Zero-Shot Natural Language Generation from Knowledge Graphs. In: Proceedings of the Workshop on Multimodal, Multilingual Natural Language Generation and Multilingual WebNLG Challenge (MM-NLG 2023): . Paper presented at 2023 Workshop on Multimodal, Multilingual Natural Language Generation and Multilingual WebNLG Challenge, MM-NLG 2023, Prague, Czechia, Sep 12 2023 (pp. 39-54). Association for Computational Linguistics (ACL)
Open this publication in new window or tab >>Using Large Language Models for Zero-Shot Natural Language Generation from Knowledge Graphs
2023 (English)In: Proceedings of the Workshop on Multimodal, Multilingual Natural Language Generation and Multilingual WebNLG Challenge (MM-NLG 2023), Association for Computational Linguistics (ACL) , 2023, p. 39-54Conference paper, Published paper (Refereed)
Abstract [en]

In any system that uses structured knowledgegraph (KG) data as its underlying knowledge representation, KG-to-text generation is a useful tool for turning parts of the graph data into text that can be understood by humans. Recent work has shown that models that make use of pretraining on large amounts of text data can perform well on the KG-to-text task, even with relatively little training data on the specific graph-to-text task. In this paper, we build on this concept by using large language models to perform zero-shot generation based on nothing but the model’s understanding of the triple structure from what it can read. We show that ChatGPT achieves near state-of-the-art performance on some measures of the WebNLG 2020 challenge, but falls behind on others. Additionally, we compare factual, counter-factual and fictional statements, and show that there is a significant connection between what the LLM already knows about the data it is parsing and the quality of the output text.

Place, publisher, year, edition, pages
Association for Computational Linguistics (ACL), 2023
Keywords
large language model, llm, lexicalisation, kg-to-text, data-to-text, bias, hallucination, triples, triple, knowledge graph, KG, webnlg, wikidata, språkmodell, llm, lexikalisering, data till text, kg till text, hallucination, triplett, kunskapsgraf, KG, webnlg, wikidata
National Category
Natural Language Processing
Research subject
Speech and Music Communication
Identifiers
urn:nbn:se:kth:diva-338176 (URN)2-s2.0-85175688234 (Scopus ID)
Conference
2023 Workshop on Multimodal, Multilingual Natural Language Generation and Multilingual WebNLG Challenge, MM-NLG 2023, Prague, Czechia, Sep 12 2023
Projects
Social robots accelerating the transition to sustainable transport (50276-1)
Funder
Swedish Energy Agency, P2020-90133
Note

QC 20231017

Available from: 2023-10-16 Created: 2023-10-16 Last updated: 2025-02-07Bibliographically approved
Ahlberg, S., Axelsson, A., Yu, P., Shaw Cortez, W. E., Gao, Y., Ghadirzadeh, A., . . . Dimarogonas, D. V. (2022). Co-adaptive Human-Robot Cooperation: Summary and Challenges. Unmanned Systems, 10(02), 187-203
Open this publication in new window or tab >>Co-adaptive Human-Robot Cooperation: Summary and Challenges
Show others...
2022 (English)In: Unmanned Systems, ISSN 2301-3850, E-ISSN 2301-3869, Vol. 10, no 02, p. 187-203Article in journal (Refereed) Published
Abstract [en]

The work presented here is a culmination of developments within the Swedish project COIN: Co-adaptive human-robot interactive systems, funded by the Swedish Foundation for Strategic Research (SSF), which addresses a unified framework for co-adaptive methodologies in human-robot co-existence. We investigate co-adaptation in the context of safe planning/control, trust, and multi-modal human-robot interactions, and present novel methods that allow humans and robots to adapt to one another and discuss directions for future work.

Place, publisher, year, edition, pages
World Scientific Pub Co Pte Ltd, 2022
Keywords
Co-adaptive systems, human-in-the-loop systems, human-robot interaction
National Category
Robotics and automation
Identifiers
urn:nbn:se:kth:diva-310041 (URN)10.1142/S230138502250011X (DOI)000761503800006 ()2-s2.0-85116890059 (Scopus ID)
Note

QC 20220321

Available from: 2022-03-21 Created: 2022-03-21 Last updated: 2025-02-09Bibliographically approved
Axelsson, A., Buschmeier, H. & Skantze, G. (2022). Modeling Feedback in Interaction With Conversational Agents—A Review. Frontiers in Computer Science, 4, Article ID 744574.
Open this publication in new window or tab >>Modeling Feedback in Interaction With Conversational Agents—A Review
2022 (English)In: Frontiers in Computer Science, E-ISSN 2624-9898, Vol. 4, article id 744574Article, review/survey (Refereed) Published
Abstract [en]

Intelligent agents interacting with humans through conversation (such as a robot, embodied conversational agent, or chatbot) need to receive feedback from the human to make sure that its communicative acts have the intended consequences. At the same time, the human interacting with the agent will also seek feedback, in order to ensure that her communicative acts have the intended consequences. In this review article, we give an overview of past and current research on how intelligent agents should be able to both give meaningful feedback toward humans, as well as understanding feedback given by the users. The review covers feedback across different modalities (e.g., speech, head gestures, gaze, and facial expression), different forms of feedback (e.g., backchannels, clarification requests), and models for allowing the agent to assess the user's level of understanding and adapt its behavior accordingly. Finally, we analyse some shortcomings of current approaches to modeling feedback, and identify important directions for future research.

Place, publisher, year, edition, pages
Frontiers Media SA, 2022
Keywords
feedback, grounding, spoken dialogue, multimodal signals, human-agent interaction, review, återmatning, reaktion, dialog, multimodala signaler, multimodalitet, människa-agent-kommunikation, recension
National Category
Human Computer Interaction
Research subject
Human-computer Interaction; Speech and Music Communication
Identifiers
urn:nbn:se:kth:diva-310401 (URN)10.3389/fcomp.2022.744574 (DOI)000778821100001 ()2-s2.0-85127522996 (Scopus ID)
Projects
COIN-SSFConstructing Explainability (438445824)
Funder
Swedish Foundation for Strategic ResearchGerman Research Foundation (DFG)
Note

QC 20220429

Available from: 2022-03-30 Created: 2022-03-30 Last updated: 2022-06-25Bibliographically approved
Axelsson, A. & Skantze, G. (2022). Multimodal User Feedback During Adaptive Robot-Human Presentations. Frontiers in Computer Science, 3
Open this publication in new window or tab >>Multimodal User Feedback During Adaptive Robot-Human Presentations
2022 (English)In: Frontiers in Computer Science, E-ISSN 2624-9898, Vol. 3Article in journal (Refereed) Published
Abstract [en]

Feedback is an essential part of all communication, and agents communicating with humans must be able to both give and receive feedback in order to ensure mutual understanding. In this paper, we analyse multimodal feedback given by humans towards a robot that is presenting a piece of art in a shared environment, similar to a museum setting. The data analysed contains both video and audio recordings of 28 participants, and the data has been richly annotated both in terms of multimodal cues (speech, gaze, head gestures, facial expressions, and body pose), as well as the polarity of any feedback (negative, positive, or neutral). We train statistical and machine learning models on the dataset, and find that random forest models and multinomial regression models perform well on predicting the polarity of the participants' reactions. An analysis of the different modalities shows that most information is found in the participants' speech and head gestures, while much less information is found in their facial expressions, body pose and gaze. An analysis of the timing of the feedback shows that most feedback is given when the robot makes pauses (and thereby invites feedback), but that the more exact timing of the feedback does not affect its meaning.

Place, publisher, year, edition, pages
Frontiers Media SA, 2022
Keywords
feedback, presentation, agent, robot, grounding, polarity, backchannel, multimodal, återmatning, presentation, agent, robot, återkoppling, polaritet, reaktion, multimodal
National Category
Human Computer Interaction
Research subject
Speech and Music Communication
Identifiers
urn:nbn:se:kth:diva-307105 (URN)10.3389/fcomp.2021.741148 (DOI)000745131900001 ()2-s2.0-85123110812 (Scopus ID)
Projects
Co-adaptive Human-Robot Interactive Systems
Funder
Swedish Foundation for Strategic Research, COIN
Note

QC 20220112 QC 20220216

Available from: 2022-01-11 Created: 2022-01-11 Last updated: 2022-06-25Bibliographically approved
Axelsson, N. & Skantze, G. (2020). Using knowledge graphs and behaviour trees for feedback-aware presentation agents. In: Proceedings of Intelligent Virtual Agents 2020: . Paper presented at Intelligent Virtual Agents 2020 University of Glasgow Glasgow, UK, October 19-23, 2020. Association for Computing Machinery (ACM)
Open this publication in new window or tab >>Using knowledge graphs and behaviour trees for feedback-aware presentation agents
2020 (English)In: Proceedings of Intelligent Virtual Agents 2020, Association for Computing Machinery (ACM) , 2020Conference paper, Published paper (Refereed)
Abstract [en]

In this paper, we address the problem of how an interactive agent (such as a robot) can present information to an audience and adaptthe presentation according to the feedback it receives. We extend a previous behaviour tree-based model to generate the presentation from a knowledge graph (Wikidata), which allows the agent to handle feedback incrementally, and adapt accordingly. Our main contribution is using this knowledge graph not just for generating the system’s dialogue, but also as the structure through which short-term user modelling happens. In an experiment using simulated users and third-party observers, we show that referring expressions generated by the system are rated more highly when they adapt to the type of feedback given by the user, and when they are based on previously grounded information as opposed to new information.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2020
National Category
Computer graphics and computer vision Computer Sciences
Research subject
Human-computer Interaction
Identifiers
urn:nbn:se:kth:diva-281996 (URN)10.1145/3383652.3423884 (DOI)000728153600026 ()2-s2.0-85096989501 (Scopus ID)
Conference
Intelligent Virtual Agents 2020 University of Glasgow Glasgow, UK, October 19-23, 2020
Projects
Co-adaptive human-robot interactive systems
Funder
Swedish Foundation for Strategic Research
Note

QC 20200929

conference ISBN: 9781450375863

Available from: 2020-09-29 Created: 2020-09-29 Last updated: 2025-02-01Bibliographically approved
Axelsson, N. & Skantze, G. (2019). Modelling Adaptive Presentations in Human-Robot Interaction using Behaviour Trees. In: Satoshi Nakamura (Ed.), 20th Annual Meeting of the Special Interest Group on Discourse and Dialogue: Proceedings of the Conference. Paper presented at 20th Annual SIGdial Meeting on Discourse and Dialogue, SIGdial 2019, Stockholm, Sweden, September 11-13, 2019 (pp. 345-352). Stroudsburg, PA: Association for Computational Linguistics (ACL)
Open this publication in new window or tab >>Modelling Adaptive Presentations in Human-Robot Interaction using Behaviour Trees
2019 (English)In: 20th Annual Meeting of the Special Interest Group on Discourse and Dialogue: Proceedings of the Conference / [ed] Satoshi Nakamura, Stroudsburg, PA: Association for Computational Linguistics (ACL) , 2019, p. 345-352Conference paper, Published paper (Refereed)
Abstract [en]

In dialogue, speakers continuously adapt their speech to accommodate the listener, based on the feedback they receive. In this paper, we explore the modelling of such behaviours in the context of a robot presenting a painting. A Behaviour Tree is used to organise the behaviour on different levels, and allow the robot to adapt its behaviour in real-time; the tree organises engagement, joint attention, turn-taking, feedback and incremental speech processing. An initial implementation of the model is presented, and the system is evaluated in a user study, where the adaptive robot presenter is compared to a non-adaptive version. The adaptive version is found to be more engaging by the users, although no effects are found on the retention of the presented material.

Place, publisher, year, edition, pages
Stroudsburg, PA: Association for Computational Linguistics (ACL), 2019
Keywords
human-robot interaction, presentation, acceptance, understanding, hearing, attention, robot, Furhat, presenter, adaptive, non-adaptive, retention, engagement, interaktion, presentation, acceptans, förståelse, förstånd, hörsel, uppmärksamhet, robot, Furhat, presentatör, adaptiv, ickeadaptiv, minne, ihågkomst, engagemang
National Category
Computer graphics and computer vision
Research subject
Speech and Music Communication; Human-computer Interaction
Identifiers
urn:nbn:se:kth:diva-267218 (URN)10.18653/v1/W19-5940 (DOI)000591510500040 ()2-s2.0-85083155670 (Scopus ID)
Conference
20th Annual SIGdial Meeting on Discourse and Dialogue, SIGdial 2019, Stockholm, Sweden, September 11-13, 2019
Projects
Co-adaptive Human-Robot Interactive Systems
Funder
Swedish Foundation for Strategic Research, RIT15-0133
Note

Part of proceedings: ISBN 978-1-950737-61-1

QC 20200205

Available from: 2020-02-04 Created: 2020-02-04 Last updated: 2025-02-07Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0003-0112-6732

Search in DiVA

Show all publications