Open this publication in new window or tab >>Show others...
2024 (English)In: 2024 Ieee Winter Conference On Applications Of Computer Vision Workshops, Wacvw 2024, Institute of Electrical and Electronics Engineers (IEEE) , 2024, p. 988-994Conference paper, Published paper (Refereed)
Abstract [en]
The evolution of autonomous driving has made remarkable advancements in recent years, evolving into a tangible reality. However, a human-centric large-scale adoption hinges on meeting a variety of multifaceted requirements. To ensure that the autonomous system meets the user's intent, it is essential to accurately discern and interpret user commands, especially in complex or emergency situations. To this end, we propose to leverage the reasoning capabilities of Large Language Models (LLMs) to infer system requirements from in-cabin users' commands. Through a series of experiments that include different LLM models and prompt designs, we explore the few-shot multivariate binary classification accuracy of system requirements from natural language textual commands. We confirm the general ability of LLMs to understand and reason about prompts but underline that their effectiveness is conditioned on the quality of both the LLM model and the design of appropriate sequential prompts. Code and models are public with the link https://github.com/KTH-RPL/DriveCmd_LLM.
Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2024
Series
IEEE Winter Conference on Applications of Computer Vision Workshops, ISSN 2572-4398
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:kth:diva-351635 (URN)10.1109/WACVW60836.2024.00108 (DOI)001223022200040 ()2-s2.0-85188691382 (Scopus ID)
Conference
IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), JAN 04-08, 2024, Waikoloa, HI
Note
QC 20240813
Part of ISBN 979-8-3503-7028-7, 979-8-3503-7071-3
2024-08-132024-08-132024-10-11Bibliographically approved