The paper presents an iteratively constructed framework for integrating LLMs as speech interfaces for physically assistive robots. The framework was developed and refined through multiple stages of testing, culminating in a user study with 11 older adults at an independent living facility.
The initial version of the framework identified 5 key components: Environment Description, Robot Functions, Function Applications, Code Specifications, and Safety. Through pilot testing with lab members and a demonstration with community members, the framework was expanded to include additional components such as Robot Variables, Instructional Materials, User Control Functions, and Feedback.
The final user study with older adults validated the effectiveness of the framework. Participants found the speech interface easy to learn and use, with low reported workload. However, challenges were identified around consistently executing user-provided modifiers (e.g. "feed me a larger scoop") and processing non-predefined commands (e.g. mixing foods).
The paper concludes by presenting 5 design guidelines based on the user study findings: 1) Customization, 2) Multi-Step Instruction, 3) Consistency, 4) Comparable Time to Caregiver, and 5) Social Capability. These guidelines highlight the importance of human-centric considerations when integrating LLMs as assistive speech interfaces, beyond just prompt engineering.
לשפה אחרת
מתוכן המקור
arxiv.org
תובנות מפתח מזוקקות מ:
by Akhil Padman... ב- arxiv.org 04-08-2024
https://arxiv.org/pdf/2404.04066.pdfשאלות מעמיקות