Gesture and Audio-Haptic Guidance Techniques to Direct Conversations with Intelligent Voice Interfaces

要旨

Advances in large language models (LLMs) empower new interactive capabilities for wearable voice interfaces, yet traditional voice-and-audio I/O techniques limit users' ability to flexibly navigate information and manage timing for complex conversational tasks. We developed a suite of gesture and audio-haptic guidance techniques that enable users to control conversation flows and maintain awareness of possible future actions, while simultaneously contributing and receiving conversation content through voice and audio. A 14-participant exploratory study compared our parallelized I/O techniques to a baseline of voice-only interaction. The results demonstrate the efficiency of gestures and haptics for information access, while allowing system speech to be redirected and interrupted in a socially acceptable manner. The techniques also raised user awareness of how to leverage intelligent capabilities. Our findings inform design recommendations to facilitate role-based collaboration between multimodal I/O techniques and reduce users' perception of time pressure when interleaving interactions with system speech.

著者
Shwetha Rajaram
Meta, Toronto, Ontario, Canada
Hemant Bhaskar. Surale
Meta, Toronto, Ontario, Canada
Codie McConkey
Meta, Toronto, Ontario, Canada
Carine Rognon
Meta, Redmond, Washington, United States
Hrim Mehta
Meta, Toronto, Ontario, Canada
Michael Glueck
Meta, Toronto, Ontario, Canada
Christopher Collins
Meta, Toronto, Ontario, Canada
DOI

10.1145/3706598.3714310

論文URL

https://dl.acm.org/doi/10.1145/3706598.3714310

動画

会議: CHI 2025

The ACM CHI Conference on Human Factors in Computing Systems (https://chi2025.acm.org/)

セッション: Vibration Vibes

G316+G317
7 件の発表
2025-05-01 18:00:00
2025-05-01 19:30:00
日本語まとめ
読み込み中…