Beyond Words: Measuring User Experience through Speech Analysis in Voice User Interfaces

要旨

Voice assistants (VAs) are typically evaluated through task performance metrics and self-report questionnaires, but people’s voices themselves carry rich paralinguistic cues that reveal affect, effort, and interaction breakdowns. We present a within-subjects study (N=49) that systematically compared three VA personas across three usage scenarios to investigate whether speech-derived audio features can serve as a proxy for user experience (UX). Participants’ speech was analyzed for temporal, spectral, and linguistic markers, alongside standardized UX measures, brief mood and stress ratings, and a post-study questionnaire. We found correlations between specific speech features and self-reported satisfaction and experience. Furthermore, a machine learning model trained on speech features achieved promising accuracy in classifying UX levels, indicating that this might be a reasonable alternative to self-report instruments. Our findings establish speech as a viable, real-time signal for implicitly measuring UX and point toward adaptive VUIs that respond dynamically to emotional and usability-related vocal cues.

受賞
Honorable Mention
著者
Yong Ma
University of Bergen, Bergen, Norway
Xuesong Zhang
Southern University of Science and Technology, Shenzhen, China
Xuedong Zhang
LMU Munich, Munich, Germany
Natalia Bartłomiejczyk
University of Neuchâtel, Neuchâtel, Switzerland
Seungwoo Je
Southern University of Science and Technology, Shenzhen, China
Adrian Holzer
University of Neuchâtel, Neuchâtel, Switzerland
Morten Fjeld
University of Bergen, Bergen, Norway
Andreas Martin. Butz
LMU Munich, Munich, Germany

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: Research Methodology & UX

P1 - Room 134
7 件の発表
2026-04-16 20:15:00
2026-04-16 21:45:00