Enhancing UX Evaluation Through Collaboration with Conversational AI Assistants: Effects of Proactive Dialogue and Timing

要旨

Usability testing is vital for enhancing the user experience (UX) of interactive systems. However, analyzing test videos is complex and resource-intensive. Recent AI advancements have spurred exploration into human-AI collaboration for UX analysis, particularly through natural language. Unlike user-initiated dialogue, our study investigated the potential of proactive conversational assistants to aid UX evaluators through automatic suggestions at three distinct times: before, in sync with, and after potential usability problems. We conducted a hybrid Wizard-of-Oz study involving 24 UX evaluators, using ChatGPT to generate automatic problem suggestions and a human actor to respond to impromptu questions. While timing did not significantly impact analytic performance, suggestions appearing after potential problems were preferred, enhancing trust and efficiency. Participants found the automatic suggestions useful, but they collectively identified more than twice as many problems, underscoring the irreplaceable role of human expertise. Our findings also offer insights into future human-AI collaborative tools for UX evaluation.

著者
Emily Kuang
Rochester Institute of Technology, Rochester, New York, United States
Minghao Li
Nanyang Technological University, Singapore, Singapore
Mingming Fan
The Hong Kong University of Science and Technology (Guangzhou), Guangzhou, China
Kristen Shinohara
Rochester Institute of Technology, Rochester, New York, United States
論文URL

doi.org/10.1145/3613904.3642168

動画

会議: CHI 2024

The ACM CHI Conference on Human Factors in Computing Systems (https://chi2024.acm.org/)

セッション: AI and Interaction Design

320 'Emalani Theater
5 件の発表
2024-05-13 20:00:00
2024-05-13 21:20:00