Agentic Audio Moderators vs Humans in Think-Aloud Usability Testing

要旨

Agentic AI holds promise for usability testing, yet its role as an audio moderator in think-aloud protocols is not well understood. This study explores: (1) how to design and develop an agentic audio moderator for think-aloud usability testing, and (2) how participants moderated by an agentic moderator differ from those moderated by a human regarding task performance, verbalization behaviors, user experience, and social perceptions of the moderator. Using a design-based research approach, we interviewed nine UX experts, iteratively developed an AI moderator, and evaluated it in a randomized controlled trial (N=60) with a note-taking application. Results suggest that significant differences were not observed between AI and human moderators in task performance or verbalization behaviors, though AI moderators received lower social perception ratings. This work contributes the first design-oriented evaluation of AI moderators in usability testing, offering implications for developing more acceptable and effective agentic audio moderators.

著者
Wangda Zhu
The Hong Kong Polytechnic University, Kowloon, Hong Kong
Guang Chen
The Hong Kong Polytechnic University, Hong Kong, Hong Kong
Yao Wang
The Hong Kong Polytechnic University, Hung Hom, Hong Kong, China
Pengcheng An
Southern University of Science and Technology, Shenzhen, China
Jiachun Du
Piipivo Technology, Hangzhou, China
Chen Li
The Hong Kong Polytechnic University , Hong Kong, Hong Kong

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: HCAI and Collaboration

P1 - Room 130
6 件の発表
2026-04-15 18:00:00
2026-04-15 19:30:00