Lost in Transcription: Subtitle Errors in Automatic Speech Recognition Reduce Speaker and Content Evaluations

要旨

Researchers have demonstrated that Automatic Speech Recognition (ASR) systems perform differently across demographic groups. In this work, we examined how subtitle errors affect evaluations of speakers and their content using a preregistered online experiment (N=207, U.S.-based crowdworkers). Participants watched speakers with various accents deliver a talk in which the subtitles were accurate or error-prone. Our results indicate that error-prone subtitles consistently reduce both speaker and content evaluations for all speakers. We did not see disparate impact between the accent groups, controlling for subtitle quality. Taken together, though, the findings of this short paper imply that speakers with accents for which ASR systems perform poorly are likely to be further penalized by viewers with lower evaluations.

著者
Kowe Kadoma
Cornell University, New York, New York, United States
Priyal Shrivastava
Carnegie Mellon University , Pittsburgh, Pennsylvania, United States
Mor Naaman
Cornell Tech, New York, New York, United States

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: Modeling Spatial, Linguistic, and Sensory Errors

P1 - Room 128
6 件の発表
2026-04-14 20:15:00
2026-04-14 21:45:00