Explainable Moderation in Multiplayer Games: Player Responses to Explanations of an Automated Temporary Ban

要旨

The opaqueness of moderation systems can leave offenders of toxic behaviour disaffected and without recourse for change. We examined whether explainability, the means by which an automated system explains its decisions, can improve player responses to automated moderation decisions within the context of multiplayer games. Through a mixed methods experiment we evaluated players' perceptions of six explanations of an automated temporary ban decision. Despite finding only minor benefits to explainable AI methods over the best current explanation used in practice, we found that justification, i.e. providing evidence within an explanation, is fundamental for improving players' perceived fairness and emotional response to moderation. We contextualise these results through a reflexive thematic analysis in which we identify four themes that reflect players' competing understandings of both explainability and moderation. We conclude by proposing four design implications for researchers and developers to consider when designing explainability for future community management systems.

著者
Timothy Holland
The University of Melbourne, Melbourne, VIC, Australia
Lucy A.. Sparrow
The University of Melbourne, Melbourne, VIC, Australia
Wally Smith
The University of Melbourne, Melbourne, Victoria, Australia
Sarah Webber
University of Melbourne, Melbourne, Victoria, Australia

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: Toxicity and Moderation in Online Games

P1 - Room 113
7 件の発表
2026-04-13 20:15:00
2026-04-13 21:45:00