Editable XAI: Toward Bidirectional Human-AI Alignment with Co-Editable Explanations of Interpretable Attributes

要旨

While Explainable AI (XAI) helps users understand AI decisions, misalignment in domain knowledge can lead to disagreement. This inconsistency hinders understanding, and because explanations are often read-only, users lack the control to improve alignment. We propose making XAI editable, allowing users to write rules to improve control and gain deeper understanding through the generation effect of active learning. We developed CoExplain, leveraging a neural network for universal representation and symbolic rules for intuitive reasoning on interpretable attributes. CoExplain explains the neural network with a faithful proxy decision tree, parses user-written rules as an equivalent neural network graph, and collaboratively optimizes the decision tree. In a user study (N=43), CoExplain and manually editable XAI improved user understanding and model alignment compared to read-only XAI. CoExplain was easier to use with fewer edits and less time. This work contributes Editable XAI for bidirectional AI alignment, improving understanding and control.

著者
Haoyang Chen
National University of Singapore, Singapore, Singapore
Jingwen Bai
National University of Singapore, Singapore, Singapore
Fang Tian
National University of Singapore, Singapore, Singapore, Singapore
Brian Y. Lim
National University of Singapore, Singapore, Singapore
動画

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: Personalization and Human-AI Alignment

P1 - Room 130
7 件の発表
2026-04-14 18:00:00
2026-04-14 19:30:00