VisGuardian: A Lightweight Group-based Visual Privacy Control Technique For Smart Glasses in Home Environments

要旨

Always-on sensing of AI applications on AR glasses makes traditional permission techniques inefficient for context-dependent private visual data within home environments. Home presents a challenging privacy context due to massive sensitive objects and the intimate nature of daily routines. We propose VisGuardian, a fine-grained content-based visual permission technique for AR glasses. VisGuardian features a group-based control mechanism that enables users to efficiently manage permissions for multiple private objects. VisGuardian detects objects using YOLO and adopts a pre-classified schema to group them. By selecting a single object, users can obscure groups of related objects based on criteria including privacy sensitivity, object category, or spatial proximity. A technical evaluation shows VisGuardian achieves mAP50 of 0.6704 with only 14.0 ms latency and a 1.7% increase in battery consumption per hour. Furthermore, a user study (N=24) comparing VisGuardian to slider-based and object-based baselines found it to be significantly faster for setting permissions and was preferred by users for its efficiency, effectiveness, and ease of use.

著者
Shuning Zhang
Tsinghua University, Beijing, China
Qucheng Zang
Institute of Computational Arts, Hangzhou, China, China
Yongquan 'Owen' Hu
National University of Singapore, Singapore, Singapore
Jiachen Du
The Future Laboratory, Tsinghua University, Beijing, China
Xueyang Wang
Tsinghua University, Beijing, China
Yan Kong
CS, Beijing, China, China
Xinyi Fu
Tsinghua University, Beijing, China
Suranga Nanayakkara
School of Computing, National University of Singapore, Singapore, Singapore
Xin Yi
Tsinghua University, Beijing, China
Hewu Li
Tsinghua University, Beijing, China

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: AI Governance and Accountability

Area 1 + 2 + 3: theatre
7 件の発表
2026-04-13 20:15:00
2026-04-13 21:45:00