TouchScribe: Augmenting Non-Visual Hand-Object Interactions with Automated Live Visual Descriptions

要旨

People who are blind or have low vision regularly use their hands to interact with the physical world to gain access to objects' shape, size, weight, and texture. However, many rich visual features remain inaccessible through touch alone, making it difficult to distinguish similar objects, interpret visual affordances, and form a complete understanding of objects. In this work, we present TouchScribe, a system that augments hand-object interactions with automated live visual descriptions. We trained a custom egocentric hand interaction model to recognize both common gestures (e.g., grab to inspect, hold side-by-side to compare) and unique ones by blind people (e.g., point to explore color, or swipe to read available texts). Furthermore, TouchScribe provides real-time and adaptive feedback based on hand movement, from hand interaction states, to object labels, and to visual details. Our user study and technical evaluations demonstrate that TouchScribe can provide rich and useful descriptions to support object understanding. Finally, we discuss the implications of making live visual descriptions responsive to users' physical reach.

著者
Ruei-Che Chang
University of Michigan, Ann Arbor, Michigan, United States
Rosiana Natalie
University of Michigan, Ann Arbor, Michigan, United States
Wenqian Xu
University of Michigan, Ann Arbor, Michigan, United States
Jovan Zheng Feng Yap
University of Michigan, Ann Arbor, Michigan, United States
Tiange Luo
University of Michigan, Ann Arbor, Michigan, United States
Venkatesh Potluri
University of Michigan, Ann Arbor, Michigan, United States
Anhong Guo
University of Michigan, Ann Arbor, Michigan, United States

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: Haptic and Multisensory Feedback

P1 - Room 118
7 件の発表
2026-04-13 20:15:00
2026-04-13 21:45:00