Supporting Accessibility of Text, Image and Video A

会議の名前
CHI 2024
“It’s Kind of Context Dependent”: Understanding Blind and Low Vision People’s Video Accessibility Preferences Across Viewing Scenarios
要旨

While audio description (AD) is the standard approach for making videos accessible to blind and low vision (BLV) people, existing AD guidelines do not consider BLV users’ varied preferences across viewing scenarios. These scenarios range from how-to videos on YouTube, where users seek to learn new skills, to historical dramas on Netflix, where a user’s goal is entertainment. Additionally, the increase in video watching on mobile devices provides an opportunity to integrate nonverbal output modalities (e.g., audio cues, tactile elements, and visual enhancements). Through a formative survey and 15 semi-structured interviews, we identified BLV people’s video accessibility preferences across diverse scenarios. For example, participants valued action and equipment details for how-to videos, tactile graphics for learning scenarios, and 3D models for fantastical content. We define a six-dimensional video accessibility design space to guide future innovation and discuss how to move from “one-size-fits-all” paradigms to scenario-specific approaches.

著者
Lucy Jiang
Cornell University, Ithaca, New York, United States
Crescentia Jung
Cornell University, New York, New York, United States
Mahika Phutane
Cornell University, New York, New York, United States
Abigale Stangl
Georgia Institute of Technology, Atlanta, Georgia, United States
Shiri Azenkot
Cornell Tech, New York, New York, United States
論文URL

doi.org/10.1145/3613904.3642238

動画
GazePrompt: Enhancing Low Vision People's Reading Experience with Gaze-Aware Augmentations
要旨

Reading is a challenging task for low vision people. While conventional low vision aids (e.g., magnification) offer certain support, they cannot fully address the difficulties faced by low vision users, such as locating the next line and distinguishing similar words. To fill this gap, we present GazePrompt, a gaze-aware reading aid that provides timely and targeted visual and audio augmentations based on users' gaze behaviors. GazePrompt includes two key features: (1) a Line-Switching support that highlights the line a reader intends to read; and (2) a Difficult-Word support that magnifies or reads aloud a word that the reader hesitates with. Through a study with 13 low vision participants who performed well-controlled reading-aloud tasks with and without GazePrompt, we found that GazePrompt significantly reduced participants' line switching time, reduced word recognition errors, and improved their subjective reading experiences. A follow-up silent-reading study showed that GazePrompt can enhance users' concentration and perceived comprehension of the reading contents. We further derive design considerations for future gaze-based low vision aids.

著者
Ru Wang
University of Wisconsin-Madison, Madison, Wisconsin, United States
Zach Potter
University of Wisconsin-Madison, Madison, Wisconsin, United States
Yun Ho
University of Wisconsin-Madison, Madison, Wisconsin, United States
Daniel Killough
University of Wisconsin-Madison, Madison, Wisconsin, United States
Linxiu Zeng
University of Wisconsin-Madison, Madison, Wisconsin, United States
Sanbrita Mondal
University of Wisconsin-Madison, Madison, Wisconsin, United States
Yuhang Zhao
University of Wisconsin-Madison, Madison, Wisconsin, United States
論文URL

doi.org/10.1145/3613904.3642878

動画
Constrained Highlighting in a Document Reader can Improve Reading Comprehension
要旨

Highlighting text in a document is a common active reading strategy to remember information from documents. Learning theory suggests that for highlights to be effective, readers must be selective with what they choose to highlight. We investigate if an imposed user interface constraint limiting the number of highlighted words in a document reader can improve reading comprehension. A large-scale between-subjects experiment shows that constraining the number of words that can be highlighted leads to higher reading comprehension scores than highlighting nothing or highlighting an unlimited number of words. Our work empirically validates theories in psychology, which in turn enables several new research directions within HCI.

受賞
Best Paper
著者
Nikhita Joshi
University of Waterloo, Waterloo, Ontario, Canada
Daniel Vogel
University of Waterloo, Waterloo, Ontario, Canada
論文URL

doi.org/10.1145/3613904.3642314

動画
Making Short-Form Videos Accessible with Hierarchical Video Summaries
要旨

Short videos on platforms such as TikTok, Instagram Reels, and YouTube Shorts (i.e. short-form videos) have become a primary source of information and entertainment. Many short-form videos are inaccessible to blind and low vision (BLV) viewers due to their rapid visual changes, on-screen text, and music or meme-audio overlays. In our formative study, 7 BLV viewers who regularly watched short-form videos reported frequently skipping such inaccessible content. We present ShortScribe, a system that provides hierarchical visual summaries of short-form videos at three levels of detail to support BLV viewers in selecting and understanding short-form videos. ShortScribe allows BLV users to navigate between video descriptions based on their level of interest. To evaluate ShortScribe, we assessed description accuracy and conducted a user study with 10 BLV participants comparing ShortScribe to a baseline interface. When using ShortScribe, participants reported higher comprehension and provided more accurate summaries of video content.

著者
Tess Van Daele
University of Texas at Austin, Austin, Texas, United States
Akhil Iyer
University of Texas at Austin, Austin, Texas, United States
Yuning Zhang
The University of Texas at Austin, Austin, Texas, United States
Jalyn C. Derry
University of Texas at Austin, Austin, Texas, United States
Mina Huh
University of Texas, Austin, Austin, Texas, United States
Amy Pavel
University of Texas, Austin, Austin, Texas, United States
論文URL

doi.org/10.1145/3613904.3642839

動画