WordGesture-GAN: Modeling Word-Gesture Movement with Generative Adversarial Network

要旨

Word-gesture production models that can synthesize word-gestures are critical to the training and evaluation of word-gesture keyboard decoders. We propose WordGesture-GAN, a conditional generative adversarial network that takes arbitrary text as input to generate realistic word-gesture movements in both spatial (i.e., $(x,y)$ coordinates of touch points) and temporal (i.e., timestamps of touch points) dimensions. WordGesture-GAN introduces a Variational Auto-Encoder to extract and embed variations of user-drawn gestures into a Gaussian distribution which can be sampled to control variation in generated gestures. Our experiments on a dataset with 38k gesture samples show that WordGesture-GAN outperforms existing gesture production models including the minimum jerk model [37] and the style-transfer GAN [31,32] in generating realistic gestures. Overall, our research demonstrates that the proposed GAN structure can learn variations in user-drawn gestures, and the resulting WordGesture-GAN can generate word-gesture movement and predict the distribution of gestures. WordGesture-GAN can serve as a valuable tool for designing and evaluating gestural input systems.

受賞
Honorable Mention
著者
Jeremy Chu
Stony Brook University, Stony Brook, New York, United States
Dongsheng An
Stony Brook University, Stony brook, New York, United States
Yan Ma
Stony Brook University, Stony Brook, New York, United States
Wenzhe Cui
Stony Brook University, Stony Brook, New York, United States
Shumin Zhai
Google, Mountain View, California, United States
Xianfeng David Gu
Stony Brook University, Stony Brook, New York, United States
Xiaojun Bi
Stony Brook University, Stony Brook, New York, United States
論文URL

https://doi.org/10.1145/3544548.3581279

動画

会議: CHI 2023

The ACM CHI Conference on Human Factors in Computing Systems (https://chi2023.acm.org/)

セッション: GUIs, Gaze, and Gesture-based Interaction

Hall C
6 件の発表
2023-04-25 18:00:00
2023-04-25 19:30:00