Effective 2D Stroke-based Gesture Augmentation for RNNs

要旨

Recurrent neural networks (RNN) require large training datasets from which they learn new class models. This limitation prohibits their use in custom gesture applications where only one or two end user samples are given per gesture class. One common way to enhance sparse datasets is to use data augmentation to synthesize new samples. Although there are numerous known techniques, they are often treated as standalone approaches when in reality they are often complementary. We show that by intelligently chaining augmentation techniques together that simulate different gesture production variability types, such as those affecting the temporal and spatial qualities of a gesture, we can significantly increase RNN accuracy without sacrificing training time. Through experimentation on four public 2D gesture datasets, we show that RNNs trained with our data augmentation chaining technique achieves state-of-the-art recognition accuracy in both writer-dependent and writer-independent test scenarios.

著者
Mykola Maslych
University of Central Florida, Orlando, Florida, United States
Eugene Matthew. Taranta
University of Central Florida, Orlando, Florida, United States
Mostafa Aldilati
University of Central Florida, Orlando, Florida, United States
Joseph LaViola
University of Central Florida, Orlando, Florida, United States
論文URL

https://doi.org/10.1145/3544548.3581358

動画

会議: CHI 2023

The ACM CHI Conference on Human Factors in Computing Systems (https://chi2023.acm.org/)

セッション: GUIs, Gaze, and Gesture-based Interaction

Hall C
6 件の発表
2023-04-25 18:00:00
2023-04-25 19:30:00