GestureMap: Supporting Visual Analytics and Quantitative Analysis of Motion Elicitation Data by Learning 2D Embeddings

Abstract

This paper presents GestureMap, a visual analytics tool for gesture elicitation which directly visualises the space of gestures. Concretely, a Variational Autoencoder embeds gestures recorded as 3D skeletons on an interactive 2D map. GestureMap further integrates three computational capabilities to connect exploration to quantitative measures: Leveraging DTW Barycenter Averaging (DBA), we compute average gestures to 1) represent gesture groups at a glance; 2) compute a new consensus measure (variance around average gesture); and 3) cluster gestures with k-means. We evaluate GestureMap and its concepts with eight experts and an in-depth analysis of published data. Our findings show how GestureMap facilitates exploring large datasets and helps researchers to gain a visual understanding of elicited gesture spaces. It further opens new directions, such as comparing elicitations across studies. We discuss implications for elicitation studies and research, and opportunities to extend our approach to additional tasks in gesture elicitation.

Authors
Hai Duong. Dang
University of Bayreuth, Bayreuth, Germany
Daniel Buschek
University of Bayreuth, Bayreuth, Germany
DOI

10.1145/3411764.3445765

Paper URL

https://doi.org/10.1145/3411764.3445765

Video

Conference: CHI 2021

The ACM CHI Conference on Human Factors in Computing Systems (https://chi2021.acm.org/)

Session: Engineering Development Support

[A] Paper Room 05, 2021-05-10 17:00:00~2021-05-10 19:00:00 / [B] Paper Room 05, 2021-05-11 01:00:00~2021-05-11 03:00:00 / [C] Paper Room 05, 2021-05-11 09:00:00~2021-05-11 11:00:00
Paper Room 05
14 items in this session
2021-05-10 08:00:00
2021-05-10 10:00:00
Japanese summary
ジェスチャを2次元空間にマッピングする視覚分析ツールGestureMapを開発。ジェスチャを統計グラフ、散布図、密度図などで表示し、他のジェスチャと比較するなどして分析を行う。被験者8人でユーザスタディを行い知見を整理。
2021-06-20 10:00:29
湯村 翼 Tsubasa YUMURA