Immersive Touch and Gesture Interaction

会議の名前
CHI 2025
Sketch2Terrain: AI-Driven Real-Time Terrain Sketch Mapping in Augmented Reality
要旨

Sketch mapping is an effective technique to externalize and communicate spatial information. However, it has been limited to 2D mediums, making it difficult to represent 3D information, particularly for terrains with elevation changes. We present Sketch2Terrain, an intuitive generative-3D-sketch-mapping system combining freehand sketching with generative Artificial Intelligence that radically changes sketch map creation and representation using Augmented Reality. Sketch2Terrain empowers non-experts to create unambiguous sketch maps of natural environments and provides a homogeneous interface for researchers to collect data and conduct experiments. A between-subject study (N=36) revealed that generative-3D-sketch-mapping improved efficiency by 38.4%, terrain-topology accuracy by 12.5%, and landmark accuracy by up to 12.1%, with only a 4.7% trade-off in terrain-elevation accuracy compared to freehand 3D-sketch-mapping. Additionally, generative-3D-sketch-mapping reduced perceived strain by 60.5% and stress by 39.5% over 2D-sketch-mapping. These findings underscore potential applications of generative-3D-sketch-mapping for in-depth understanding and accurate representation of vertically complex environments. The implementation is publicly available.

著者
Tianyi Xiao
Institute of Cartography and Geoinformation, ETH Zurich, Zürich, Switzerland
Yizi Chen
IKG, Zurich, Switzerland
Sailin Zhong
ETH Zürich, Zürich, Switzerland
Peter Kiefer
ETH Zurich, Zurich, ZH, Switzerland
Jakub Krukar
University of Muenster, Muenster, Germany
Kevin Gonyop Kim
FHNW University of Applied Sciences and Arts Northwestern Switzerland, Windisch, Switzerland
Lorenz Hurni
ETH Zurich, Zurich, Switzerland
Angela Schwering
University of Muenster, Muenster, Germany
Martin Raubal
ETH Zurich, Zurich, ZH, Switzerland
DOI

10.1145/3706598.3713467

論文URL

https://dl.acm.org/doi/10.1145/3706598.3713467

動画
PatternTrack: Multi-Device Tracking Using Infrared, Structured-Light Projections from Built-in LiDAR
要旨

As augmented reality devices (e.g., smartphones and headsets) proliferate in the market, multi-user AR scenarios are set to become more common. Co-located users will want to share coherent and synchronized AR experiences, but this is surprisingly cumbersome with current methods. In response, we developed PatternTrack, a novel tracking approach that repurposes the structured infrared light patterns emitted by VCSEL-driven depth sensors, like those found in the Apple Vision Pro, iPhone, iPad, and Meta Quest 3. Our approach is infrastructure-free, requires no pre-registration, works on featureless surfaces, and provides the real-time 3D position and orientation of other users' devices. In our evaluation --- tested on six different surfaces and with inter-device distances of up to 260 cm --- we found a mean 3D positional tracking error of 11.02 cm and a mean angular error of 6.81°.

著者
Daehwa Kim
Carnegie Mellon University, Pittsburgh, Pennsylvania, United States
Robert Xiao
University of British Columbia, Vancouver, British Columbia, Canada
Chris Harrison
Carnegie Mellon University, Pittsburgh, Pennsylvania, United States
DOI

10.1145/3706598.3713388

論文URL

https://dl.acm.org/doi/10.1145/3706598.3713388

動画
Online-EYE: Multimodal Implicit Eye Tracking Calibration for XR
要旨

Unlike other inputs for extended reality (XR) that work out of the box, eye tracking typically requires custom calibration per user or session. We present a multimodal inputs approach for implicit calibration of eye tracker in VR, leveraging UI interaction for continuous, background calibration. Our method analyzes gaze data alongside controller interaction with UI elements, and employing ML techniques it continuously refines the calibration matrix without interrupting users from their current tasks. Potentially eliminating the need for explicit calibration. We demonstrate the accuracy and effectiveness of this implicit approach across various tasks and real time applications achieving comparable eye tracking accuracy to native, explicit calibration. While our evaluation focuses on VR and controller-based interactions, we anticipate the broader applicability of this approach to various XR devices and input modalities.

著者
Baosheng James HOU
Google, Seattle, Washington, United States
Lucy Abramyan
Google, Mountain View, California, United States
Prasanthi Gurumurthy
Google, Mountain View, California, United States
Haley Adams
Google, Mountain View, California, United States
Ivana Tosic Rodgers
Google, Mountain View, California, United States
Eric J. Gonzalez
Google, Seattle, Washington, United States
Khushman Patel
Google Inc, Mountain View, California, India
Andrea Colaço
Google, Mountain View, California, United States
Ken Pfeuffer
Aarhus University, Aarhus, Denmark
Hans Gellersen
Lancaster University, Lancaster, United Kingdom
Karan Ahuja
Google, Seattle, Washington, United States
Mar Gonzalez-Franco
Google, Seattle, Washington, United States
DOI

10.1145/3706598.3713461

論文URL

https://dl.acm.org/doi/10.1145/3706598.3713461

動画
Mid-Air Gestures for Proactive Olfactory Interactions in Virtual Reality
要旨

Olfactory experiences are increasingly in demand due to their immersive benefits. However, most interaction implementations are passive and rely on conventions established for other modalities. In this work, we investigated proactive olfactory interactions, where users actively engage with scents, focusing on mid-air gestures as an input modality miming real-world object- and scent-manipulation, e.g., fanning away an odor. Our study had participants develop a user-defined gesture set for interacting with scents in Virtual Reality (VR), covering various object types (solid, liquid, gas) and interaction modes (out-of-reach, \revision{not graspable}, graspable), participants compared interacting with scents in VR using traditional controllers versus proactive gestures, revealing that proactive gestures enhanced user experience, presence, and task performance. Finally, an exploratory study showed strong participants' preferences for personalization, enhanced interaction capabilities, and multi-sensory integration. Based on these findings, we propose design guidelines and applications for proactive interactions with scents.

著者
Junxian Li
Zhejiang University, Hangzhou, Zhejiang, China
Yanan Wang
Donghua University, Shanghai , China
Zhitong Cui
Zhejiang University, Hangzhou, Zhejiang, China
Jas Brooks
University of Chicago, Chicago, Illinois, United States
Yifan Yan
Donghua University, Shanghai, China
Zhengyu Lou
Colledge of Fashion and Design, Shanghai, Shanghai, China
Yucheng Li
Donghua University, Shanghai, China
DOI

10.1145/3706598.3713964

論文URL

https://dl.acm.org/doi/10.1145/3706598.3713964

動画
HaloTouch: Using IR Multi-Path Interference to Support Touch Interactions with General Surfaces
要旨

Sensing touch on arbitrary surfaces has long been a goal of ubiquitous computing, but often requires instrumenting the surface. Depth camera-based systems have emerged as a promising solution for minimizing instrumentation, but at the cost of high touch-down detection error rates, high touch latency, and high minimum hover distance, limiting them to basic tasks. We developed HaloTouch, a vision-based system which exploits a multipath interference effect from an off-the-shelf time-of-flight depth camera to enable fast, accurate touch interactions on general surfaces. HaloTouch achieves a 99.2% touch-down detection accuracy across various materials, with a motion-to-photon latency of 150 ms. With a brief (20s) user-specific calibration, HaloTouch supports millimeter-accurate hover sensing as well as continuous pressure sensing. We conducted a user study with 12 participants, including a typing task demonstrating text input at 26.3 AWPM. HaloTouch shows promise for more robust, dynamic touch interactions without instrumenting surfaces or adding hardware to users.

著者
Ziyi Xia
University of British Columbia, Vancouver, British Columbia, Canada
Xincheng Huang
University of British Columbia, Vancouver, British Columbia, Canada
Sidney S. Fels
University of British Columbia, Vancouver, British Columbia, Canada
Robert Xiao
University of British Columbia, Vancouver, British Columbia, Canada
DOI

10.1145/3706598.3714179

論文URL

https://dl.acm.org/doi/10.1145/3706598.3714179

動画
Understanding Usability of VR Pointing Methods with a Handheld-style HMD for Onsite Exhibitions
要旨

Handheld-style head-mounted displays (HMDs) are becoming increasingly popular as a convenient option for onsite exhibitions. However, they lack established practices for basic interactions, particularly pointing methods. Through our formative study involving practitioners, we discovered that controllers and hand gestures are the primary pointing methods being utilized. Building upon these findings, we conducted a usability study to explore seven different pointing methods, incorporating insights from the formative study and current virtual reality (VR) practices. The results showed that while controllers remain a viable option, hand gestures are not recommended. Notably, dwell time-based methods, which are not fast and are not commonly recognized by practitioners, demonstrate high usability and user confidence, particularly for inexperienced VR users. We recommend the use of dwell-based methods for onsite exhibition contexts. This research provides insights for the adoption of handheld-style HMDs, laying the groundwork for improving user interaction in exhibition environments, thereby potentially enhancing visitor experiences.

受賞
Honorable Mention
著者
Yuki Abe
Hokkaido University, Sapporo, Japan
Kan Kusakabe
HokkaidoUniversity, Sapporo, Japan
Myungguen Choi
University of Tsukuba, Tsukuba, Ibaraki, Japan
Daisuke Sakamoto
Hokkaido University, Sapporo, Japan
Tetsuo Ono
Hokkaido University, Sapporo, Hokkaido, Japan
DOI

10.1145/3706598.3713874

論文URL

https://dl.acm.org/doi/10.1145/3706598.3713874

動画
Palmpad: Enabling Real-Time Index-to-Palm Touch Interaction with a Single RGB Camera
要旨

Index-to-palm interaction plays a crucial role in Mixed Reality(MR) interactions. However, achieving a satisfactory inter-hand interaction experience is challenging with existing vision-based hand tracking technologies, especially in scenarios where only a single camera is available. Therefore, we introduce Palmpad, a novel sensing method utilizing a single RGB camera to detect the touch of an index finger on the opposite palm. Our exploration reveals that the incorporation of optical flow techniques to extract motion information between consecutive frames for the index finger and palm leads to a significant improvement in touch status determination. By doing so, our CNN model achieves 97.0% recognition accuracy and a 96.1% F1 score. In usability evaluation, we compare Palmpad with Quest's inherent hand gesture algorithms. Palmpad not only delivers superior accuracy 95.3% but also reduces operational demands and significantly improves users’ willingness and confidence. Palmpad aims to enhance accurate touch detection for lightweight MR devices.

著者
Zhe He
Tsinghua University, Beijing, Beijing, China
Xiangyang Wang
Tsinghua University, Beijing, China
Yuanchun Shi
Tsinghua University, Beijing, China
Chi Hsia
Tsinghua University, Beijing, China
Chen Liang
The Hong Kong University of Science and Technology (Guangzhou), Guangzhou, Guangdong, China
Chun Yu
Tsinghua University, Beijing, China
DOI

10.1145/3706598.3714130

論文URL

https://dl.acm.org/doi/10.1145/3706598.3714130

動画