Current touch-based interactions on earphones are limited by hygiene concerns and the small interaction surface. Recent works attempt to bypass these issues with mid-air gesture systems using active acoustic sensing. However, these signals may be audible and pose potential hearing risks. To address this, we propose FingerBar, a mid-air gesture recognition system for earphones that relies solely on microphones without active signal transmission. FingerBar leverages the distinctive friction sounds generated by finger gestures to achieve gesture recognition. We design a gesture filtering pipeline to maintain robustness against daily noise. An adversarial training strategy further enhances user-independent performance. From a set of 16 gestures, we identify the 7 most suitable for FingerBar based on user acceptability. Extensive evaluations demonstrate high accuracy and robustness. Furthermore, a user study confirms the practicality and acceptability of the system. Our findings highlight the promise of passive acoustic sensing as a user-friendly interaction modality for earphones.
ACM CHI Conference on Human Factors in Computing Systems