Automated Class Discovery and One-Shot Interactions for Acoustic Activity Recognition

要旨

Acoustic activity recognition has emerged as a foundational element for imbuing devices with context-driven capabilities, enabling richer, more assistive, and more accommodating computational experiences. Traditional approaches rely either on custom models trained in situ, or general models pre-trained on preexisting data, with each approach having accuracy and user burden implications. We present Listen Learner, a technique for activity recognition that gradually learns events specific to a deployed environment while minimizing user burden. Specifically, we built an end-to-end system for self-supervised learning of events labelled through one-shot interaction. We describe and quantify system performance 1) on preexisting audio datasets, 2) on real-world datasets we collected, and 3) through user studies which uncovered system behaviors suitable for this new type of interaction. Our results show that our system can accurately and automatically learn acoustic events across environments (e.g., 97% precision, 87% recall), while adhering to users' preferences for non-intrusive interactive behavior.

受賞
Honorable Mention
キーワード
Automatic class discovery
Acoustic activity recognition
著者
Jason Wu
Carnegie Mellon University & Apple Inc., Pittsburgh, PA, USA
Chris Harrison
Carnegie Mellon University, Pittsburgh, PA, USA
Jeffrey P. Bigham
Apple Inc. & Carnegie Mellon University, Cupertino, CA, USA
Gierad Laput
Apple Inc. & Carnegie Mellon University, Cupertino, CA, USA
DOI

10.1145/3313831.3376875

論文URL

https://doi.org/10.1145/3313831.3376875

動画

会議: CHI 2020

The ACM CHI Conference on Human Factors in Computing Systems (https://chi2020.acm.org/)

セッション: Sensing the human

Paper session
312 NI'IHAU
5 件の発表
2020-04-29 18:00:00
2020-04-29 19:15:00
日本語まとめ
読み込み中…