SonifyAR: Context-Aware Sound Generation in Augmented Reality

要旨

Sound plays a crucial role in enhancing user experience and immersiveness in Augmented Reality (AR). However, current platforms lack support for AR sound authoring due to limited interaction types, challenges in collecting and specifying context information, and difficulty in acquiring matching sound assets. We present SonifyAR, an LLM-based AR sound authoring system that generates context-aware sound effects for AR experiences. SonifyAR expands the current design space of AR sound and implements a Programming by Demonstration (PbD) pipeline to automatically collect contextual information of AR events, including virtual-content-semantics and real-world context. This context information is then processed by a large language model to acquire sound effects with Recommendation, Retrieval, Generation, and Transfer methods. To evaluate the usability and performance of our system, we conducted a user study with eight participants and created five example applications, including an AR-based science experiment, and an assistive application for low-vision AR users.

著者
Xia Su
University of Washington, Seattle, Washington, United States
Jon E.. Froehlich
University of Washington, Seattle, Washington, United States
Eunyee Koh
Adobe Research, San Jose, California, United States
Chang Xiao
Adobe Research, San Jose, California, United States
論文URL

https://doi.org/10.1145/3654777.3676406

動画

会議: UIST 2024

ACM Symposium on User Interface Software and Technology

セッション: 2. Sound & Music

Westin: Allegheny 2
5 件の発表
2024-10-16 23:00:00
2024-10-17 00:15:00