2. Shared Spaces

会議の名前
UIST 2024
BlendScape: Enabling End-User Customization of Video-Conferencing Environments through Generative AI
要旨

Today’s video-conferencing tools support a rich range of professional and social activities, but their generic meeting environments cannot be dynamically adapted to align with distributed collaborators’ needs. To enable end-user customization, we developed BlendScape, a rendering and composition system for video-conferencing participants to tailor environments to their meeting context by leveraging AI image generation techniques. BlendScape supports flexible representations of task spaces by blending users’ physical or digital backgrounds into unified environments and implements multimodal interaction techniques to steer the generation. Through an exploratory study with 15 end-users, we investigated whether and how they would find value in using generative AI to customize video-conferencing environments. Participants envisioned using a system like BlendScape to facilitate collaborative activities in the future, but required further controls to mitigate distracting or unrealistic visual elements. We implemented scenarios to demonstrate BlendScape's expressiveness for supporting environment design strategies from prior work and propose composition techniques to improve the quality of environments.

受賞
Honorable Mention
著者
Shwetha Rajaram
Microsoft Research, Redmond, Washington, United States
Nels Numan
Microsoft Research, Redmond, Washington, United States
Bala Kumaravel
Microsoft Research, Redmond, Washington, United States
Nicolai Marquardt
Microsoft Research, Redmond, Washington, United States
Andrew D. Wilson
Microsoft Research, Redmond, Washington, United States
論文URL

https://doi.org/10.1145/3654777.3676326

動画
MyWebstrates: Webstrates as Local-first Software
要旨

Webstrates are web substrates, a practical realization of shareable dynamic media under which distributability, shareability, and malleability are fundamental software principles. Webstrates blur the distinction between application and document in a way that enables users to share, repurpose, and refit software across a variety of domains, but its reliance on a central server constrains its use; it is at odds with personal and collective control of data; and limits applications to the web. We extend the fundamental principles to include interoperability and sovereignty over data and propose MyWebstrates, an implementation of Webstrates on top of a new, lower-level substrate for synchronization built around local-first software principles. MyWebstrates registers itself in the user’s browser and function as a piece of local software that can selectively synchronise data over sync servers or peer-to-peer connections. We show how MyWebstrates extends Webstrates to enable offline collaborative use, interoperate between Webstrates on non-web technologies such as Unity, and maintain personal and collective sovereignty over data. We demonstrate how this enables new types of applications of Webstrates and discuss limitations of this approach and new challenges that it reveals.

著者
Clemens Nylandsted. Klokmose
Aarhus University, Aarhus, Denmark
James R.. Eagan
Institut Polytechnique de Paris, Paris, France
Peter van Hardenberg
Ink & Switch, San Francisco, California, United States
論文URL

https://doi.org/10.1145/3654777.3676445

動画
SituationAdapt: Contextual UI Optimization in Mixed Reality with Situation Awareness via LLM Reasoning
要旨

Mixed Reality is increasingly used in mobile settings beyond controlled home and office spaces. This mobility introduces the need for user interface layouts that adapt to varying contexts. However, existing adaptive systems are designed only for static environments. In this paper, we introduce SituationAdapt, a system that adjusts Mixed Reality UIs to real-world surroundings by considering environmental and social cues in shared settings. Our system consists of perception, reasoning, and optimization modules for UI adaptation. Our perception module identifies objects and individuals around the user, while our reasoning module leverages a Vision-and-Language Model to assess the placement of interactive UI elements. This ensures that adapted layouts do not obstruct relevant environmental cues or interfere with social norms. Our optimization module then generates Mixed Reality interfaces that account for these considerations as well as temporal constraints The evaluation of SituationAdapt is two-fold: We first validate our reasoning component’s capability in assessing UI contexts comparable to human expert users. In an online user study, we then established our system’s capability of producing context-aware MR layouts, where it outperformed adaptive methods from previous work. We further demonstrate the versatility and applicability of SituationAdapt with a set of application scenarios.

著者
Zhipeng Li
ETH, Zurich, Switzerland
Christoph Gebhardt
ETH Zurich, Zurich, Switzerland
Yves Inglin
ETH Zürich, Zürich, Switzerland
Nicolas Steck
ETH Zürich, Zurich, Switzerland
Paul Streli
ETH, Zurich, Switzerland
Christian Holz
ETH Zürich, Zurich, Switzerland
論文URL

https://doi.org/10.1145/3654777.3676470

動画
Desk2Desk: Optimization-based Mixed Reality Workspace Integration for Remote Side-by-side Collaboration
要旨

Mixed Reality enables hybrid workspaces where physical and virtual monitors are adaptively created and moved to suit the current environment and needs. However, in shared settings, individual users’ workspaces are rarely aligned and can vary significantly in the number of monitors, available physical space, and workspace layout, creating inconsistencies between workspaces which may cause confusion and reduce collaboration. We present Desk2Desk, an optimization-based approach for remote collaboration in which the hybrid workspaces of two collaborators are fully integrated to enable immersive side-by-side collaboration. The optimization adjusts each user’s workspace in layout and number of shared monitors and creates a mapping between workspaces to handle inconsistencies between workspaces due to physical constraints (e.g. physical monitors). We show in a user study how our system adaptively merges dissimilar physical workspaces to enable immersive side-by-side collaboration, and demonstrate how an optimization-based approach can effectively address dissimilar physical layouts.

著者
Ludwig Sidenmark
University of Toronto, Toronto, Ontario, Canada
Tianyu Zhang
University of Toronto, Toronto, Ontario, Canada
Leen Al Lababidi
University of Toronto, Toronto, Ontario, Canada
Jiannan Li
Singapore Management University , Singapore, Singapore
Tovi Grossman
University of Toronto, Toronto, Ontario, Canada
論文URL

https://doi.org/10.1145/3654777.3676339

動画
SpaceBlender: Creating Context-Rich Collaborative Spaces Through Generative 3D Scene Blending
要旨

There is increased interest in using generative AI to create 3D spaces for virtual reality (VR) applications. However, today’s models produce artificial environments, falling short of supporting collaborative tasks that benefit from incorporating the user's physical context. To generate environments that support VR telepresence, we introduce SpaceBlender, a novel pipeline that utilizes generative AI techniques to blend users' physical surroundings into unified virtual spaces. This pipeline transforms user-provided 2D images into context-rich 3D environments through an iterative process consisting of depth estimation, mesh alignment, and diffusion-based space completion guided by geometric priors and adaptive text prompts. In a preliminary within-subjects study, where 20 participants performed a collaborative VR affinity diagramming task in pairs, we compared SpaceBlender with a generic virtual environment and a state-of-the-art scene generation framework, evaluating its ability to create virtual spaces suitable for collaboration. Participants appreciated the enhanced familiarity and context provided by SpaceBlender but also noted complexities in the generative environments that could detract from task focus. Drawing on participant feedback, we propose directions for improving the pipeline and discuss the value and design of blended spaces for different scenarios.

著者
Nels Numan
Microsoft Research, Redmond, Washington, United States
Shwetha Rajaram
Microsoft Research, Redmond, Washington, United States
Bala Kumaravel
Microsoft Research, Redmond, Washington, United States
Nicolai Marquardt
Microsoft Research, Redmond, Washington, United States
Andrew D. Wilson
Microsoft Research, Redmond, Washington, United States
論文URL

https://doi.org/10.1145/3654777.3676361

動画