Generative Design

会議の名前
UIST 2022
OPAL: Multimodal Image Generation for News Illustrations
要旨

Advances in multimodal AI have presented people with powerful ways to create images from text. Recent work has shown that text-to-image generations are able to represent a broad range of subjects and artistic styles. However, finding the right visual language for text prompts is difficult. In this paper, we address this challenge with Opal, a system that produces text-to-image generations for news illustration. Given an article, Opal guides users through a structured search for visual concepts and provides a pipeline allowing users to generate illustrations based on an article's tone, keywords, and related artistic styles. Our evaluation shows that Opal efficiently generates diverse sets of news illustrations, visual assets, and concept ideas. Users with Opal generated two times more usable results than users without. We discuss how structured exploration can help users better understand the capabilities of human AI co-creative systems.

著者
Vivian Liu
Columbia University, New York, New York, United States
Han Qiao
Columbia University, New York, New York, United States
Lydia B. Chilton
Columbia University, New York, New York, United States
論文URL

https://doi.org/10.1145/3526113.3545621

Social Simulacra: Creating Populated Prototypes for Social Computing Systems
要旨

Social computing prototypes probe the social behaviors that may arise in an envisioned system design. This prototyping practice is currently limited to recruiting small groups of people. Unfortunately, many challenges do not arise until a system is populated at a larger scale. Can a designer understand how a social system might behave when populated, and make adjustments to the design before the system falls prey to such challenges? We introduce social simulacra, a prototyping technique that generates a breadth of realistic social interactions that may emerge when a social computing system is populated. Social simulacra take as input the designer's description of a community’s design---goal, rules, and member personas---and produce as output an instance of that design with simulated behavior, including posts, replies, and anti-social behaviors. We demonstrate that social simulacra shift the behaviors that they generate appropriately in response to design changes, and that they enable exploration of "what if?" scenarios where community members or moderators intervene. To power social simulacra, we contribute techniques for prompting a large language model to generate thousands of distinct community members and their social interactions with each other; these techniques are enabled by the observation that large language models' training data already includes a wide variety of positive and negative behavior on social media platforms. In evaluations, we show that participants are often unable to distinguish social simulacra from actual community behavior and that social computing designers successfully refine their social computing designs when using social simulacra.

著者
Joon Sung Park
Stanford University, Palo Alto, California, United States
Lindsay Popowski
Stanford University, Stanford, California, United States
Carrie J. Cai
Google, Mountain View, California, United States
Meredith Ringel. Morris
Google Research, Seattle, Washington, United States
Percy Liang
Stanford University, Stanford, California, United States
Michael S.. Bernstein
Stanford University, Stanford, California, United States
論文URL

https://doi.org/10.1145/3526113.3545616

GANzilla: User-Driven Direction Discovery in Generative Adversarial Networks
要旨

Generative Adversarial Network (GAN) is being widely adopted in numerous application areas, such as data preprocessing, image editing, and creativity support. However, GAN's 'black box' nature prevents non-expert users from controlling what data a model generates, spawning a plethora of prior work that focused on algorithm-driven approaches to automatically extract editing directions to control GAN. Complementarily, we propose a GANzilla---a user-driven tool that empowers a user with the classic scatter/gather technique to iteratively discover directions to meet their editing intents. In a work session with 12 participants, GANzilla users were able to discover directions that (i) edited images to match provided examples (closed-ended tasks) and that (ii) met a high-level goal, e.g., making the face happier, while showing diversity across individuals (open-ended tasks).

著者
Noyan Evirgen
University of California Los Angeles, Los Angeles, California, United States
Xiang 'Anthony' Chen
UCLA, Los Angeles, California, United States
論文URL

https://doi.org/10.1145/3526113.3545638

We-toon: A Communication Support System between Writers and Artists in Collaborative Webtoon Sketch Revision
要旨

We present a communication support system, namely \textit{We-toon}, that can bridge the webtoon writers and artists during sketch revision (i.e., character design and draft revision). In the highly iterative design process between the webtoon writers and artists, writers often have difficulties in precisely articulating their feedback on sketches owing to their lack of drawing proficiency. This drawback makes the writers rely on textual descriptions and reference images found using search engines, leading to indirect and inefficient communications. Inspired by a formative study, we designed \textit{We-toon} to help writers revise webtoon sketches and effectively communicate with artists. Through a GAN-based image synthesis and manipulation, \textit{We-toon} can interactively generate diverse reference images and synthesize them locally on any user-provided image. Our user study with 24 professional webtoon authors demonstrated that \textit{We-toon} outperforms the traditional methods in terms of communication effectiveness and the writers' satisfaction level related to the revised image.

著者
Hyung-Kwon Ko
Webtoon AI, NAVER WEBTOON Corp., Seongnam, Gyeonggi, Korea, Republic of
Subin An
Seoul National University, Seoul, Korea, Republic of
Gwanmo Park
Seoul National University, Seoul, Korea, Republic of
Seung Kwon Kim
NAVER WEBTOON, SEOUL, Korea, Republic of
Daesik Kim
Naver Webtoon Ltd, SeongNam, Korea, Republic of
Bohyoung Kim
Hankuk University of Foreign Studies, Yongin-si, Korea, Republic of
Jaemin Jo
Sungkyunkwan University, Suwon, Korea, Republic of
Jinwook Seo
Seoul National University, Seoul, Korea, Republic of
論文URL

https://doi.org/10.1145/3526113.3545612

BO as Assistant: Using Bayesian Optimization for Asynchronously Generating Design Suggestions
要旨

Many design tasks involve parameter adjustment, and designers often struggle to find desirable parameter value combinations by manipulating sliders back and forth. For such a multi-dimensional search problem, Bayesian optimization (BO) is a promising technique because of its intelligent sampling strategy; in each iteration, BO samples the most effective points considering both exploration (i.e., prioritizing unexplored regions) and exploitation (i.e., prioritizing promising regions), enabling efficient searches. However, existing BO-based design frameworks take the initiative in the design process and thus are not flexible enough for designers to freely explore the design space using their domain knowledge. In this paper, we propose a novel design framework, BO as Assistant, which enables designers to take the initiative in the design process while also benefiting from BO's sampling strategy. The designer can manipulate sliders as usual; the system monitors the slider manipulation to automatically estimate the design goal on the fly and then asynchronously provides unexplored-yet-promising suggestions using BO's sampling strategy. The designer can choose to use the suggestions at any time. This framework uses a novel technique to automatically extract the necessary information to run BO by observing slider manipulation without requesting additional inputs. Our framework is domain-agnostic, demonstrated by applying it to photo color enhancement, 3D shape design for personal fabrication, and procedural material design in computer graphics.

著者
Yuki Koyama
National Institute of Advanced Industrial Science and Technology (AIST), Tsukuba, Japan
Masataka Goto
National Institute of Advanced Industrial Science and Technology (AIST), Tsukuba, Japan
論文URL

https://doi.org/10.1145/3526113.3545664

Chatbots Facilitating Consensus-building in Asynchronous Co-Design
要旨

Consensus-building is an essential process for the success of co-design projects. To build consensus, stakeholders need to discuss conflicting needs and viewpoints, converge their ideas toward shared interests, and grow their willingness to commit to group decisions. However, managing group discussions is challenging in large co-design projects with multiple stakeholders. In this paper, we investigate the interaction design of a chatbot that can mediate consensus-building conversationally. By interacting with individual stakeholders, the chatbot collects ideas for satisfying conflicting needs and engages stakeholders to consider others' viewpoints, without having stakeholders directly interact with each other. Results from an empirical study in an educational setting (N = 12) suggest that the approach can increase stakeholders' commitment to group decisions and maintain the effect even on the group decisions that conflict with personal interests. We conclude that chatbots can facilitate consensus-building in small-to-medium-sized projects, but more work is needed to scale up to larger projects.

著者
Joon Gi Shin
Aalto University, Espoo, Finland
Michael A.. Hedderich
Saarland Informatics Campus, Saarbrücken , Germany
Andrés Lucero
Aalto University, Espoo, Finland
Antti Oulasvirta
Aalto University, Helsinki, Finland
論文URL

https://doi.org/10.1145/3526113.3545671