OPAL: Multimodal Image Generation for News Illustrations

要旨

Advances in multimodal AI have presented people with powerful ways to create images from text. Recent work has shown that text-to-image generations are able to represent a broad range of subjects and artistic styles. However, finding the right visual language for text prompts is difficult. In this paper, we address this challenge with Opal, a system that produces text-to-image generations for news illustration. Given an article, Opal guides users through a structured search for visual concepts and provides a pipeline allowing users to generate illustrations based on an article's tone, keywords, and related artistic styles. Our evaluation shows that Opal efficiently generates diverse sets of news illustrations, visual assets, and concept ideas. Users with Opal generated two times more usable results than users without. We discuss how structured exploration can help users better understand the capabilities of human AI co-creative systems.

著者
Vivian Liu
Columbia University, New York, New York, United States
Han Qiao
Columbia University, New York, New York, United States
Lydia B. Chilton
Columbia University, New York, New York, United States
論文URL

https://doi.org/10.1145/3526113.3545621

会議: UIST 2022

The ACM Symposium on User Interface Software and Technology

セッション: Generative Design

6 件の発表
2022-11-02 18:00:00
2022-11-02 19:30:00