AI for Visual Generation

会議の名前
CHI 2023
GANravel: User-Driven Direction Disentanglement in Generative Adversarial Networks
要旨

Generative adversarial networks (GANs) have many application areas including image editing, domain translation, missing data imputation, and support for creative work. However, GANs are considered `black boxes'. Specifically, the end-users have little control over how to improve editing directions through disentanglement. Prior work focused on new GAN architectures to disentangle editing directions. Alternatively, we propose GANravel --a user-driven direction disentanglement tool that complements the existing GAN architectures and allows users to improve editing directions iteratively. In two user studies with 16 participants each, GANravel users were able to disentangle directions and outperformed the state-of-the-art direction discovery baselines in disentanglement performance. In the second user study, GANravel was used in a creative task of creating dog memes and was able to create high-quality edited images and GIFs.

著者
Noyan Evirgen
HCI Group, Los Angeles, California, United States
Xiang 'Anthony' Chen
UCLA, Los Angeles, California, United States
論文URL

https://doi.org/10.1145/3544548.3581226

動画
The Elements of Visual Art Recommendation: Learning Latent Semantic Representations of Paintings
要旨

Artwork recommendation is challenging because it requires understanding how users interact with highly subjective content, the complexity of the concepts embedded within the artwork, and the emotional and cognitive reflections they may trigger in users. In this paper, we focus on efficiently capturing the elements (i.e., latent semantic relationships) of visual art for personalized recommendation. We propose and study recommender systems based on textual and visual feature learning techniques, as well as their combinations. We then perform a small-scale and a large-scale user-centric evaluation of the quality of the recommendations. Our results indicate that textual features compare favourably with visual ones, whereas a fusion of both captures the most suitable hidden semantic relationships for artwork recommendation. Ultimately, this paper contributes to our understanding of how to deliver content that suitably matches the user's interests and how they are perceived.

著者
Bereket A.. YILMA
University of Luxembourg, Luxembourg, ESCH/ALZETTE, Luxembourg
Luis A.. Leiva
University of Luxembourg, Esch-sur-Alzette, Luxembourg
論文URL

https://doi.org/10.1145/3544548.3581477

動画
StyleMe: Towards Intelligent Fashion Generation with Designer Style
要旨

Hand-drawn sketches and sketch colourization are the most laborious but necessary steps for fashion designers to design exquisite clothes, especially when the fashion design requires distinctive and personal characteristics from designer style. This paper presents an artificial intelligent aided fashion design system, namely StyleMe, to support the automatic generation of clothing sketches with designer style. Given the clothing pictures specified by the designer, StyleMe can use deep learning based generative model to generate clothing sketches that are consistent with the designer style. The system also supports intelligent colourization on clothing sketch by style transfer, according to specified styles from the real fashion images. Through a series of performance evaluations and user studies, we found that our system can generate effective clothing sketches as good as fashion designers' human work, and significantly improve the efficiency of fashion design with its sketch colourization method.

著者
Di Wu
Hunan University, Changsha, Hunan, China
Zhiwang Yu
Hunan University, Changsha, Hunan, China
Nan Ma
Hunan University, Changsha city, China
Jianan Jiang
Hunan University, Changsha, China
Yuetian Wang
Hunan University, Changsha city, China
Guixiang Zhou
Hunan University, Changsha city, China
Hanhui Deng
Hunan University, Changsha, Hunan, China
Yi Li
Hunan University, Changsha city, China
論文URL

https://doi.org/10.1145/3544548.3581377

動画
InfinitePaint: Painting in Virtual Reality with Passive Haptics Using Wet Brushes and a Physical Proxy Canvas
要旨

Digital painting interfaces require an input fidelity that preserves the artistic expression of the user. Drawing tablets allow for precise and low-latency sensing of pen motions and other parameters like pressure to convert them to fully digitized strokes. A drawback is that those interfaces are rigid. While soft brushes can be simulated in software, the haptic sensation of the rigid pen input device is different compared to using a soft wet brush on paper. We present InfinitePaint, a system that supports digital painting in Virtual Reality on real paper with a real wet brush. We use special paper that turns black wherever it comes into contact with water and turns blank again upon drying. A single camera captures those temporary strokes and digitizes them while applying properties like color or other digital effects. We tested our system with artists and compared the subjective experience with a drawing tablet.

著者
Andreas Rene. Fender
ETH Zürich, Zurich, Switzerland
Thomas Roberts
ETH Zürich, Zurich, Switzerland
Tiffany Luong
ETH Zürich, Zürich, Switzerland
Christian Holz
ETH Zürich, Zurich, Switzerland
論文URL

https://doi.org/10.1145/3544548.3580927

動画
I think I don't feel sick: Exploring the Relationship Between Cognitive Demand and Cybersickness in Virtual Reality using fNIRS
要旨

Virtual Reality (VR) applications commonly use the illusion of self-motion (vection) to simulate experiences such as running, driving, or flying. However, this can lead to cybersickness, which diminishes the experience of users, and can even lead to disengagement with this platform. In this paper we present a study in which we show that users performing a cognitive task while experiencing a VR rollercoaster reported reduced symptoms of cybersickness. Furthermore, we collected and analysed brain activity data from our participants during their experience using functional near infra-red spectroscopy (fNIRS): \hl{preliminary analysis suggests the possibility} that this technology may be able to detect the experience of cybersickness. Together, these results can assist the creators of VR experiences, both through mitigation of cybersickness in the design process, and by better understanding the experiences of their users.

著者
Katharina Margareta Theresa. Pöhlmann
University of Glasgow, Glasgow, United Kingdom
Horia A.. Maior
University of Nottingham, Nottingham, Nottinghamshire, United Kingdom
Julia Föcker
University of Lincoln, Lincoln, United Kingdom
Louise O'Hare
Nottingham Trent University, Nottingham, United Kingdom
Adrian Parke
University of the West of Scotland, Glasgow, United Kingdom
Aleksandra Ladowska
University of Nottingham, Nottingham, United Kingdom
Patrick Dickinson
University of Lincoln, Lincoln, United Kingdom
論文URL

https://doi.org/10.1145/3544548.3581063

動画
RePrompt: Automatic Prompt Editing to Refine AI-Generative Art Towards Precise Expressions
要旨

Generative AI models have shown impressive ability to produce images with text prompts, which could benefit creativity in visual art creation and self-expression. However, it is unclear how precisely the generated images express contexts and emotions from the input texts. We explored the emotional expressiveness of AI-generated images and developed RePrompt, an automatic method to refine text prompts toward precise expression of the generated images. Inspired by crowdsourced editing strategies, we curated intuitive text features, such as the number and concreteness of nouns, and trained a proxy model to analyze the feature effects on the AI-generated image. With model explanations of the proxy model, we curated a rubric to adjust text prompts to optimize image generation for precise emotion expression. We conducted simulation and user studies, which showed that RePrompt significantly improves the emotional expressiveness of AI-generated images, especially for negative emotions.

著者
Yunlong Wang
National University of Singapore, Singapore, Singapore
Shuyuan Shen
National University of Singapore, Singapore, Singapore
Brian Y. Lim
National University of Singapore, Singapore, Singapore
論文URL

https://doi.org/10.1145/3544548.3581402

動画