Papers
arxiv:2410.02055
Using Style Ambiguity Loss to Improve Aesthetics of Diffusion Models
Published on Oct 2
Authors:
Abstract
Teaching text-to-image models to be creative involves using style ambiguity loss. In this work, we explore using the style ambiguity training objective, used to approximate creativity, on a diffusion model. We then experiment with forms of style ambiguity loss that do not require training a classifier or a labeled dataset, and find that the models trained with style ambiguity loss can generate better images than the baseline diffusion models and GANs. Code is available at https://github.com/jamesBaker361/clipcreate.
Models citing this paper 0
No model linking this paper
Cite arxiv.org/abs/2410.02055 in a model README.md to link it from this page.
Datasets citing this paper 0
No dataset linking this paper
Cite arxiv.org/abs/2410.02055 in a dataset README.md to link it from this page.
Spaces citing this paper 0
No Space linking this paper
Cite arxiv.org/abs/2410.02055 in a Space README.md to link it from this page.