stable-diffusion stable-diffusion-diffusers text-to-image diffusers

Text-to-image finetuning - ohicarip/sd-deepfashion-baseline-model

This pipeline was finetuned from CompVis/stable-diffusion-v1-4 on the ohicarip/deepfashion_bl2 dataset. Below are some example images generated with the finetuned pipeline using the following prompts: ['This man wears a long-sleeve sweater with pure color patterns. The sweater is with cotton fabric. It has a round neckline. The pants this man wears is of long length. The pants are with denim fabric and solid color patterns. The outer clothing the gentleman wears is with cotton fabric and solid color patterns. There is an accessory on his wrist.', 'This person is wearing a short-sleeve shirt with pure color patterns. The shirt is with cotton fabric. It has a round neckline. This person wears a long trousers. The trousers are with denim fabric and lattice patterns.', 'This guy is wearing a short-sleeve shirt with solid color patterns and a long pants. The shirt is with cotton fabric and its neckline is crew. The pants are with denim fabric and solid color patterns.', 'This female is wearing a tank tank shirt with plaid patterns and a three-point shorts. The tank shirt is with cotton fabric. The neckline of the tank shirt is crew. The shorts are with cotton fabric and plaid patterns. This lady wears socks in shoes.']:

val_imgs_grid

Pipeline usage

You can use the pipeline like so:

from diffusers import DiffusionPipeline
import torch

pipeline = DiffusionPipeline.from_pretrained("ohicarip/sd-deepfashion-baseline-model", torch_dtype=torch.float16)
prompt = "This man wears a long-sleeve sweater with pure color patterns. The sweater is with cotton fabric. It has a round neckline. The pants this man wears is of long length. The pants are with denim fabric and solid color patterns. The outer clothing the gentleman wears is with cotton fabric and solid color patterns. There is an accessory on his wrist."
image = pipeline(prompt).images[0]
image.save("my_image.png")

Training info

These are the key hyperparameters used during training:

More information on all the CLI arguments and the environment are available on your wandb run page.