coreml stable-diffusion text-to-image

Core ML Converted Model:

<br>

epiCRealism-pureEvolution-V3_cn:

Source(s): CivitAI<br>

V3 is here!

Since SDXL is right around the corner, let's say it is the final version for now since I put a lot effort into it and probably cannot do much more.

I tried to refine the understanding of the Prompts, Hands and of course the Realism. Let's see what you guys can do with it.

Thanks to @drawaline for the in-depth review, so i'd like to give some advices to use this model.

Advices

Use simple prompts

No need to use keywords like "masterpiece, best quality, 8k, intricate, high detail" or "(extremely detailed face), (extremely detailed hands), (extremely detailed hair)" since it doesn't produce appreciable change

Use simple negatives or small negative embeddings. gives most realistic look (check samples to get an idea of negatives i used)

Add "asian, chinese" to negative if you're looking for ethnicities other than Asian

Light, shadows, and details are excellent without extra keywords

If you're looking for a natural effect, avoid "cinematic"

Avoid using "1girl" since it pushes things to render/anime style

To much description of the face will turn out bad mostly

For a more fantasy like output use 2M Karras Sampler

No extra noise-offset needed, but u can if you like to 😉

How to use?

Prompt: simple explanation of the image (try first without extra keywords)

Negative: "cartoon, painting, illustration, (worst quality, low quality, normal quality:2)"

Steps: >20 (if image has errors or artefacts use higher Steps)

CFG Scale: 5 (higher config scale can lose realism, depends on prompt, sampler and Steps)

Sampler: Any Sampler (SDE, DPM-Sampler will result in more realism)

Size: 512x768 or 768x512

Hires upscaler: 4x_NMKD-Superscale-SP_178000_G (Denoising: 0.35, Upscale: 2x)<br><br>

image

image

image

image