coreml stable-diffusion text-to-image

Core ML converted model

This model was converted to Core ML for use on Apple Silicon devices by following Apple's instructions here.
Provide the model to an app such as Mochi Diffusion to generate images.

split_einsum versions are compatible with all compute units.
original versions are only compatible with CPU & GPU.

3DKX V2

Sources: Hugging Face - CivitAI

<img width="604px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076941914538004581/ezgif-3-7ac098980a.gif"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932535445098626/1.png"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932536120385657/2.png"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932536795664405/3.png"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932537416433835/4.png"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932538041372782/5.png"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932538783768626/6.png"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932539664580688/7.png"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932540327268413/8.png"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932541052895292/9.png"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932541614923818/10.png"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932544534151178/11.png"> <img width="768px" src="https://cdn.discordapp.com/attachments/1051410188592226364/1076932545117179924/12.png">

Model Description

<!-- Provide a longer summary of what this model is. -->

3DKX V2 is a model that was trained on highly detailed 3D rendered pictures of various subjects such as landscapes, scenes, models, textures, and more. Our aim is to provide a useful tool that can produce consistent and high-resolution renders for creative purposes such as storyboarding, sketching, templates, wallpapers, and more.

Uses

<!-- 3DKX is a model that was trained on highly detailed 3D rendered pictures of various subjects such as landscapes, scenes, models, textures, and more. Our aim is to provide a useful tool that can produce consistent and high-resolution renders for creative purposes such as storyboarding, sketching, templates, wallpapers, and more. -->

Get started with the model, use our cheat sheet !

Use the guide in the link below to get started with the model ! 3DKX_V2 Presentation/Guide

License & User Restrictions

You agree not to use the Model or Derivatives of the Model:

Important notes:

Training Details

Base Model: SD 1.5 Steps: 15,000 Training Method: Finetuning Trigger Keywords: None really, but the main styles trained are "3d render" and "3d cartoon" see our presentation slide for more details.

<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> Big shoutout to this new code that allowed us to fix the diffusion noise, allowing for more depth, contrast, and white level balance in the outputs. https://www.crosslabs.org//blog/diffusion-with-offset-noise