# SoteDiffusion V2
An anime finetune of Würstchen V3 / Stable Cascade.
# Release Notes
- This release is sponsored by fal.ai/grants
- Trained on 12M text & image paris including WD tags and natural language captions for a single epoch on 8xH100 80GB SXM5 GPUs.
- Trained with Full FP32 and MAE Loss.
# ComfyUI
Use these arguments when starting ComfyUI: --fp16-vae --fp16-unet
Download the Stage C to unet folder: https://civarchive.com/api/download/models/703072?type=Model&format=SafeTensor&size=pruned&fp=fp16
Download the Stage C Text Encoder to clip folder: https://civarchive.com/api/download/models/704783?type=Model&format=SafeTensor&size=pruned&fp=fp16
Download the Stage B to unet folder: https://civarchive.com/api/download/models/704646?type=Model&format=SafeTensor&size=pruned&fp=fp16
Download the Stage A to vae folder: https://huggingface.co/madebyollin/stage-a-ft-hq/resolve/main/stage_a_ft_hq.safetensors
Download the workflow and load it: https://huggingface.co/Disty0/sotediffusion-v2/resolve/main/comfyui_workflow.json?download=true
Stage C sampler: DPMPP 2M or DPMPP 2M SDE with SGM Uniform scheduler
Stage C steps: 28
Stage C CFG: 6.0
Stage B sampler: LCM with Exponential scheduler
Stage B steps: 14
Stage B CFG: 1.0
## SD.Next
URL: https://github.com/vladmandic/automatic/
Go to Models -> Huggingface and type Disty0/sotediffusion-v2 into the model name and press download.
Load Disty0/sotediffusion-v2 after the download process is complete.
Prompt:
```
your prompt goes here
very aesthetic, best quality, newest,
```
(New lines act the same way as BREAK in SD.Next)
Negative Prompt:
```
very displeasing, displeasing, worst quality, bad quality, low quality, realistic, monochrome, comic, sketch, oldest, early, artist name, signature, blurry, simple background, upside down,
```
Parameters:
Sampler: Default
Steps: 28
Refiner Steps: 14
CFG: 5.0 to 6.0
Secondary CFG: 1.0 to 1.5
Resolution: 1280x1280, 1024x1536, 1024x2048, 2048x1152
Anything works as long as it's a multiply of 128.
## License
SoteDiffusion models falls under [Fair AI Public License 1.0-SD](https://freedevproject.org/faipl-1.0-sd/) license, which is compatible with Stable Diffusion models’ license. Key points:
1. Modification Sharing: If you modify SoteDiffusion models, you must share both your changes and the original license.
2. Source Code Accessibility: If your modified version is network-accessible, provide a way (like a download link) for others to get the source code. This applies to derived models too.
3. Distribution Terms: Any distribution must be under this license or another with similar rules.
4. Compliance: Non-compliance must be fixed within 30 days to avoid license termination, emphasizing transparency and adherence to open-source values.
Notes: Anything not covered by Fair AI license is inherited from Stability AI Non-Commercial license.
Description
Trained with Full FP32 and MAE Loss.
FAQ
Comments (6)
Very good results considering it was only one epoch.
there are a character/artist list?
Could you please convert your models to "single files" format (separate stages, of course), as in Stable Cascade by Stability.AI (single files) - Stage B (for ComfyUI)?
I have a problem using your provided workflow in SwarmUI, but "single model" versions of Stable Cascade can be just loaded and used as normal checkpoints.
No. Explained why in here already: https://huggingface.co/Disty0/sotediffusion-v2/discussions/1
Basically, checkpoints doesn't work for anything other than ComfyUI and there is no clear way to revert them into UNets.
The model have "middle style" with a lot of agressive gradients by default - it have a good form but coloring doesn't look good enough. Is it difficult to teach it for more anime-like coloring style by default (sharp shadows, softer gradients)?
From what I know, style dataset should be very diverse and all images should have similar style along with it. Also it's recommended to use downscaled big resolution images instead of default lowres or middle resolution to avoid artifacts and blurry lines in output.
The ComfyUI workflow i provided uses DPMPP 2M SDE GPU for the Stage C. You probably want DPMPP 2M instead.
Edit: Now updated the ComfyUI workflow to use DPMPP 2M as well since i decided for it to be the default on Diffusers.
Details
Available On (1 platform)
Same model published on other platforms. May have additional downloads or version variants.

