Hey guys,
I was trying SDXL 1.0 but my laptop with a RTX 3050 Laptop 4GB vRAM was not able to generate in less than 3 minutes, so I spent some time to get a good configuration in ComfyUI, now I get can generate in 55s (batch images) - 70s (new prompt detected) getting a great images after the refiner kicks in.
Is the best balanced I could find between image size (1024x720), models, steps (10+5 refiner), samplers/schedulers, so we can use SDXL on our laptops without those expensive/bulky desktop GPUs.
I wanted to share my configuration for ComfyUI, since many of us are using our laptops most of the time. I think this is the best balanced I could find.
Add params in "run_nvidia_gpu.bat" --normalvram --fp16-vae
Face fix fast version?:
SDXL has many problems for faces when the face is away from the "camera" (small faces), so this version fixes faces detected and takes 5 extra steps only for the face.
Face fix no fast version?:
For fix face (no fast version), faces will be fixed after the upscaler, better results, specially for very small faces, but adds 20 seconds compared to fast version.
If the face fix output does not generate a different image (maybe you are using a 4x upscaler), and console prints "segment skip [determined upscale factor=0.9875809267424535]" , in module "FaceDetailer" increase "guide_size" from 1280 to 1408 or more until it activates the FaceDetailer.
Difference between LORA and LORA fast?:
If you choose LORA fast, you can save 20-30 seconds.
The LORA fast does not have 3 extra steps after refiner to retouch the LORA effect, refiner dims the effect of the LORAs, in some cases for LORA with custom styles is needed the last 3 steps to add again the effect. Not needed in most cases, recommended for LORAs with custom styles were styles changes a lot the image.
Generation time (after first image):
No LORA: 55-60 seconds
With LORA: 85-115 seconds
With LORA Fast: 75-80 seconds
With face fix: 80 seconds (fix faces FAST) - 110 seconds (Fix faces slow version)
Files to Download:
Refiner SDXL 1.0: https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0
Model Dreamshaper SDXL 1.0 (or any other): https://civarchive.com/models/112902/dreamshaper-xl10
Fixed SDXL VAE 16FP: https://huggingface.co/madebyollin/sdxl-vae-fp16-fix (config.json, diffusion_pytorch_model.safetensors and sdxl_vae.safetensors files inside the same folder under VAE/<NEW_FOLDER>).
Upscale model (RealESRGAN or Swin2SR):
https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.1/RealESRGAN_x2plus.pth
https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_Lightweight_X2_64.pthFor the LORA version: https://civarchive.com/models/117635/greg-rutkowski-style-lora-sdxl
For face fix version (FaceDetailer):
ComfyUI/models/sams: https://dl.fbaipublicfiles.com/segment_anything/sam_vit_b_01ec64.pth

Description
Finally a workflow with LCM using the LCM-LORA.
All base models compatible.
Great quality in 10 steps! (7 base model + 3 refined with 1.5x upscaled and contrast fix).
No sdxl_refiner model needed (uses same base_model to refine).
Final refined image will be scaled 1.5x with a contrast and quality trick.
Don´t forget to download the LCM-LORA and select it in the "Enable LCM LORA" group, you can chain more loras below in the "More LORAs" group.
In case of using a base model type LCM, you have to bypass the group "Enable LCM LORA" by selecting the 2 nodes and CTRL+B.
FAQ
Comments (14)
LCM version errors:
lora key not loaded
Wrong version pytorch_lora_weights?
If your model is a LCM model, you have to disable the 2 nodes in the ENABLE LCM box.
Thanks I really enjoyed this. I've added control net to your workflow if you interested.
Could you drop the control-net workflow with this
Thanks for the workflow. But adding --fp16-vae gives black images.
Are you using the external 16bit VAE?
It must be loaded in the "load VAE" node.
still gives black images.
How can I use the sd1.5 model as a refiner for the basic sdxl model? Like Fooocus
You cannot.
It said missing Integer node, what is it? I can't find it to install
same error, got any solution?
@DangerVirat1767 Same here T_T
Did you find any solution?
It was a deprecated comfy node, I fixed it in v4.





