CivArchive
    Preview 111907554
    Preview 111907569
    Preview 111907559
    Preview 111907565
    Preview 111907564
    Preview 111907563
    Preview 111907562
    Preview 111907577
    Preview 111907568
    Preview 111907566
    Preview 111907560
    Preview 111907576
    Preview 111907575
    Preview 111907572
    Preview 111907578
    Preview 111907573
    Preview 111907558
    Preview 111907556
    Preview 111907574

    🚀 Z-Image AIO Collection

    ⚡ Base & Turbo • All-in-One • Bilingual Text • Qwen3-4B


    ⚠️ IMPORTANT: Requires ComfyUI v0.11.0+

    📥 Download ComfyUI


    ✨ What is Z-Image AIO?

    Z-Image AIO is an All-in-One repackage of Alibaba Tongyi Lab's 6B parameter image generation models.

    Everything integrated:

    • ✅ VAE already built-in

    • ✅ Qwen3-4B Text Encoder integrated

    • ✅ Just download and generate!


    🎯 Available Versions


    🔥 Z-Image-Turbo-AIO (8 Steps • CFG 1.0)

    Ultra-fast generation for production & daily use


    ⚫ NVFP4-AIO (7.8 GB) 🆕

    🎯 ONLY for NVIDIA Blackwell GPUs (RTX 50xx)!
    ⚡ Maximum speed optimized
    💾 Smallest file size
    🚀 FP4 precision - blazing fast
    

    Perfect for: RTX 5070, 5080, 5090 owners who want maximum speed


    ✅ Best balance of size & quality
    ✅ Works on 8GB VRAM
    ✅ Fast downloads
    ✅ Ideal for most users
    

    Perfect for: Daily use, testing, RTX 3060/4060/4070


    🔵 FP16-AIO (20 GB)

    💾 Same file size as BF16
    🔄 ComfyUI auto-casts to BF16 for compute
    ⚠️ Does NOT enable FP16 compute mode
    📦 Alternative download option
    

    Note: Z-Image does not support FP16 compute - activation values exceed FP16's max range, causing NaN/black images. Weights are cast to BF16 during inference regardless of file format.

    Perfect for: Alternative to BF16 download (identical inference behavior)


    ✅ BFloat16 full precision
    ✅ Absolute best quality
    ✅ Professional projects
    ✅ Also works on 8GB VRAM
    

    Perfect for: Professional work, maximum quality


    🎨 Z-Image-Base-AIO (28-50 Steps • CFG 3-5)

    Full creative control for pros & LoRA training


    🟡 FP8-AIO (10 GB)

    ✅ Efficient for daily use
    ✅ Full CFG control
    ✅ Negative prompts supported
    ✅ 8GB VRAM compatible
    

    Perfect for: Daily work with full control


    🔵 FP16-AIO (20 GB)

    💾 Same file size as BF16
    🔄 ComfyUI auto-casts to BF16 for compute
    ⚠️ Does NOT enable FP16 compute mode
    📦 Alternative download option
    

    Note: See technical explanation in FAQ below.

    Perfect for: Alternative to BF16 download (identical inference behavior)


    ✅ Maximum quality
    ✅ Ideal for LoRA training
    ✅ Professional projects
    ✅ Highest precision
    

    Perfect for: LoRA training, professional work


    🆚 Turbo vs Base - When to Use?


    ⚡ Use TURBO when:

    ⚡ Speed is priority → 8 steps = 3-10 seconds
    📸 Production workflows → Consistent high quality
    💾 Quick iterations → Rapid prototyping
    🎯 Simple prompts → Less complex scenes
    

    🎨 Use BASE when:

    🎨 Creative exploration → Higher diversity
    🔧 LoRA/ControlNet dev → Undistilled foundation
    📝 Complex prompting → Full CFG control
    🚫 Negative prompts needed → Remove unwanted elements
    

    ⚙️ Recommended Settings


    ⚡ Turbo Settings (incl. NVFP4)

    📊 Steps: 8
    🎚️ CFG: 1.0 (don't change!)
    🎲 Sampler: res_multistep OR euler_ancestral
    📈 Scheduler: simple OR beta
    📐 Resolution: 1920×1088 (recommended)
    🚫 Negative Prompt: ❌ Not used!
    

    🎨 Base Settings

    📊 Steps: 28-50
    🎚️ CFG: 3.0-5.0 (start with 4.0)
    🎲 Sampler: euler ⭐ OR dpmpp_2m
    📈 Scheduler: normal ⭐ OR karras
    📐 Resolution: 512×512 to 2048×2048
    🚫 Negative Prompt: ✅ Fully supported!
    

    📊 Quick Overview


    Turbo Versions

    ⚫ NVFP4  │ 7.8 GB  │ RTX 50xx only  │ Max Speed 🆕
    🟡 FP8   │ 10 GB   │ 8GB VRAM       │ Recommended ⭐
    🔵 FP16  │ 20 GB   │ → BF16 compute │ See FAQ ⚠️
    🌟 BF16  │ 20 GB   │ 8GB VRAM       │ Max Quality ⭐
    

    Base Versions

    🟡 FP8   │ 10 GB   │ 8GB VRAM       │ Efficient
    🔵 FP16  │ 20 GB   │ → BF16 compute │ See FAQ ⚠️
    🌟 BF16  │ 20 GB   │ 8GB VRAM       │ LoRA Training ⭐
    

    💡 Prompting Guide


    ✅ Good Example:

    Professional food photography of artisan breakfast plate. 
    Golden poached eggs on sourdough toast, crispy bacon, fresh 
    avocado slices. Morning sunlight creating warm glow. Shallow 
    depth of field, magazine-quality presentation.
    

    ❌ Bad Example:

    breakfast, eggs, bacon, toast, food, morning, plate
    

    📝 Tips

    DO:

    • ✅ Use natural language

    • ✅ Be detailed (100-300 words)

    • ✅ Describe lighting & mood

    • ✅ Specify camera angle

    • ✅ English OR Chinese (or both!)

    DON'T:

    • ❌ Tag-style prompts (tag1, tag2, tag3)

    • ❌ Very short prompts (under 50 words)

    • ❌ Negative prompts with Turbo


    🌐 Bilingual Text Rendering


    English:

    Neon sign reading "OPEN 24/7" in bright blue letters 
    above entrance. Modern sans-serif font, glowing effect.
    

    中文:

    Traditional tea house entrance with sign reading 
    "古韵茶坊" in elegant gold Chinese calligraphy.
    

    Both:

    Modern cafe with bilingual sign. "Morning Brew" in 
    white script above, "晨曦咖啡" in Chinese below.
    

    📥 Installation


    Step 1: Download

    Choose your version based on:

    • GPU: RTX 50xx → NVFP4 possible

    • VRAM: 8GB → FP8 recommended

    • Purpose: LoRA Training → Base BF16


    Step 2: Place File

    ComfyUI/models/checkpoints/
    └── Z-Image-Turbo-FP8-AIO.safetensors
    

    Step 3: Load & Generate

    1. Open ComfyUI (v0.11.0+!)

    2. Use "Load Checkpoint" node

    3. Select your AIO version

    4. Generate!

    No separate VAE or Text Encoder needed!


    🙏 Credits


    Original Model

    👨‍💻 Developer: Tongyi Lab (Alibaba Group)
    🏗️ Architecture: Single-Stream DiT (6B parameters)
    📜 License: Apache 2.0
    

    🔗 Z-Image Base: https://huggingface.co/Tongyi-MAI/Z-Image

    🔗 Z-Image Turbo: https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

    🧠 Text Encoder: https://huggingface.co/Qwen/Qwen3-4B


    📈 Version History


    v2.2 - FP16 Clarification

    📝 Updated FP16 descriptions for technical accuracy
    ⚠️ Clarified: FP16 weights ≠ FP16 compute
    🔄 FP16 files are cast to BF16 during inference
    

    v2.1 - NVFP4 Release 🆕

    ➕ Z-Image-Turbo-NVFP4-AIO (7.8GB)
    ⚡ Optimized for NVIDIA Blackwell (RTX 50xx)
    🚀 Maximum speed generation
    

    v2.0 - Base AIO Release

    ➕ Z-Image-Base-BF16-AIO
    ➕ Z-Image-Base-FP16-AIO
    ➕ Z-Image-Base-FP8-AIO
    🔄 ComfyUI v0.11.0+ support
    📝 Qwen3-4B Text Encoder
    

    v1.1 - FP16 Added

    ➕ Z-Image-Turbo-FP16-AIO
    🔧 Wider GPU compatibility
    

    v1.0 - Initial Release

    ✅ Z-Image-Turbo-FP8-AIO
    ✅ Z-Image-Turbo-BF16-AIO
    ✅ Integrated VAE + Text Encoder
    

    ❓ FAQ


    Q: Which version should I choose?

    RTX 50xx + Speed → NVFP4 🆕
    Most users       → Turbo FP8 ⭐
    Full precision   → BF16 ⭐
    LoRA Training    → Base BF16
    

    Q: Turbo or Base?

    Fast & simple    → Turbo ⚡
    Full control     → Base 🎨
    

    Q: Will NVFP4 work on my RTX 4090?

    No! NVFP4 is only for RTX 50xx (Blackwell architecture).

    Use FP8 instead for RTX 40xx and older.


    Q: Do I need separate VAE/Text Encoder?

    No! Everything is already integrated.

    Just Load Checkpoint and go!


    Q: Works on 8GB VRAM?

    Yes! All versions work on 8GB VRAM.

    (NVFP4 requires RTX 50xx regardless of VRAM)


    ⚠️ Q: What about FP16 for older GPUs (RTX 2000/3000)?

    Important technical clarification:

    Z-Image does NOT support FP16 compute type. Here's why:

    📊 Technical reason:
    - FP16 max value: ~65,504
    - BF16 max value: ~3.39e+38 (same as FP32)
    - Z-Image's activation values exceed FP16's range
    - Result: Overflow → NaN → Black images
    

    What actually happens:

    • ComfyUI automatically casts weights to BF16 for computation

    • You can see this in logs: "model weight dtype X, manual cast: torch.bfloat16"

    • "Weight dtype" (file format) ≠ "Compute dtype" (actual calculation)

    For RTX 20xx users (no native BF16):

    • BF16 is emulated via FP32 = slower but works

    • There is no way to run Z-Image in true FP16 compute

    • FP8 with CPU offload may be a better option for limited VRAM

    TL;DR: FP16 and BF16 files behave identically during inference. Choose based on download preference, not GPU compatibility.


    🚀 Get Started Now!

    Download → Load Checkpoint → Generate!

    Recommended versions:

    • 🟡 FP8 for most users (best size/quality balance)

    • 🌟 BF16 for maximum quality

    • NVFP4 for RTX 50xx speed

    All versions work on 8GB VRAM


    Happy generating! 🎨

    Description

    null

    FAQ

    Comments (138)

    kubilayanNov 28, 2025· 1 reaction
    CivitAI

    Where is FP8 AIO version?
    Thank you for your effort.

    SeeSeeLP
    Author
    Nov 28, 2025

    Currently uploading, should be finished in 19 minutes.

    SeeSeeLP
    Author
    Nov 28, 2025· 1 reaction

    @kubilayan is online 😁👍

    kubilayanNov 29, 2025· 1 reaction

    @SeeSeeLP Thank you Thank you Thank you :)

    1639992813Nov 28, 2025· 1 reaction
    CivitAI

    Is this the model you trained? Is it different from the original z-image?

    BastardOGNov 28, 2025· 2 reactions

    this is the same model as original it just has the clip and vae baked in so you could load in a checkpoint node not a diffusion node + clip + vae ... trying it right now but the lora keep giving me errors

    SeeSeeLP
    Author
    Nov 28, 2025· 2 reactions

    ✅ Yep, that’s exactly how it is.
    The only exception is the FP8 AIO:

    🔧 The FP8 AIO is actually the BF16 version that I manually downscaled to FP8
    ✨ I did the same process for the text encoder, and from that the FP8 AIO was created.

    1639992813Nov 28, 2025· 9 reactions

    @Bujieog I see, thank you for your hard work. Looking forward to your Lora; I think what's most urgently needed right now is a Lora that can generate perfect pussy and penis. hahah

    SeeSeeLP
    Author
    Nov 28, 2025

    @1639992813 There is an urgent need for action. 😂👍

    Bars0199Nov 28, 2025· 2 reactions
    CivitAI

    FP8 isn’t working — the example workflow throws a size-mismatch error.
    Hopefully it’ll be supported in the future.

    SeeSeeLP
    Author
    Nov 28, 2025

    @Bars0199 For you 😊:

    https://civitai.com/models/2174008?modelVersionId=2448186

    The standard workflow also works with this.

    SeeSeeLP
    Author
    Nov 28, 2025

    @Bars0199 You also need to replace Load Diffusion Model, Load CLIP, and Load VAE with a Load Checkpoint. Otherwise, using an AIO version wouldn't make sense 😊.

    https://drive.google.com/drive/folders/1hzCorrSXDoqTBp3AyPGWL-R1n1cx3xhY?usp=sharing

    fre19861Dec 1, 2025· 1 reaction

    I'm using a 2060s card, also got the size-mismatch error, is it because the model doesn't support it?

    SeeSeeLP
    Author
    Dec 1, 2025

    @fre19861 No, thank you, you need to update your ComfyUI to the latest version.

    CoolGeniusNov 29, 2025
    CivitAI

    Getting size mismatch error with fp8

    even after following this workflow https://civitai.com/models/2173571/z-image-turbo-aio?modelVersionId=2448013

    with this workflow https://civitai.com/models/2174008?modelVersionId=2448186

    tany6666372Nov 29, 2025

    It worked for me when I updated the workflow.

    CoolGeniusNov 29, 2025

    @tany6666372 could you tell

    what changes you done, can provide the workflow

    SeeSeeLP
    Author
    Nov 29, 2025

    @CoolGenius

    The workflow should work immediately - just drag the .json or .png into ComfyUI,

    place the model file in the correct folder, and press "Run". That's it!

    Here's what might be wrong:

    🔴 MOST LIKELY ISSUE: Wrong Loader Node

    The AIO version MUST use "Load Checkpoint" node, NOT "Load Diffusion Model"!

    Quick Fix Steps:

    1. ✅ Place Z-Image-Turbo-FP8-AIO.safetensors in: ComfyUI/models/checkpoints/

    2. ✅ Drag workflow JSON into ComfyUI

    ---

    If still getting "size mismatch" error, try these:

    🔧 Solution 1: Update ComfyUI

    Your ComfyUI might be outdated!

    Update via ComfyUI Manager:

    - Open ComfyUI Manager

    - Click "Update ComfyUI"

    - Restart ComfyUI

    OR manually:

    cd ComfyUI

    git pull

    restart ComfyUI

    🔧 Solution 2: Verify File Location

    ❌ DON'T put in: models/diffusion_models/

    ❌ DON'T put in: models/unet/

    ✅ MUST be in: models/checkpoints/

    SeeSeeLP
    Author
    Nov 29, 2025

    @CoolGenius

    I tried reproducing your mismatch error but couldn't trigger it in any scenario.

    The workflow loads and runs perfectly on my end.

    Most likely causes:

    1. Corrupted download → Re-download the file (FP8 should be ~10GB)

    2. Wrong folder → Must be in models/checkpoints/

    3. Outdated ComfyUI → Update it

    Try re-downloading the model file first - that's usually the issue when it

    works for others but not for you.

    Season_4Nov 29, 2025· 2 reactions

    可以尝试更新Comfyui

    SeeSeeLP
    Author
    Nov 29, 2025· 1 reaction

    @Season_4 👍

    qekNov 29, 2025· 1 reaction

    @SeeSeeLP AI generated text with Qwen?

    SeeSeeLP
    Author
    Nov 29, 2025

    @qek nope claude

    tany6666372Nov 29, 2025· 2 reactions

    @CoolGenius , ComfyUI is outdated → Update it...That's how it happened to me, but the quality is not very good

    qekNov 29, 2025· 1 reaction

    @SeeSeeLP My local LLMs gen better text and no censorship

    SeeSeeLP
    Author
    Nov 29, 2025

    @qek Cool, what LLMs do you use? I'm still a big fan of Ollama for Locale LLMs.

    CoolGeniusNov 29, 2025· 1 reaction

    @SeeSeeLP @tany6666372 Thanks ComfyUI needs to be updated rest of the steps I am doing it correctly.vIts working now

    qekNov 29, 2025· 1 reaction

    @SeeSeeLP Really? Ok
    CaptainErisNebula-12B-Chimera-v1.1
    magnum-v4-9b-abliterated
    Hathor_Sofit-L3-8B-v1
    Nyanade_Stunna-Maid-7B
    Impish_LLAMA_4B
    Dolphin3.0-Qwen2.5-3b
    EXAONE-3.5-2.4B-Instruct-abliterated
    2B-ad_HA_NL
    Qwen3-VL-8B-NSFW-Caption-V4.5
    Huihui-Qwen3-VL-4B-Instruct-abliterated
    Huihui-Qwen3-VL-2B-Instruct-abliterated
    Nanonets-OCR2-3B
    Qwen2.5-Coder-7B-Instruct-abliterated

    SeeSeeLP
    Author
    Nov 29, 2025

    @qek Nice list!

    I've used some of them before, like CaptainErisNebula, but mostly for Sillytarvern.

    Thanks for your time and reply! 👍

    qekNov 29, 2025· 1 reaction

    @SeeSeeLP They are extremely uncensored, but models trained on erp are more suitable for it

    LordSambaladoDec 2, 2025· 2 reactions

    update comfyui to 3.75

    sunnydubadws87671Nov 29, 2025· 1 reaction
    CivitAI

    Does this work well with lora? as I know, qwen edit AIO doesn't work well with lora

    qekNov 29, 2025· 1 reaction

    It isn't the kind of it. You're about Qwen Rapid, this Z-Image AIO is clean

    MaybeAINov 30, 2025· 4 reactions
    CivitAI

    Impressive, a full workin checkpoint with Clip and Vae and under 10GB, hope you get your hands on the base model soon :D

    SeeSeeLP
    Author
    Dec 1, 2025· 3 reactions

    I can definitely tell you that I'm currently setting up a new dataset for model and lora training and creating new anime loras.

    techbiz221Dec 1, 2025· 2 reactions
    CivitAI

    Kudos !!...It looks great. I will give this a try. Question: Can you use weights in the prompts? I know you cant use negative prompts just like Flux but not sure about the positive text encoder

    SeeSeeLP
    Author
    Dec 1, 2025

    @techbiz221 Thank you! Glad you like it.
    About your question: as far as I know, you can’t really rely on weights in the prompt here. It depends on how the Qwen text encoder interprets them, and if the model wasn’t trained to use that kind of weighting, it won’t have any effect.

    Negative prompts also won’t work if you plug them directly into the KSampler, just like with Flux. At that stage they basically do nothing. If you want your negatives to actually work, you need to inject them before the main prompt. There’s a node for that — one with both a model input and a conditioning input. You connect your negatives there, and then send the model output into the “AuraFlow” node. Unfortunately, I don’t remember the exact name of the node.

    Hope that helps! 😊👍

    techbiz221Dec 1, 2025· 1 reaction

    @SeeSeeLP I see. I am not well verse with those nodes, I'll do my research. Thanks for your efforts, it does help!

    LordSambaladoDec 2, 2025· 3 reactions
    CivitAI

    I was about to make my own checkpoint when I found this. haha thanks this save my time

    SeeSeeLP
    Author
    Dec 2, 2025

    That makes me happy, have fun 👍😊

    LordSambaladoDec 2, 2025· 4 reactions
    CivitAI

    For anyone got size-mismatch error :

    Update comfyui to v3.75
    if you're using web ui version not desktop, go check the git branch , if it's not on master it probably will stuck and can't update .
    At least that's what worked for me

    SeeSeeLP
    Author
    Dec 2, 2025

    That's exactly right 👍

    prodriver8301991Dec 6, 2025· 1 reaction

    Thank you so much!! That is so weird. Does it only work with 3.75? I had the newest version and it wouldn't work... that same ridiculously long size mismatch error

    SeeSeeLP
    Author
    Dec 6, 2025

    @prodriver8301991 I'm glad it's working for you now. I'm currently on version 3.77 because of ControlNet, so you need version 3.75 or higher.

    Best regards and happy holidays! 🎄

    anadarco12227Dec 3, 2025· 1 reaction
    CivitAI

    iam using the BF16 version using your workflow but i get black screen

    anadarco12227Dec 3, 2025· 1 reaction

    got the issue need to load model in bf16

    SeeSeeLP
    Author
    Dec 3, 2025

    @anadarco12227 I'm familiar with this error, as I use this workflow myself. Could it be that you don't have the latest version of ComfyUI? If not, try downloading a second version and testing with that; it could also be a custom node issue.

    anadarco12227Dec 4, 2025· 1 reaction

    @SeeSeeLP I found the issue you need to include dtype node and set it to bf16.then it worked like charm.yes comfyui update is must

    SeeSeeLP
    Author
    Dec 4, 2025

    @anadarco12227 Great, thanks for your feedback, it might also help others who have a problem with Z-Image.

    batartDec 3, 2025· 3 reactions
    CivitAI

    Which text encoder is integrated inside? Quantised Qwen?

    SeeSeeLP
    Author
    Dec 3, 2025· 1 reaction

    Yes, it's the same, just adapted by me and scaled down to FP8.

    anadarco12227Dec 4, 2025· 4 reactions
    CivitAI

    The quality in bf16 is fantastic and also optimised on my 3060.the good thing is that the vae and textencoder are both merged.That gives me advantage of reducing load on my card.i guess yours is the only one which has textencoder merged as on other checkpoints I dont see merged ones.

    SeeSeeLP
    Author
    Dec 4, 2025

    Thank you for your comment and feedback.

    AlderDeanDec 9, 2025· 1 reaction

    How do you fit bf16 into a 3060? How much VRAM do you have? As for me, bf16 crashes Python on my 3060 with 12GB VRAM and 32GB RAM.

    SeeSeeLP
    Author
    Dec 9, 2025

    @AlderDean It's probably due to the setup. I did all the pictures, tests, and so on with 8GB of VRAM. If you tell me a bit more about your system, like your software and workflow, I might be able to help you find the problem.

    SeeSeeLP
    Author
    Dec 9, 2025

    @AlderDean Otherwise, you can download the new ComfyUI as a second system for testing purposes. My workflow uses:

    Z-Image-Turbo-AIO Official Workflow v2

    https://civitai.com/models/2174008?modelVersionId=2478180

    and download the two custom nodes:

    rgthree-comfy

    comfyui_image_metadata_extension

    This is also mentioned in the workflow.

    AlderDeanDec 9, 2025· 1 reaction

    @SeeSeeLP Thank you for your swift reply and willingness to help, I really appreciate it. No matter how I try, the larger version keeps crashing my Python. I might have to look into some deeper settings and apply some fine-tuning tricks but I am too old and lazy for that. I think I'll stick to the FP8 version for the time being. FP8 works fine, BTW. Thanks again for a great checkpoint.

    anadarco12227Dec 10, 2025

    @AlderDean .Bf16 takes time just follow the workflow provided it works fine. Iam having 12Gb version

    syphonfilterargAIDec 5, 2025· 3 reactions
    CivitAI

    Hi, I loved this model, not only because it's easier to use, but also because it works perfectly on a GTX 960 with 2GB of VRAM. Could you upload them to the TensorArt platform? Thanks.

    SeeSeeLP
    Author
    Dec 5, 2025· 1 reaction

    Hey, thank you very much and I'm glad it's working for you. Great feedback 😊👍

    I don't have an account at TensorArt, but I'll take a look there.

    Thanks for the suggestion

    korederomaine356Dec 8, 2025· 1 reaction

    I like these comments. I have 2070 8gb, 32gb ram, and get oom. Yes, perfect on 2gb 960, my ass.

    SeeSeeLP
    Author
    Dec 8, 2025

    @korederomaine356 That's due to your ComfyUI version! I've even had it running with 8GB VRAM and 32GB DDR4 RAM, and hey, let's keep things civil!

    SeeSeeLP
    Author
    Dec 8, 2025

    @korederomaine356 Since the only information I have is that you’re getting an OOM (Out Of Memory) error in ComfyUI when using my Z-Image version, I’ll try to give you some general troubleshooting steps:

    Check if your ComfyUI installation is up to date.
    To be sure, go into the update folder and run the ComfyUI update script/file.

    Disable your custom nodes.
    It’s enough to disable the display-related nodes first. Then test again to see if the issue persists.

    Windows and GPU vendors have different memory-management settings.
    Depending on your system, Windows version, and GPU manufacturer (NVIDIA/AMD/Intel), certain settings can affect VRAM usage. Try searching for GPU memory-management options for your setup—other users have had similar issues and found solutions that way.

    Try a clean ComfyUI installation for testing.
    Download a fresh version of ComfyUI in a new folder so you don’t overwrite your current setup. Test Z-Image with that clean installation to see if the issue is related to your current environment.

    Have you tried my workflow yet?
    This is the workflow I currently use for everything related to Z-Image, so testing with it could help determine whether the issue is workflow-specific.

    Additional tips:

    Monitor your VRAM usage while generating.
    Tools like MSI Afterburner, NVIDIA-SMI, or Windows Task Manager can show you how quickly you’re hitting VRAM limits.

    Reduce input resolution or batch size.
    High resolutions or multi-batch generation can easily cause OOM errors, especially on GPUs with limited VRAM.

    Check for conflicting or outdated custom nodes.
    Some nodes leak memory or allocate more VRAM than expected. Updating or removing them can help.

    Clear ComfyUI’s cache.
    Deleting the ComfyUI\cache folder (while ComfyUI is closed) can sometimes reduce unnecessary memory usage.

    korederomaine356Dec 8, 2025

    @SeeSeeLP I used Forge Neo( i think it has bad memory management). I enabled never OOM, it works with it, just ultraslow. When the model gets speed improvement tools, i will switch to atleast for swamui. But, no, 2gb vram is bullshit.

    SeeSeeLP
    Author
    Dec 8, 2025

    @korederomaine356 First of all, thank you for your feedback — I’m glad to hear it’s working for you.

    About the user with 2 GB of VRAM: I really can’t say for sure whether it will work or not, since I don’t have a card with that little VRAM to test myself. But it’s also not as unbelievable as it might sound.

    For example, I often work with QWEN-Edit, which is over 20 GB even in FP8 — and that’s only the image part. Once you add the text encoder and the VAE, the full model is around 27 GB. And it still runs on my GPU with only 8 GB VRAM without any issues:
    (1024×1536 | Steps: 6 | CFG: 1.0 | Time: ~32.07s)

    So why shouldn’t a model that’s about 6 GB run on a GPU with 2 GB VRAM, especially when the text encoder stays on the CPU and parts of the model get unloaded during image generation?

    What I’m trying to say is: it might work, and it might not — I honestly don’t know. But as my example shows, it’s definitely not impossible.

    adhrithfarrell850Dec 9, 2025· 1 reaction

    @SeeSeeLP yes, that only works in comfy(offload fully the encoders and vae). Forge tries first the Vram, then offload, what not fits into it. That is huge waste of time.

    syphonfilterargAIDec 10, 2025· 1 reaction

    @SeeSeeLP  Thank you, 😊

    syphonfilterargAIDec 10, 2025· 1 reaction

    @SeeSeeLP It works with driver 531.61, my processor is an AMD A8 7680, 15.4 GB of RAM, Windows 10

    syphonfilterargAIDec 10, 2025· 1 reaction

    I just updated ComfyUI to the latest version, and it created an image faster than before.

    SyntheticScenesDec 5, 2025· 3 reactions
    CivitAI

    Keeps crashing Python (12 GB VRAM + 16 system RAM) on latest swarm UI

    SeeSeeLP
    Author
    Dec 5, 2025

    Unfortunately, I can’t really help you any further, except to suggest updating your SwarmUI again. The “Model Support.md” list was last updated on GitHub yesterday, and z-image is now included. I don’t see any reason why it shouldn’t work there, but I’m also not very familiar with SwarmUI. It’s possible that your 16 GB of system RAM just isn’t enough for the text encoder, SwarmUI, the operating system, and everything else running in the background.

    SyntheticScenesDec 6, 2025· 1 reaction

    @SeeSeeLP Thanks! I iupdated Swarm UI and reduced refine upscale size, it was the culprit! :D Probably my RAM was "just enough" ..

    SeeSeeLP
    Author
    Dec 6, 2025· 1 reaction

    @SyntheticScenes Great! 👍 It's wonderful that it's working for you now.

    Have fun generating your images.

    If you'd like, feel free to share them here.

    Happy Holidays!

    VisionaryAI_StudioDec 6, 2025· 4 reactions
    CivitAI

    Really well done!😊

    SeeSeeLP
    Author
    Dec 6, 2025· 2 reactions

    Thank you so much 👍😊👍

    grisha199432602Dec 7, 2025· 3 reactions
    CivitAI

    I just got into neural networks yesterday)) I seem to have downloaded everything correctly,but the results are worse in quality than in the examples
    Can you tell me what I'm doing wrong? I have an RTX 4060 Ti with 16GB.
    And how do I transfer comfy:20 Nodes? It's just copied as text.
    Thank you very much in advance for your help


    SeeSeeLP
    Author
    Dec 8, 2025

    Hey, I'm using my workflow v2. It also contains some tips.

    https://civitai.com/models/2174008?modelVersionId=2478180

    SiegwaldDec 7, 2025· 3 reactions
    CivitAI

    Can I use this model on automatic1111?

    SeeSeeLP
    Author
    Dec 8, 2025

    @Siegwald I haven't tested it, but if you like, you can try it and let me know if it worked.

    SiegwaldDec 8, 2025· 2 reactions

    @SeeSeeLP Not working

    SeeSeeLP
    Author
    Dec 8, 2025· 1 reaction

    @Siegwald First of all, thanks for testing and for your feedback.

    I just noticed that the AUTOMATIC1111 versions — including the Forge version — are already several months old. With versions that outdated, Z-Image definitely won’t run!

    Official repos:
    https://github.com/AUTOMATIC1111/stable-diffusion-webui
    https://github.com/lllyasviel/stable-diffusion-webui-forge

    However, I did find this:
    https://github.com/Haoming02/sd-webui-forge-classic
    Starting from version 2.6 (released last week), Z-Image should work there.

    If you're still using the older versions mentioned above, you should switch to the newer one linked below — or consider using ComfyUI instead if you’re okay with using node-based workflows.

    Here’s also a tutorial on how to install it (or check GitHub directly):
    https://www.stablediffusiontutorials.com/2025/11/forge-neo-installation.html

    kainnosgoth1988Dec 8, 2025· 1 reaction

    @SeeSeeLP yeah tested on Auto1111 and it indeed doesn't work... Forge Classic or Neo looks like its using the exact same UI as auto1111, which is good news to me... I hope it won't be difficult to get the same parameters as i had before cause they worked perfectly for my computer! Thanks for the links

    SiegwaldDec 9, 2025

    @SeeSeeLP Thank you very much for the advice

    qekDec 10, 2025

    @SeeSeeLP SD WEBUI Forge is a fork, not official

    SeeSeeLP
    Author
    Dec 7, 2025· 4 reactions
    CivitAI

    !!!UPDATE!!!

    The workflow for ControlNet is complete:

    https://civitai.com/api/download/models/2480804?type=Archive&format=Other

    Rixh666Dec 9, 2025· 3 reactions
    CivitAI

    Is the ram you are using 64gb DDR5 6000MHz?

    SeeSeeLP
    Author
    Dec 9, 2025

    two of these kits, Corsair Dominator Platinum RGB white UDIMM 32GB Kit, DDR5-6200, CL36-39-39-76

    IggortDec 12, 2025· 1 reaction

    Можно и на дряхлой машине. 3050 (8 гб) 32 RAM

    SeeSeeLP
    Author
    Dec 12, 2025

    @Iggort Thank you for your feedback!

    vichikaru978Dec 10, 2025· 3 reactions
    CivitAI

    any sample workflow? thanks a lot

    SeeSeeLP
    Author
    Dec 10, 2025
    Phoenix69420Dec 10, 2025· 3 reactions
    CivitAI

    The large version crashes when i load it, i have 32GB ram and RTX 4070 Super

    SeeSeeLP
    Author
    Dec 12, 2025

    @Phoenix69420 Maybe try running it with my workflow to see if it helps. If that doesn’t fix it, you could also download a clean, fresh ComfyUI installation and test it as a second setup. Here’s my workflow if you want to try it: https://civitai.com/models/2174008?modelVersionId=2478180

    Phoenix69420Feb 14, 2026

    @SeeSeeLP nvm, 32GB ram wasn't enough, i needed 50+ lol

    kainnosgoth1988Dec 12, 2025· 4 reactions
    CivitAI

    can-t this model do NSFW stuff? Like i can do naked and kissing, but sex acts, penetration, fellatio and all that just doesn't work

    IggortDec 12, 2025· 1 reaction

    Оно и не будет работать. Ищите спец лоры. Здесь же, на сайте.

    SeeSeeLP
    Author
    Dec 12, 2025

    @kainnosgoth1988 It’s an uncensored model, yes — but it can only generate what the base version is capable of. If you want to go deeper into NSFW themes, you’ll need the right LoRAs, since there isn’t a fully fine-tuned model for that yet.

    kainnosgoth1988Dec 12, 2025· 1 reaction

    @SeeSeeLP Ah i see, any recommendations? Thanks!

    SeeSeeLP
    Author
    Dec 12, 2025

    @kainnosgoth1988 No, at the moment things aren’t really usable, because the LoRAs are being trained with an adapter to avoid losing the Turbo performance. So basically we have to wait until the base version releases — I think things will get better after that. Unless another AI company suddenly drops an even faster model with more training data on the market 😅👍

    SeeSeeLP
    Author
    Dec 12, 2025

    @kainnosgoth1988 As an alternative for newer models, you can use Pony 7 or Chroma1HD. Otherwise, the usual AI models like SDXL and ILL work fine. I also have an anime version of Chroma that can handle everything: https://civitai.com/models/2022057/chroma-anime-aio?modelVersionId=2288507

    qekDec 12, 2025· 1 reaction

    @SeeSeeLP Pony 7 is terrible

    SeeSeeLP
    Author
    Dec 12, 2025

    @qek I don't really like it either 😅 you're better off going with the old pony models.

    kainnosgoth1988Dec 12, 2025

    @SeeSeeLP ah... i tried some of the new Zimageturbo lora releasing here, but they 0 effect at all, like it's not activated, i don't know if i'm doing anything wrong with it...in that case i'll keep using illustrioous models as i always had and zimageturbo separately for realistic stuff in the meantime

    PilanokuDec 13, 2025· 3 reactions
    CivitAI

    Hi! Can someone explain what the difference between this and Z-Image-Turbo is, please? :)

    SeeSeeLP
    Author
    Dec 13, 2025· 6 reactions

    Hi @Salciano ! Great question! 👋

    Main difference:

    Z-Image-Turbo (original):

    - Separate files: Model + VAE + Text Encoder

    - You need to download and load 3 different files

    Z-Image-Turbo-AIO (this version):

    - All-in-One - VAE and Text Encoder integrated into the model

    - Just ONE file to download and load

    - Much simpler to use!

    Two versions available:

    🟡 FP8-AIO (10GB) - Recommended for most users

    - Text encoder also scaled to FP8 (smaller file size)

    - Introduces slight noise in high/low frequency areas

    - Creates a "grittier" look - perfect for photorealistic images!

    - Adds extra touch of realism

    - Works great on 8GB VRAM

    🟢 BF16-AIO (20GB) - Maximum precision

    - Full BFloat16 precision throughout

    - Cleaner, smoother output

    - Better for anime or very clean lines

    - Also works on 8GB VRAM

    TL;DR:

    AIO = everything in one file, easier setup. FP8 for photorealism, BF16 for anime/clean styles. Same speed! 🚀

    Hope that helps! Let me know if you have more questions! 😊

    SeeSeeLP
    Author
    Dec 13, 2025· 1 reaction

    and you can use my workflows and other AIO workflows. 😊👍 https://civitai.com/models/2174008/z-image-turbo-aio-workflow

    PilanokuDec 13, 2025

    @SeeSeeLP Thank you very much!

    So I guess this has better quality than the base model for photorealism/anime, depending on the version or overall?

    Is it abliterated/uncensored in some way? :)

    Also, is the 10GB/FP8 better in only photorealism or also anime than the original?

    What about the 20GB is it only better in anime or also photorealism?

    SeeSeeLP
    Author
    Dec 13, 2025· 5 reactions

    @Salciano Great questions!

    Quality: Both versions match the base model 1:1. You can do photorealism AND anime with either! No quality loss.

    The difference is texture:

    FP8 adds slight noise (due to smaller size + merged encoder) which actually looks great for photorealism - less "AI-generated", more natural. For anime with clean lines, you might prefer BF16's cleaner output. But honestly, the difference is subtle - you'll only really notice when comparing side-by-side with identical settings.

    Censorship: Both are uncensored! Z-Image-Turbo base is already uncensored (Apache 2.0), and AIO maintains this. 🔓

    Bottom line:

    - FP8: Perfect for photorealism, still great for anime

    - BF16: Perfect for clean anime lines, still great for photorealism

    Both work for everything - just pick based on your main use case! 😊

    ovatographyDec 15, 2025· 8 reactions
    CivitAI

    its worth downloading , what a nice model 👌
    it works good with NVIDIA 3060 ti 8gb vram..

    SeeSeeLP
    Author
    Dec 15, 2025· 1 reaction

    @ovatography Thank you so much for your kind feedback, and it's great that it's working for you!

    BeeKhayDec 18, 2025· 9 reactions
    CivitAI

    This model is amazing, fast and high enough quality for GPUs with 8GB VRAM.

    I use WebUI Forge Neo. I hope this model has an optimal version for Forge Neo.

    MisccDec 22, 2025· 2 reactions
    CivitAI

    Thank you! ily

    Wendy_EarthDec 23, 2025· 1 reaction

    fr no creator explains like this much in detail, he has explained very beautifully.

    makoshark1975Dec 23, 2025· 3 reactions
    CivitAI

    Any chance of an FP16 version?

    SeeSeeLP
    Author
    Dec 23, 2025

    Thanks for the question 🙂

    At the moment I don’t see a strong technical reason to release an FP16 version, since a BF16 version is already provided.

    BF16 is generally superior to FP16 in terms of numerical stability because it uses the same exponent width as FP32. If your hardware supports BF16, there is usually no practical benefit in using FP16 instead.

    FP16 is mainly relevant for older GPUs or setups that do not support BF16. If that’s the case for you and you have a specific compatibility issue, feel free to let me know and I can consider it.

    makoshark1975Dec 24, 2025

    @SeeSeeLP Yeah I am running on a 1660 super which doesn't support bf16 so my system has to convert to fp32 behind the scenes which is super slow. I know I can use the fp8 but the 1660 super has acceleration for fp16 so figured I would ask.

    SeeSeeLP
    Author
    Dec 30, 2025

    @makoshark1975 is online 😊

    makoshark1975Dec 31, 2025

    @SeeSeeLP Awesome, I really appreciate it :)

    assmanagerDec 26, 2025· 4 reactions
    CivitAI

    I can't seem to load this model. It breaks almost immediately while trying to load the checkpoint, with this error message (I only included the beginning, it's VERY long):

    CheckpointLoaderSimple Error(s) in loading state_dict for NextDiT: size mismatch for x_embedder.weight: copying a param with shape torch.Size([3840, 64]) from checkpoint, the shape in current model is torch.Size([2304, 64]). size mismatch for x_embedder.bias: copying a param with shape torch.Size([3840]) from checkpoint, the shape in current model is torch.Size([2304]). size mismatch for noise_refiner.0.attention.qkv.weight: copying a param with shape torch.Size([11520, 3840]) from checkpoint, the shape in current model is torch.Size([3840, 2304]). size mismatch for noise_refiner.0.attention.out.weight: copying a param with shape torch.Size([3840, 3840]) from checkpoint, the shape in current model is torch.Size([2304, 2304]). size mismatch for noise_refiner.0.attention.q_norm.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([96]).

    What am I doing wrong? :(

    SeeSeeLP
    Author
    Dec 26, 2025· 1 reaction

    The "size mismatch" error occurs with an outdated ComfyUI version. Update your ComfyUI.

    assmanagerDec 26, 2025· 1 reaction

    Yeah, looks like that was the problem. Thanks for the quick response!

    SeeSeeLP
    Author
    Dec 26, 2025· 1 reaction

    @assmanager Always a pleasure, and I'm glad that Comfyui is now working for you 👌 Happy holidays 🎄

    masteratoms692Dec 29, 2025· 3 reactions
    CivitAI

    Pretty good, thanks

    Im not liking Z because you need to create a different prompt to generate a different image, the randomness is not there but overall the quality is amazing

    SeeSeeLP
    Author
    Dec 29, 2025

    Hey, thanks for your feedback!

    Many others are experiencing the same issue, but have you checked out my "Seed Variance Enhancer Workflow" for this problem? 😏

    https://civitai.com/models/2174008?modelVersionId=2500337

    Adds diversity to outputs by introducing controlled noise to text embeddings. Compensates for low seed variance – get more varied results with the same prompt. Includes manual seed control for reproducibility. Requires a SeedVarianceEnhancer custom node.

    I also have this in my anime workflow:

    https://civitai.com/models/2174008?modelVersionId=2544130

    masteratoms692Dec 29, 2025· 1 reaction

    Thanks a lot I will try it :) @SeeSeeLP 

    z1297416031672Dec 30, 2025· 3 reactions
    CivitAI

    请问工作流哪里下载?

    SeeSeeLP
    Author
    Dec 30, 2025· 1 reaction
    z1297416031672Jan 1, 2026· 1 reaction

    @SeeSeeLP 谢谢

    SeeSeeLP
    Author
    Dec 30, 2025· 14 reactions
    CivitAI

    ! Update ! FP16 version released

    FP16 offers the widest compatibility and works on virtually all GPUs.

    BF16 is supported starting with NVIDIA RTX 3000 (Ampere) and newer.
    FP8 is primarily optimized for NVIDIA RTX 4000 series and newer, where it can be used most efficiently.

    Choose the version that best matches your hardware for the best experience.

    AieditorDec 30, 2025· 1 reaction

    Thank you! Did you use the regular VAE or UltraFlux?

    SeeSeeLP
    Author
    Dec 30, 2025· 1 reaction

    @Aieditor the regular VAE 😊

    p4tt3n832Jan 1, 2026

    Me and my Radeon thank you! o7

    actionmaster679854Jan 17, 2026

    I have an RTX 3060 with 12GB Ram, trying out the fp8, seems to work fine (around 25 seconds for a 832x1216 image)

    Can I make the FP16 version work? I have 32 GB Ram. But I don't know if that'll make it much slower.

    alucardnoir941Dec 30, 2025· 4 reactions
    CivitAI

    Not to sound like a dick - because this IS my go to checkpoint for Stability Matrix, so really, thanks for the work you did - but using different seeds and samplers even if the prompts are the same make the differences between the checkpoints look a lot bigger than they are. Take for example the dragon image. You're using 3 different seeds and 2 different samplers. Even if everything else is the same, that's still enough to give us different images. Is FP16 softer than FP8 and BF16? or is it just euler + Beta that gives it that look? Are those really the differences between FP8 and BF16 or is it just to the seed.

    I've tested both the FP8 and BF16 earlier in the month and there are differences, but for the same seeds the composition is mostly the same, FP8 being just more prone to error than BF16 (and extra hand here, and extra leg there). I unironically preferred the FP8 images in most cases I tested, even if they were more likely to have error. But based on the images you're showcasing, the differences seem a LOT bigger because of different seeds. (Of note though, ComfyUI seems to have had an update with their KSampler and it seems to no longer have a fixed seed option in it and I haven't had much luck with forxing a fixed seed with other nodes, so in case you though you were using the same seeds, that might be why your'e not)

    BTW, the FP8, did you do any calibration on those or just weight quantization without calibration?

    SeeSeeLP
    Author
    Dec 30, 2025· 3 reactions

    Thanks for the detailed comment — and no worries at all, I didn’t take it the wrong way 🙂
    I really appreciate you taking the time to test things thoroughly, especially since you’re actively using this checkpoint in Stability Matrix.

    You’re absolutely right that different seeds, samplers, schedulers, and step counts can significantly change the output, even when the prompt itself stays the same. In my showcases, I intentionally keep the prompts mostly identical but vary settings like seed, sampler, scheduler, and steps, because all of these factors play a major role in the final image. The resulting differences are therefore expected and, to some extent, intentional.

    Regarding FP8 vs. BF16:
    The main difference I observe is that FP8 introduces more noise. FP8 images often appear dirtier or grainier, while BF16 tends to be cleaner and more stable. With identical settings, composition between FP8 and BF16 usually remains very similar, but FP8 is more prone to structural errors.

    So FP16 or BF16 themselves aren’t inherently “softer” — what often gets interpreted as softness is usually a result of noise characteristics combined with sampler and scheduler behavior, rather than the numeric format alone.

    On the topic of fixed seeds in ComfyUI: I just tested this again using two checkpoint loaders (FP8 and BF16) with the same seed, and in my setup the results remain fully deterministic. I wasn’t able to reproduce the issue you mentioned, so at least on my end, fixed seeds are still behaving as expected.

    As for FP8 specifically: these builds are weight-quantized without an additional calibration pass. That choice was mainly about maintaining iteration speed and compatibility, but calibrated FP8 variants are definitely something I may explore further.

    Thanks again for the thoughtful feedback — discussions like this are genuinely helpful and push the project forward.

    alucardnoir941Dec 31, 2025

    @SeeSeeLP It's the Node 2.0 thing comfyui is testing. Turn that on and you lose access to the ksampler's control after generation option for seeds. The new node 2.0 UI hides A LOT of stuff and it's usually for the worse.

    As for my point about seeds, what I meant was that since the only difference between these checkpoints is precision (as opposed to your anime checkpoint), people would be better served were they able to see the same set of images generated under the same conditions bar the checkpoint and judge the precision trade-off without having to download and test each checkpoint individually.