CivArchive
    QuadForge / Wan 2.2 I2V / SVI 2.0 Pro / Prompt Enhancer / automated multi-part ComfyUI Workflow - v1.2 (SVI)
    NSFW
    Preview 116090578
    Preview 116090577

    v1.3:

    • Prompt enhancer added that takes the input image as a reference (LLM-based, Gemma 3 abliterated).

      • Works for SFW as well as NSFW

      • Your own prompt can be very basic and short.

      • You can feed it back its output and write what you want to have changed, to - for example - include trigger words.

      • If needed, its custom instruction prompt can be adjusted in each "Enhancer" node.

      • If you dont want to use it, the group needs to be manuall bypassed for each part.

    • Global LoRAs (SVI, lightx2v) only need to be loaded once.

    • Aspect ratio of the input image is automatically taken over for the video.

      • You only need to adjust the desired resolution tier in the CustomResolution node.

    Troubleshooting for the prompt enhancer if you get errors such as "LTXVGemmaClipModelLoader No package metadata was found for bitsandbytes":

    • Update ComfyUI

    • Update all custom nodes

    • open cmd in your ComfyUI portable folder and type:

      • python_embedded\python.exe -m pip install -U bitsandbytes

      • or try: python_embedded\python.exe -m pip install -U bitsandbytes accelerate


    Troubleshooting/best practice tips:

    • The SVI-LoRA is mandatory in each part for the workflow to function (link in the workflow).

    • The SVI-node does not seem to update with ComfyUI-manager, you`ll need to install it manually via Git (link in the workflow).

    • For Sage Attention, this guide on YouTube helped me a lot: "How to Install Sage Attention 2.2 on Latest ComfyUI Portable And Deskop Version".

    • Reduce the output bitrate to about 10 if you`re having trouble uploading to CivitAI.

    • Start with 2x, 3x, 4x turned off and at a lower resolution (e.g. k20) and then go from there.

    v1.2:

    • Workflow reworked to support SVI 2.0 Pro, which uses the starting image and latent video from the previous clip to:

      • Improve transitions

      • Enable cross-clip memory for faces and other attributes

    • Resolution can now only be set once for part 1; it is carried over to all subsequent clips

    • Added switches for quickly switching between GGUF and diffusion/.safetensors models (e.g., SmoothMix). Deactivate LightX2v when using SmoothMix, as it is already embedded.

      • Note: SmoothMix changes faces and appearances quite a bit with this workflow compared to the GGUFs of base WAN.

    • Sage Attention and patched Torch are now part of the model inputs and can be bypassed globally.

    • Output folder structure updated: partial videos are now saved in the same folder as last-frame images.

    • Separate switches for last frame and all frames image preview.


    v1.1:

    • Major declutter: Removed most crossing lines using set/get, subgraphs, and anything-everywhere nodes.

    • Single model input propagated everywhere.

    • Color-coded sections + cleaner layout.

    CustomResolution Node Update:

    • 9:16 & 16:9 support with shared resolution tiers.

    • 24 & 30 FPS support.

    • Auto-rounding to multiples of 8 for manual inputs.

    https://github.com/AugustusLXIII/ComfyUI_CustomResolution_I2V/tree/main

    v1.01:

    Changes:

    • Missing save last frame node added for scene 1.

    • Switches for image preview (reduces generation time) and interpolation added.


    Update: My CustomResolutionI2V node is now available via ComfyUI-manager. You might need to update ComfyUI-manager if you can`t find it.


    QuadForge WAN 2.2 – ComfyUI Workflow

    A powerful and flexible img2vid chaining workflow optimized for WAN 2.2 GGUF with the lightx2v-LoRA (6 steps).

    Start with a single input reference image and generate up to four consecutive video segments. Each segment automatically uses the last frame of the previous clip as its starting image for perfect continuity. All segments are stitched together seamlessly, with final video interpolation (RIFE) applied as the last step for a smooth, high-quality output.

    Key features:

    • Adjustable number of generations (1–4) via simple switches

      • Tip: Start with 2x, 3x, and 4x turned off and add them until you like the result of the previous part. For this to work, you have to keep the seeds, LoRAs, and prompts fixed. Bypass the interpolation video node until you like your final result to save time.

    • Independent prompt and LoRA control for each segment

    • SageAttention enabled for approx. 40% speedup (bypass the two nodes in each step if not installed yet or if not supported by your GPU)

    • Automatic last-frame extraction and saving after every generation

    • Quick resolution and clip length adjustments using my own CustomResolution node (available via ComfyUI Manager): https://github.com/AugustusLXIII/ComfyUI_CustomResolution_I2V)

    • Fully automated stitching

    • RAM cleanup and model unload nodes for better stability

    Ideal for creating long, consistent tracking shots, progressive scene reveals, or any multi-stage animated sequence from one starting image.

    Notes:

    • Lots of spaghetti but it works (will be polished in future updates)

    • Framerate output in the CustomResolution node is currently fixed to 16 fps (I will make that adjustable as well)

    • How-to manual: TBD

    Description

    v1.2:

    • Workflow reworked to support SVI 2.0 Pro, which uses the starting image and latent video from the previous clip to:

      • Improve transitions

      • Enable cross-clip memory for faces and other attributes

    • Resolution can now only be set once; it is carried over to all subsequent clips

    • Added switches for quickly switching between GGUF and diffusion models (e.g., SmoothMix). Deactivate LightX2v when using SmoothMix, as it is already embedded.

      • Note: SmoothMix changes faces and appearances quite a bit with this workflow compared to the GGUFs of base WAN.

    • Sage Attention and patched Torch are now part of the model inputs and can be bypassed globally

    • Output folder structure updated: partial videos are now saved in the same folder as last-frame images

    • Separate switches for last frame and all frames image preview

    FAQ

    Comments (10)

    FferrettJan 3, 2026· 3 reactions
    CivitAI

    Quite the evolution of your original quadforge. Now to add Ollama goonsai prompt autogeneration!

    aiibelJan 3, 2026· 1 reaction
    CivitAI

    Perfect workflow! In terms of stability, ease of use, and generation quality, this is by far the best one I've ever used.

    (By the way, I spent an entire afternoon troubleshooting to get the right sageattention version for my setup, but it was totally worth it!)

    NUVoizeJan 3, 2026
    CivitAI

    getting group node missing anyone else?


    Node 1091: UUID 77440340-3a38-4bdf-85bd-e86771cc43a9

    Node 1371: UUID f4ff6954-7d48-46cb-8a12-4f1be0e49862

    Node 1414: UUID 2f3c6f77-e06d-4c0e-9bf1-6d6b0e4b67fc

    qbdownJan 3, 2026· 1 reaction
    CivitAI

    one of the best workflows, but missing some prompt enhancer and perhaps mmaudio to make it complete

    Cloud_RonJan 3, 2026· 1 reaction
    CivitAI

    Good idea to load lightnings with the Power Lora Loader (rgthree) but there is no need to load it again and again if you inject it at the beginning. Also as it has a toggle the switch is not needed. Maybe you can use it once without a switch so you can load SVI lora also only once at the beginning. Overall still good memory management of your workflow. It took some time to figure out that the SVI lora is mandatory for this worflow as I got only grey rush foggy videos without :)

    jafdeth2030105Jan 4, 2026· 1 reaction
    CivitAI

    Guys, and what if my first start frame in video must begins with my character stands from back and turning in front? How can i set his face before turning?

    SECGTJan 6, 2026· 1 reaction
    CivitAI

    Never saw V1, but I can tell the amount of work needed to get this fine product done GJ! I noticed that each segment will change colors slightly. In my case red colors start going magenta at the end. Does this flow have a "Color Matching" node?

    NRubricJan 9, 2026
    CivitAI

    I don't know why, but I only getting a fully black output.
    I didn't changed anything and I'm using all the right models.

    Any idea?

    TheBodyHunterJan 15, 2026· 5 reactions
    CivitAI

    Thank you for your great work !
    Have you guys noticed that videos generated from SVI Pro have one pretty obvious flaw: the further you go toward the later clips, the video quality drops noticeably? The details of objects get way blurrier compared to the first clip, and in some more complex videos, the character's facial features even start getting messed up.

    Does anyone know if there's any method to reduce or fix this issue? Please help me out !

    roderick2690Jan 26, 2026· 1 reaction
    CivitAI

    I keep getting:

    [GetNode] ✗ Variable 'Last Frame P2' not found! Available: (none) Tip: Make sure SetNode runs BEFORE GetNode in the graph.

    Can someone help me fix it and explain to me like I'm 10? 😭

    Workflows
    Wan Video 14B i2v 720p

    Details

    Downloads
    1,025
    Platform
    CivitAI
    Platform Status
    Deleted
    Created
    1/2/2026
    Updated
    4/27/2026
    Deleted
    4/27/2026