CivArchive
    Wan2.2 I2V V2V Video Extend Quant 14B - v1.0
    NSFW

    Description (What this workflow is)

    This is my Wan2.2 I2V → V2V “Video Extend” cascade workflow built around Wan2.2 14B quant (GGUF) High/Low noise pair, designed to generate a first clip, then extend it in multiple stages while keeping motion continuity and prompt control.

    Core idea: “make a strong first clip → re-inject last-frame → extend again”. You get a practical pipeline for longer videos without turning your generation into a 1-hour single sampling job.

    • Uses GGUF quant 14B High/Low noise models (wan2.2_i2v_high_noise_14B_Q3_K_M.gguf + wan2.2_i2v_low_noise_14B_Q3_K_M.gguf).

      Wan2.2 I2V Quant 14B Cascade 1…

    • Output is saved via VideoHelperSuite (VHS_VideoCombine) at 16 fps, H.264 MP4, CRF 19, yuv420p, metadata on.

    • Built around the common Wan2.2 I2V “HighNoise/LowNoise” split that ComfyUI also documents in their official Wan2.2 workflows.


    What’s inside (Features / “what’s cool here”)

    1) True multi-stage extension (cascade)

    Instead of trying to brute-force one long clip in a single run, this workflow:

    • generates the start clip,

    • extracts/uses the last frame as the seed for the next clip,

    • repeats extension multiple times,

    • saves each stage as its own MP4 with separate filename prefixes (so you can compare and pick the best stage quickly).

    2) Pinned “don’t touch anything else” controls

    The key nodes are pinned so you don’t hunt through spaghetti:

    • GGUF loaders for High/Low.

      Wan2.2 I2V Quant 14B Cascade 1…

    • Multiple LoRA loader slots (stack-style) so you

      Wan2.2 I2V Quant 14B Cascade 1…

      uickly.

      Wan2.2 I2V Quant 14B Cascade 1…

    • Pinned video combine/export nodes per cascade s

      Wan2.2 I2V Quant 14B Cascade 1…

    3) Fast mode: LightX2V logic (4-step mindset)

    This workflow is designed to support the Light2X / LightX2V low-step style that people use to make Wan2.2 I2V practical. Community workflows show the same “4 steps with Light2X LoRA” approach as the main speed hack.

    Your practical rule (what you wrote, translated into “how to actually run it”):

    • ON (LightX2V-ish) → low steps, fast iterations.

    • OFF (no LightX2V) → you can use random custom LoRAs, but step count jumps to ~20–40 and render time explodes.

    4) Export quality defaults that don’t sabotage you

    • CRF 19 + yuv420p for compatibility.

    • 16 fps as a stable baseline for Wan2.2 I2V style clips.

    • Metadata saving enabled.


    Usage (How to run it)

    0) Required baseline (ComfyUI side)

    You need a working Wan2.2 I2V setup (models in the right folders). The official ComfyUI docs list the standard Wan2.2 model components and structure.

    1) Load workflow

    Drag & drop the JSON into ComfyUI.

    2) Set the models (GGUF quant)

    Select your HighNoise + LowNoise 14B GGUF in the pinned loaders. The workflow is already pointed at:

    • wan2.2_i2v_high_noise_14B_Q3_K_M.gguf

    • wan2.2_i2v_low_noise_14B_Q3_K_M.gguf

      Wan2.2 I2V Quant 14B Cascade 1…

    3) Pick your speed strategy

    A) Fast motion mode (recommended for iteration)

    • Keep the LightX2V approach (4-step style). This is the same core trick people use in other Wan2.2 workflows to get “minutes not hours.”

      Wan2.2 I2V Quant 14B Cascade 1…

      RA freedom mode**

    • Disable LightX2V and run your own LoRAs (character/style/etc).

    • Expect 20–40 steps territory and much slower total runtime (your note matches what most people experience: low-step LoRA tricks exist specifically to avoid that).

    4) Prompt like you mean it (this workflow is prompt-driven)

    For short clips (1.5–3s), it behaves best.
    For longer clips (6s+), prompt precision becomes mandatory:

    • Write movement instructions explicitly (camera + subject motion).

    • Write the ending motion/pose explicitly (you need to know where the clip ends).

    • Treat each cascade stage as its own “shot continuation” and prompt accordingly.

    5) Run cascade stages

    Each stage outputs a separate MP4 via VHS_VideoCombine nodes (you’ll see filename prefixes for the extend steps).


    Settings (What to tweak first)

    Video export (safe defaults already set)

    • frame_rate: 16

    • format: video/h264-mp4

    • pix_fmt: yuv420p

    • crf: 19

    • save_metadata: true

    If you want smaller files: increase CRF (23–28).
    If you want higher compatibility: keep yuv420p.

    LoRA slots (fast switching)

    This workflow is set up with multiple LoRA loader nodes so you can keep options “parked” and toggle quickly.

    Wan2.2 I2V Quant 14B Cascade 1…

    Also included: an AccVid LoRA option is present in the file (Wan21_AccVid_T2V_14B_lora_rank32_fp16.safetensors).

    Wan2.2 I2V Quant 14B Cascade 1…


    Troubleshooting (the stuff that actually breaks)

    “My custom LoRA doesn’t work with LightX2V”

    That’s expected in practice: LightX2V-style motion LoRAs often don’t like being mixed with other LoRAs in the same run (behavior varies by LoRA and how it was trained). Community workflows frequently treat Light2X as the LoRA for the low-step motion trick, n

    Wan2.2 I2V Quant 14B Cascade 1…

    k casually.

    Fix:

    • Disable LightX2V → use custom LoRAs → accept higher steps.

    “Wit

    Wan2.2 I2V Quant 14B Cascade 1…

    s forever”

    Yes. That’s literally why Light2X/LightX2V workflows exist: to avoid high-step runtime.

    Fix:

    • Keep clips shorter (1.5–3s) and extend via cascade.

    • Iterate on prompts with fast mode first, then do one “final” higher-step run only when you already know the motion.

    “My extended stage drifts / changes subject”

    Common for I2V extension pipelines when the prompt is vague.
    Fix:

    • Add anchor descriptors (identity, outfit, environment, lighting) every stage.

    • Add explicit motion endpoints (final pose/camera state).

    “Video won’t play on my phone”

    Use the default export:

    • H.264 MP4, yuv420p, CRF 19.


    FAQ

    Q: What is this workflow optimized for?
    A: Practical I2V generation + multi-stage extension on Wan2.2 14B quant GGUF, with exports that just work.

    Q: Can I use other Wan2.2 quant files (Q4/Q5/Q8, etc.)?
    A: Yes — swap the GGUF model selections in the pinned High/Low loaders. The pipeline logic stays the same.

    Wan2.2 I2V Quant 14B Cascade 1…

    Q: Why two models (HighNoise + LowNoise)?
    A: That’s the standard Wan2.2 pattern: high-noise for establishing motion/structure, low-noise for refinement. ComfyUI’s official Wan2.2 docs follow the same split.

    Q: Why does the workflow feel “better” on 2–3 seconds?
    A: Short clips reduce drift and make prompt control tighter. Then you extend through cascade instead of praying that one long run stays coherent.

    Q: Can I get 32 fps?
    A: Yes, but that’s typically done with interpolation nodes after generation (many community workflows do exactly that).

    Description

    FAQ