CivArchive
    WAN 2.2 14B Multi-Phase I2V/T2V Workflow - WAN22_I2V/T2V14B_4PV143
    NSFW

    WAN 2.2 14B Multi-Phase I2V/T2V Workflow: Professional Video Generation Made Accessible

    Supports both Image-to-Video and Text-to-Video generation with the same optimized architecture

    Breaking Through Hardware Limitations

    Have you ever wanted to create longer, higher-quality AI videos but kept running into memory errors? This workflow solves that problem by splitting video generation into four independent phases that work together like a relay race. Each phase does its job, cleans up after itself, and passes the baton to the next runner.

    What's New: This workflow includes a custom-built WanSettingsController node that centralizes all video settings into one control point, eliminating the tedious process of manually updating dozens of nodes when you want to change resolution or aspect ratio.

    What Makes This Workflow Special

    Dual Mode: Image-to-Video AND Text-to-Video

    This workflow is designed to handle both I2V (Image-to-Video) and T2V (Text-to-Video) generation:

    Image-to-Video Mode - Load an input image and the workflow animates it through four refinement phases, adding motion, upscaling resolution, and interpolating frames. All four phases work seamlessly together, with each phase building on the previous one.

    Text-to-Video Mode - Disable the image input using the CR Image Input Switch (4 way) node and let the workflow generate video purely from text prompts. This is where the WildcardPromptFromString node becomes critical—T2V quality lives and dies by your prompt, and wildcards let you generate diverse, high-quality variations across large batches.

    Important T2V Limitation: Unlike I2V mode where phases can extend or enhance the video, T2V doesn't work well with multiple phases. Each phase generates a new scene from the text prompt rather than continuing the previous generation, breaking visual continuity. For T2V generation, you'll typically want to use only Phase 1 and disable Phases 2-4 using the Fast Groups Bypasser.

    The same memory management and batch processing capabilities work identically in both modes—you can still run 30+ T2V generations overnight without crashes.

    Multi-Phase Architecture: Doing More With Less

    Instead of trying to generate your entire video in one massive operation that would crash most systems, this workflow uses a smart phase-based approach with four separate video generation phases:

    The 4-Phase System - Each phase runs a complete WAN video generation cycle. This allows the workflow to:

    • Generate longer videos by chaining multiple generation cycles together

    • Overcome VRAM limitations by processing in manageable chunks

    • Clean memory between phases to prevent crashes during batch processing

    How It Works: Each WanImageToVideo node generates a video segment. The output from one phase can feed into the next, allowing you to extend video length beyond what a single generation could produce. Between each phase, RAM cleaners and model unloaders free up memory, resetting the system for the next generation cycle.

    Real-World Production Usage:

    • 1 phase = Quick single generations, T2V workflows (~6-8 seconds)

    • 2 phases = Standard production sweet spot (~12 seconds, ~15-20 minutes with lightning models)

    • 3-4 phases = Longer showcase videos (~20-34 seconds, extended generation time)

    Most daily production work uses 1-2 phases for efficiency and speed. Using 2 phases with lightning-based models, you can generate approximately 3 twelve-second videos per hour, enabling 30-50+ videos in overnight batch runs. The 3-4 phase capability is available when you need longer content for special projects.

    The Power of Independence: Each phase can be disabled individually or in groups using the Fast Groups Bypasser node. Need just one quick generation? Use only Phase 1. Want to chain generations together for longer videos? Enable multiple phases. This modular design means you can dial in exactly what you need without wasting processing time or resources.

    Note: This phase flexibility is particularly powerful for I2V workflows where phases can build on each other. For T2V workflows, you'll typically only use Phase 1 since each phase restart generates a new scene rather than continuing the previous one.

    Memory Management: The Secret Sauce

    The workflow includes aggressive RAM and VRAM cleaning at strategic points:

    RAM Cleaners (SoftFullCleanRAMAndVRAM) - Placed between phases to clear system memory, preventing the slow memory creep that eventually crashes batch processing. These nodes ensure each phase starts with a clean slate.

    Model Unloaders (SoftModelUnloader) - Actively remove models from VRAM when they're no longer needed. This is crucial for running large batches overnight without running out of video memory.

    Execution Order Controllers (ImpactExecutionOrderController) - Ensure cleanup happens at exactly the right moment, forcing the workflow to finish one phase completely before moving to the next.

    The Tradeoff: These cleanup nodes do add time to each generation cycle—models need to be reloaded from disk when the next batch starts. However, this is a strategic trade: spending a few extra seconds per video to reload models is vastly preferable to your entire batch crashing at video #15 because you ran out of memory. When you're running 30+ videos overnight, reliability trumps raw speed every time.

    Together, these create a system that can process dozens or even hundreds of videos in a single batch without filling up memory and crashing. This is the difference between babysitting your computer and waking up to completed work.

    LoRA Management System

    LoRA Loader (LoraManager) - A sophisticated LoRA loading system that tracks which LoRAs are applied, their strengths, and their trigger words. This isn't just loading LoRAs—it's managing them intelligently.

    Debug Metadata (LoraManager) - Captures all LoRA information into metadata that's compatible with Civitai and other platforms. When you upload your videos, people can see exactly what LoRAs you used.

    TriggerWord Toggle - Allows you to easily enable or disable LoRA trigger words without editing your prompt, giving you quick A/B testing capability.

    Central Control: WAN Settings Controller (Custom Node)

    The WanSettingsController is a custom-built node designed specifically for this workflow. It solves one of the biggest pain points in complex video workflows: changing settings across dozens of connected nodes.

    The Problem It Solves: In traditional workflows, adjusting your video resolution means hunting through your canvas, finding every node that needs width/height/frame settings, and manually updating each one. Miss a single node and your workflow breaks. Change your mind about aspect ratio? Start the treasure hunt again.

    The Solution: This custom controller is your command center. Instead of hunting through dozens of nodes to change resolution, you adjust one dropdown menu and the entire workflow updates automatically.

    Key Features:

    • 24 Pre-Validated Resolutions - Every resolution is tested and confirmed to work with WAN 2.2 14B, from mobile-friendly 576×1024 portrait to cinema-quality 1920×1080 landscape

    • Dimension Locking - All resolutions are mathematically locked to multiples of 16 (WAN's technical requirement), so you'll never accidentally break your workflow with invalid settings

    • Aspect Ratio Labels - Every resolution clearly shows its aspect ratio (9:16, 16:9, 1:1, etc.) so you know exactly what you're getting

    • Optimized Defaults - The 960×1216 (10:16) resolution is marked as the sweet spot for quality vs. performance

    • Five Outputs, One Source - Width, Height, Length (frame count), Frame Rate, and Batch Size all flow from this single node to everywhere they're needed

    Real-World Impact: Change your aspect ratio in one second, not five minutes. Test different resolutions without rewiring your workflow. Scale from portrait to landscape with a single dropdown selection.

    This isn't just convenience—it's what makes testing and iteration actually feasible at production scale. When you're generating 50+ videos per day, this node saves hours of workflow management time.

    Node Types Explained (The Building Blocks)

    Custom Workflow Control

    WanSettingsController - This is a custom-built node created specifically for this workflow. It centralizes all video settings into a single control point with 24 pre-validated WAN-compatible resolutions. Think of it as replacing a dozen knobs scattered across a control panel with one master dial. When you change the resolution dropdown, width, height, length, frame_rate, and batch_size outputs automatically update across your entire workflow. This eliminates the tedious (and error-prone) process of manually updating settings in multiple nodes. The node includes portrait orientations (576×1024 to 1080×1920), square formats (768×768, 1024×1024), and landscape orientations (832×480 to 1920×1080), all validated to work with WAN 2.2 14B and locked to 16-pixel multiples. This single innovation transforms workflow iteration from a chore into a single click.

    Core Processing Nodes

    KSamplerAdvanced (8 instances) - The heavy lifters that actually generate images and video frames using diffusion models. These nodes handle the AI's creative process, iteratively refining noise into coherent visuals.

    WanImageToVideo (4 instances) - Specialized nodes that convert images into video using the WAN 2.2 14B model. Each instance handles one phase of the video generation pipeline.

    VAEDecode (4 instances) - Converts latent space representations (the compressed format the AI works in) back into actual pixels you can see. Every image has to pass through a VAE to become viewable.

    Video Creation and Export

    CreateVideo (5 instances) - Assembles individual frames into video files, handling frame rates, codecs, and timing.

    SaveVideo (5 instances) - Writes the completed videos to your drive with proper naming and metadata.

    RIFE VFI - The frame interpolation engine that creates smooth in-between frames, doubling (or more) your effective frame rate using optical flow estimation.

    Workflow Organization

    ReroutePrimitive|pysssss (46 instances) - These are like junction boxes in electrical wiring. They let you connect distant nodes without spaghetti cables crossing your entire canvas. Essential for keeping complex workflows readable.

    Fast Groups Bypasser (rgthree) - Your phase control panel. This single node lets you enable or disable entire groups of nodes, making it trivial to test specific phases or skip unnecessary processing.

    Power Primitive (rgthree) - A smarter primitive node that can feed values to multiple inputs simultaneously, reducing clutter.

    ImpactExecutionOrderController (4 instances) - Forces specific execution sequences, crucial for ensuring memory cleanup happens between phases rather than at random times.

    Image Processing

    ImageScale - Resizes images while maintaining quality, used in the upscaling phase.

    ImageFromBatch - Extracts individual images from batch processing, useful for preview and quality checks.

    ImageBatchMulti (3 instances) - Combines multiple images into batches for efficient processing.

    CR Image Input Switch (4 way) - A critical router that lets you switch between four different input images OR disable image input entirely for Text-to-Video generation. This is your I2V/T2V mode selector—when you want pure T2V, this node cuts out the image input and lets the model generate from prompt alone. No rewiring needed to switch between modes.

    PreviewImage (3 instances) - Displays images during generation so you can monitor progress without waiting for final output.

    Text and Prompt Handling

    CLIPTextEncode (2 instances) - Converts your text descriptions into the mathematical format (embeddings) that the AI understands.

    Power Prompt - Simple (rgthree) - An enhanced prompt node with better formatting and organization options.

    WildcardPromptFromString - Critical for T2V generation. Enables prompt randomization using wildcards (like {adjective}, {action}, {lighting}), letting you generate diverse variations across large batches. In Text-to-Video mode, prompt quality is everything—a mediocre prompt gets mediocre results, while a well-crafted prompt with strategic wildcards produces usable, compelling video. This node is your secret weapon for batch diversity: instead of generating 30 identical videos, you're generating 30 unique variations by randomly combining different descriptive elements. Essential for maintaining quality and variety in T2V workflows.

    JoinStringMulti - Combines multiple text strings into one, useful for building complex prompts from modular pieces.

    Model Loading

    UNETLoader (2 instances) - Loads the WAN 2.2 14B model components. WAN 14B requires two separate models: a "high" model and a "low" model that work together during video generation. The two UNET loaders handle loading both model components that the WAN workflow needs.

    CLIPLoader - Loads the text encoder that converts words into concepts the AI understands.

    VAELoader - Loads the VAE (Variational Autoencoder) that converts between latent space and pixel space.

    CLIPSetLastLayer - Controls how many layers of the text encoder to use, allowing fine-tuning of how literally the AI interprets prompts.

    ModelSamplingSD3 (2 instances) - Configures the sampling behavior for Stable Diffusion 3 architecture models, controlling generation quality and characteristics.

    Utility Nodes

    MathExpression|pysssss (3 instances) - Performs calculations on values in your workflow, useful for dynamic frame counts, resolution scaling, and parameter adjustments.

    VHS_GetImageCount (3 instances) - Counts frames in video sequences, essential for phase coordination and batch processing.

    Number of Active Phases - Critical control node that must be set to match how many phases you're actually using (1-4). If you're using only Phase 1, set this to 1. Using Phases 1-3? Set it to 3. This node coordinates the workflow's execution and must match your Fast Groups Bypasser settings. Forgetting to set this correctly will cause workflow errors.

    MarkdownNote - A documentation node where you can write notes about what different sections of your workflow do. Invaluable for complex setups.

    ShowText|pysssss - Displays text values for debugging and confirmation that settings are correct.

    SaveImageWithMetaData - Saves images with embedded generation parameters, so you can always recreate your results.

    Memory Management (Critical!)

    SoftFullCleanRAMAndVRAM|LP (2 instances) - Aggressively frees both system RAM and GPU VRAM, preventing memory accumulation during batch processing. Yes, this adds a few seconds to reload models at the start of each new generation, but that's the price of reliability—without these cleaners, you'd crash mid-batch instead of completing 30+ videos overnight.

    SoftModelUnloader|LP - Removes models from VRAM when they're no longer needed, freeing space for subsequent phases. The model reload time is negligible compared to the hours you'd lose restarting a crashed batch.

    The Big Picture: How It All Works Together

    Think of this workflow like a production line in a factory:

    1. Raw materials enter (your input image/prompt and settings from WanSettingsController)

    2. Station 1 (Phase 1) runs a complete video generation cycle

    3. Cleanup crew clears the workspace (RAM/VRAM cleaning)

    4. Station 2 (Phase 2) runs another video generation cycle (optional—can be disabled)

    5. Cleanup crew strikes again

    6. Station 3 (Phase 3) runs another generation cycle (optional—can be disabled)

    7. Final cleanup

    8. Station 4 (Phase 4) runs the final generation cycle (optional—can be disabled)

    9. Quality control (preview nodes show results)

    10. Shipping (SaveVideo writes final files)

    Each station runs a complete WAN generation cycle independently. If a station isn't needed, you turn it off with the Fast Groups Bypasser. If you're running multiple products (batch processing), the cleanup crews ensure the workspace never gets cluttered between generations.

    Why This Matters

    For Beginners: You get a professional workflow that handles complexity behind the scenes. Change one setting, get reliable results. Start with 1-2 phases for quick wins.

    For Veterans: You get granular control over every phase, with the ability to disable what you don't need and batch process at scale without crashes. Optimize for your use case—quick 12-second videos for volume, or longer showcase pieces when quality demands it.

    For Everyone: You get longer videos, higher resolution, smoother motion, and the ability to run it all overnight without running out of memory. With lightning models, expect approximately 3 twelve-second videos per hour using 2 phases, enabling 30-50+ video overnight batch runs.

    This workflow represents months of optimization, testing, and problem-solving, all packaged into a system that "just works." Whether you're making content for social media, testing LoRAs, or pushing the boundaries of AI video generation, this workflow gives you the tools to do it efficiently and reliably.

    Technical Requirements

    • ComfyUI with WAN 2.2 14B model support

    • WAN 2.2 14B Models: You need both the "high" and "low" model files (WAN 14B is a two-model system)

    • VRAM: Minimum 12GB for base operation, 16GB+ recommended for higher resolutions

    • RAM: 32GB+ recommended for batch processing

    • Custom Nodes Required:

      • ComfyUI-Impact-Pack (for execution controllers)

      • ComfyUI-Custom-Scripts (for math expressions)

      • rgthree-comfy (for power nodes and bypasser)

      • LitePicker/ComfyUI-MemoryManagement (for RAM/VRAM cleaners)

      • LoraManager nodes

      • WAN Settings Controller (custom node - INCLUDED with this workflow!)

      • RIFE VFI nodes

      • Video Helpers Suite

    Getting Started

    Installing the WAN Settings Controller (Custom Node)

    The custom WanSettingsController node is included as wan_settings_controller.py. To install:

    1. Copy wan_settings_controller.py to your ComfyUI/custom_nodes/ directory

    2. Restart ComfyUI

    3. The node will appear in the video/settings category

    That's it! The node is self-contained with all 24 validated resolutions built in. No dependencies beyond base ComfyUI.

    Running the Workflow

    For Image-to-Video (I2V):

    1. Install all required custom nodes (see Technical Requirements above)

    2. Install the WanSettingsController custom node

    3. Load the workflow in ComfyUI

    4. Select your desired resolution in WanSettingsController dropdown

    5. Set the "Number of Active Phases" node to match how many phases you're using (1-4)

      • If using only Phase 1, set to 1

      • If using Phases 1-3, set to 3

      • This must match the phases you have enabled via Fast Groups Bypasser

    6. Load your input image using the CR Image Input Switch

    7. Enable/disable phases as needed using Fast Groups Bypasser

    8. Queue and let it run!

    For Text-to-Video (T2V):

    1. Follow steps 1-4 above

    2. Set the "Number of Active Phases" node to 1 (T2V only uses Phase 1)

    3. Use the CR Image Input Switch to disable image input

    4. Disable Phases 2-4 using Fast Groups Bypasser

      • Each phase generates a new scene from the prompt, breaking continuity

      • Use only Phase 1 for T2V generation

    5. Craft your prompt using the WildcardPromptFromString node

      • Use wildcards for variation: {lighting|golden hour|dramatic shadows|soft diffused}

      • Build modular prompts: {subject} in {location}, {camera angle}, {mood}

      • Remember: T2V quality depends heavily on prompt quality—invest time here

    6. Queue your batch and review results

    T2V Pro Tip: Since you're only using Phase 1, generation is much faster than multi-phase I2V. This makes T2V ideal for rapid iteration and testing—you can generate and evaluate prompts quickly, then scale up to larger batches once you've dialed in your wildcards.

    The workflow handles the rest, managing memory, coordinating phases, and producing high-quality video without requiring constant supervision.


    This workflow is designed for both learning and production. Study how the phases interact, experiment with disabling different sections, and scale up to batch processing when you're ready. The modular design means you can understand one piece at a time while still having a complete, working system from day one.

    Description

    WAN 4-Phase V1.4.3 Update

    What's New in V1.4.3

    This update adds GGUF model support and fixes model unloader connections in the final processing stage.

    New Features

    GGUF Model Support

    • Added SmartSequentialWANLoaderGGUFPhase1 node for loading quantized GGUF models

    • Added 2x CR Model Input Switch nodes to select between safetensors and GGUF loaders

    • New "GGUF Models" and "safesensor models" groups for organization

    • Users can now run the workflow with GGUF models for lower VRAM usage

    Model Input Switching

    • CR Model Input Switch nodes allow you to choose which loader type to use

    • Switch between safetensors (standard) and GGUF (quantized) models

    • Manual selector - set to input 1 for safetensors, input 2 for GGUF

    Fixed Issues

    Model Unloader Connections

    • Fixed Node 376's 'model for unload' connection (was disconnected in V1.4)

    • Model unloaders now properly chain in the Final Processing section

    • These unload WAN models after all 4 phases complete, freeing VRAM for final upscaling/interpolation operations

    System Requirements

    Minimum Requirements

    • GPU: NVIDIA GPU with 16GB+ VRAM

    • System RAM: 32GB+

    • Storage: SSD with 100GB+ free space

    • OS: Windows 10/11, Linux, or macOS

    • ComfyUI: Latest version

    GGUF Model Benefits

    • Lower VRAM requirements compared to safetensors models

    • Faster loading times

    • Same generation quality with proper quantization

    Required ComfyUI Extensions

    Install these via ComfyUI Manager:

    1. ComfyUI-VideoHelperSuite (VHS) - Video generation and processing

    2. ComfyUI-Impact-Pack - Execution control and utilities

    3. ComfyUI-LoRA-Manager - LoRA loading and management

    4. rgthree's ComfyUI Nodes - Power Prompt, Fast Groups Bypasser

    5. pysssss Custom Scripts - ShowText, ReroutePrimitive nodes

    6. ComfyUI-KJNodes - ImageBatchMulti, JoinStringMulti nodes

    7. ComfyUI-GGUF (city96) - GGUF model loading support

    8. ComfyUI LayerPeople Nodes - ModelUnloader, SoftFullCleanRAMAndVRAM

    9. ComfyRoll Custom Nodes - CR Model Input Switch, CR Image Input Switch

    Required Custom Nodes

    Download these Python files and place in ComfyUI/custom_nodes/:

    1. smart_sequential_wan_loader.py - Safetensors model loader (Phase 1)

    2. smart_sequential_wan_loader_gguf.py - GGUF model loader (new in V1.4.2)

    3. multi_phase_prompt_manager_4phase_v2.py - Multi-phase prompt management

    4. advanced_wan_settings_controller.py - Video settings control

    Restart ComfyUI after adding these files.

    Installation Instructions

    Model Setup

    • Safetensors models: Place in ComfyUI/models/diffusion_models/

    • GGUF models: Place in ComfyUI/models/unet/ or ComfyUI/models/diffusion_models/

    • VAE: Place in ComfyUI/models/vae/

    • CLIP: Place in ComfyUI/models/clip/

    Workflow Installation

    1. Download WAN_4-Phase_V1_4_2.json

    2. Open ComfyUI

    3. Drag and drop the JSON file into the ComfyUI interface

    4. ComfyUI will alert you if any required extensions are missing

    Quick Start Guide

    Step 1: Choose Your Model Type

    The workflow includes two model loader groups:

    • "safesensor models" - For standard safetensors WAN models

    • "GGUF Models" - For quantized GGUF WAN models

    To use safetensors models:

    1. Configure your models in the "safesensor models" group

    2. Use Fast Groups Bypasser to bypass/disable the "GGUF Models" group

    3. Set both CR Model Input Switch nodes to input 1

    To use GGUF models:

    1. Configure your models in the "GGUF Models" group

    2. Use Fast Groups Bypasser to bypass/disable the "safesensor models" group

    3. Set both CR Model Input Switch nodes to input 2 (current default setting)

    Step 2: Configure Video Settings

    In the Advanced WAN Settings Controller:

    • Set base resolution (default: 896x1152 - 7:9 Portrait)

    • Set frame count (default: 160 frames)

    • Set generation FPS (default: 16)

    • Set final output FPS (default: 32)

    • Choose number of active phases (default: 1, range 1-4)

    • Configure upscale multiplier (default: 1.35)

    Step 3: Set Workflow Mode

    In your Phase 1 loader:

    • I2V Mode: Load an image and generate video from it

    • T2V Mode: Generate video from text prompts only

    Step 4: Configure Prompts

    In the Multi-Phase Prompt Manager:

    • Enter positive prompts for each phase

    • Add LoRA trigger words if needed

    • Configure negative prompts per phase

    • Enable/disable individual phases

    Step 5: Configure LoRAs (Optional)

    Each phase has its own LoRA loader section:

    • Phase 1-4 each support multiple LoRAs

    • Use the LoRA Manager visual interface to select LoRAs

    • Configure LoRA strengths per phase

    CRITICAL - Avoid Double-Loading LoRAs:

    LoRAs loaded in Phase 1 remain active throughout all subsequent phases. When configuring Phases 2-4:

    • DO NOT re-enable/load LoRAs that were already loaded in previous phases

    • If you load the same LoRA in multiple phases, you're stacking its weight, which causes severe anomalies and distortions

    • Only use trigger words in later phases to activate LoRAs that are already loaded

    • Example: If you load LoRA_A with strength 1.0 in Phase 1, just add the trigger word in Phase 2 prompts - don't load LoRA_A again in Phase 2's LoRA loader

    The LoRAs stay in VRAM across all phases. Re-loading them multiplies their effect and creates very bad artifacts.

    Step 6: Generate

    1. Queue your prompt

    2. Phase 1 generates initial video

    3. Phases 2-4 extend the video sequentially

    4. Final Processing unloads models and combines all segments

    5. Outputs are saved in your configured output directory

    Using GGUF Models

    What Are GGUF Models?

    GGUF (GPT-Generated Unified Format) models are quantized versions of standard models that use less VRAM while maintaining quality.

    Configuration

    SmartSequentialWANLoaderGGUFPhase1 accepts:

    • Workflow mode selection (I2V or T2V)

    • Separate I2V high/low model selection

    • Separate T2V high/low model selection

    • Quantization settings (dequant_dtype, patch_dtype)

    The loader automatically loads the appropriate model pair based on your selected workflow mode.

    Model Switching

    The CR Model Input Switch nodes have a selector widget:

    • Input 1: Routes safetensors loader output

    • Input 2: Routes GGUF loader output (current default)

    You need two switches (one for high_model, one for low_model). Both should be set to the same input number.

    Troubleshooting

    GGUF models not appearing in loader:

    • Verify ComfyUI-GGUF extension is installed

    • Place GGUF files in models/unet/ or models/diffusion_models/

    • Restart ComfyUI

    Wrong models loading:

    • Check CR Model Input Switch selector values (both should match)

    • Verify the correct loader group is enabled in Fast Groups Bypasser

    • Check that the unused loader group is bypassed

    Out of VRAM errors:

    • Try GGUF models instead of safetensors

    • Reduce number of active phases

    • Lower base resolution

    • Reduce batch size

    • Disable upscaling phases

    Workflow won't load:

    • Install all required extensions via ComfyUI Manager

    • Ensure all custom .py files are in custom_nodes directory

    • Restart ComfyUI after installing extensions

    Technical Notes

    Model Loading Architecture

    V1.4.2 uses a dual-loader system:

    1. SmartSequentialWANLoaderPhase1 (safetensors)

    2. SmartSequentialWANLoaderGGUFPhase1 (GGUF)

    Both loaders run in parallel but only one set of outputs is used via the CR Model Input Switch nodes.

    Model Unloading

    ModelUnloader nodes are located in the "Final Processing" group. They:

    1. Unload WAN models after Phase 4 completes

    2. Free VRAM before upscaling/interpolation operations

    3. Chain together (Node 376 → Node 394 → SoftFullCleanRAMAndVRAM)

    This ensures maximum VRAM is available for RIFE upscaling and FPS interpolation.

    Phase Architecture

    All 4 phases use the same models loaded once in Phase 1. There is no model switching between phases - the loaded models remain in VRAM throughout all generation phases, then get unloaded during final processing.

    Credits

    Workflow Design: DarthRidonkulous
    Custom Nodes: DarthRidonkulous
    ComfyUI-GGUF Extension: city96
    ComfyRoll Nodes: RockOfFire (ComfyRoll Custom Nodes)
    LayerPeople Nodes: LayerPeople

    Find me on CivitAI: @DarthRidonkulous

    Version History

    V1.4.2 (Current)

    • Added GGUF model support via SmartSequentialWANLoaderGGUFPhase1

    • Added CR Model Input Switch nodes for loader selection

    • Fixed model unloader connections in Final Processing

    • Added "GGUF Models" and "safesensor models" organization groups

    V1.4

    • Previous version (see previous documentation)

    FAQ

    Workflows
    Wan Video 2.2 I2V-A14B

    Details

    Downloads
    975
    Platform
    CivitAI
    Platform Status
    Available
    Created
    12/7/2025
    Updated
    4/26/2026
    Deleted
    -

    Files

    wan2214BMultiPhaseI2V_wan22I2VT2V14B4PV142.zip

    wan2214BMultiPhaseI2V_wan22I2VT2V14B4PV143.zip

    wan2214BMultiPhaseI2V_wan22I2VT2V14B4PV142.zip