Daxamur's Wan 2.2 Workflows
If you'd like to support me, check out my Patreon!
DM to inquire about custom projects.
-NEWS-
Responses are delayed as I'm heads down working on getting my next release ready for you all - once released, responses will go back to normal!
v1.2.1 Out Now! - Update to DaxNodes via ComfyUI manager required
FLF2V added with GGUF support - no new models required
Fixed ability to independently disabled / enable upscaling and interpolation
Dedicated resolution picker nodes, added auto-resizing functionality from v1.3.1 to I2V and FLF2V
DaxNodes now available via ComfyUI Manager, no more git clone required!
Current Tracked Bugs:
KJNodes Get / Set reporting a missing error for some users, if this happens - ensure you download the latest version of DaxNodes from ComfyUI manager, and re-import the workflow! - In progress
If you see a "FileNotFoundError ([WinError 2] The system cannot find the file specified.)" from VideoSave or other video-related nodes, FFmpeg is missing or not in your system PATH.
Setup (Full Version Required):
Download the full FFmpeg build
Extract it to a stable location (e.g., C:\ffmpeg).
Add C:\ffmpeg\bin to your system PATH:
Open Edit the system environment variables -> Environment Variables....
Under System variables, select Path -> Edit....
Click New and add C:\ffmpeg\bin.
Save and exit.
Restart ComfyUI (and your terminal/command prompt).
After this, everything should work!
v1.3.1 Features
Segment-Based Prompting
Persistent Positive Prompt: Keeps consistent details across the entire video (ie. “A woman with green eyes and brown hair in her warmly lit bedroom”).
Segment Positive Prompts: Separated with +, one per segment length (ie. “She is writing in a journal + She closes the journal and stands up + She walks away”).
Gives you far more control in long-form videos and helps reduce WAN’s tendency to render weird camera movements or jutters on I2V start.
Endless-Style Looping
Segments can chain "infinitely" (I capped the node at 9999), creating effectively endless loops.
The Video Execution ID manages overwrites and stitching - just increment the ID as you generate new sequences.
Streaming RIFE VFI + Upscaling
Tweaked RIFE VFI and upscaling now stream frames instead of holding entire sequences in VRAM/RAM.
Allows much longer videos, smoother interpolation, and sharper upscales without OOM errors.
Face Detection & Drift Correction
Intelligent Mediapipe face frame detection locks focus on characters.
Drift correction ensures the final video runs at least as long as requested - but instead of cutting mid-generation, it will add full extra segments until the target framecount is met or exceeded.
This way, no generated frames are wasted, and you always end up with smooth, complete segments.
Fully toggleable, with adjustable frame look-back settings.
Resolution Handling
T2V: Standard WAN resolution presets with optional overrides.
I2V: Input image scales to WAN-native resolutions, preserving aspect ratio. “Native” passthrough supported.
QoL & Management
Toggle upscaling/interpolation independently.
Temp file output organized by execution ID - clear /output/.tmp/ periodically to save space.
Looking Ahead
This workflow is still experimental , future versions will expand on segment control, smarter handling of motion/camera behavior, more adaptive face tracking, and even integration of audio/video for cinematic sequences. Big things are coming!
Notes
I've done my best to place most nodes that you'd want to configure at the lower portion of the flow (roughly) sequentially, while most of the operational / backend stuff sits at the top. Nodes have been labeled according to their function as clearly as possible.
Beyond that;
NAG Attention is in use, so it is recommended to leave the CFG set to 1.
The sampler and scheduler are set to uni_pc // simple by default as I find this is the best balance of speed and quality. (1.1> Only) If you don't mind waiting (a lot, in my experience) longer for some slightly better results, then I'd recommend res_3s // bong_tangent from the RES4LYF custom node.
I have set the default number of steps to 8 (4 steps per sampler) as opposed to 4, as here is where I see the most significant quality / time tradeoff - but this is really up to your preference.
This flow will save finished videos to ComfyUI/output/WAN/<T2V|T2I|I2V>/ by default.
I2V
The custom node flow2-wan-video will cause a conflict with the Wan image to video node and must be removed to work. I have found that this node does not get completely removed from the custom_nodes folder when removing via the ComfyUI manager, so this must be deleted manually.
GGUF
All models used with the GGUF versions of the flows are the same with the exception of the base high and low noise model. You will need to determine which GGUF quant best fits your system, and then set the correct model in each respective Load WAN 2.2 GGUF node accordingly. As a rule of thumb, ideally your GGUF model should fit within your VRAM with a few GB to spare.
The examples for the GGUF flows were created using the Q6_K quant of WAN 2.2 I2V and T2V.
The WAN 2.2 GGUF quants tested with this flow come from the following locations on huggingface;
MMAUDIO
To set up MMAUDIO, you must download the MMAUDIO models below, create an "mmaudio" folder in your models directory (ComfyUI/models/mmaudio), and place every mmaudio model downloaded into this folder (even apple_DFN5B-CLIP-ViT-H-14-384_fp16.safetensors).
Block Swap Flows
Being discontinued as I have found that the native ComfyUI memory swapping conserves more memory and slows down the process less in my testing. If you receive OOM with the base v1.2 flows, I'd recommend trying out the GGUF versions!
Triton and SageAttention Issues
The most frequent issues I see users encounter are related to the installation of Triton and SageAttention - and while I'm happy to help out as much as I can, I am but one man and can't always get to everyone in a reasonable time. Luckily, @CRAZYAI4U has pointed me to Stability Matrix which can auto-deploy ComfyUI and has a dedicated script for installing Triton and SageAttention.
You will first need to download Stability Matrix from their repository, and download ComfyUI via their hub. Once ComfyUI has been deployed via the hub, click the three horizontal dots to the top left of the ComfyUI instance's entry, select "Package Commands" and then "Install Triton and SageAttention". Once complete, you should be able to import the flow, install any missing dependencies via ComfyUI manager, drop in your models and start generating!
Will spin up a dedicated article with screenshots on this soon.
Models Used
T2V (Text to Video)
Wan2_2-I2V-A14B-HIGH_fp8_e4m3fn_scaled_KJ.safetensors (for loop segments)
Wan2_2-I2V-A14B-LOW_fp8_e4m3fn_scaled_KJ.safetensors (for loop segments)
lightx2v_T2V_14B_cfg_step_distill_v2_lora_rank64_bf16.safetensors
Wan21_I2V_14B_lightx2v_cfg_step_distill_lora_rank64.safetensors (for loop segments)
I2V (Image to Video)
Wan21_I2V_14B_lightx2v_cfg_step_distill_lora_rank64.safetensors
model.safetensors (renamed to clip-vit-large-patch14.safetensors)
MMAUDIO
Non-Native Custom_Nodes Used
Description
Release
FAQ
Comments (17)
works on my 3090 , results are great thanks
Great workflow, thanks much. Also, I love the way you organized it, with the logic and all segregated into nice clean 'sections'. Must be a programmer...:)
RES4LYF is hardly 'slightly better'- used correctly it is transformative. And it takes TWICE as long per iteration. There is no mystery to either fact.
Fair statement - it has its pros, and I agree the output is noticeably different and often better to a degree.
The catch for me is that, for frequent use, the extra time (2× for res_2s or 3× for res_3s) basically cancels out the speed gain you get from running the lightx2v LoRA in the first place. In my testing, I just don’t prefer the res/bong_tangent look enough to justify that hit, so UniPC stays my go-to - but I do get the appeal.
@blobby99 if you can't actually define what res4lyf samplers actually do in relation to how they generate from latent then you literally have no fucking point to make. learn to code or find a new hobby.
I haven't explored all those samplers and schedulers. Can you tell me the advantages of using hte RES4LYF nodes?
filmington The "s" in res_2s / res_3s just means how many sub steps (mini-passes) it does per true step 2s runs twice, 3s runs three times. It’s roughly 2× or 3× slower, but you get extra detail / "crispiness" per step (it does typically change the composition pretty dramatically as well). So while it can and often does give somewhat better (in my very humble opinion), and very different outputs, it either nearly negates the benefit of the lora speedup in the case of res_2s, or entirely and then some in the case of res_3s. Don't get me wrong though - they absolutely do have their use cases!
Daxamur Thanks for the explanation!
The T2V Workflow is working just fine, but the I2V gives me an error : KSamplerAdvanced
Given groups=1, weight of size [5120, 36, 1, 2, 2], expected input[1, 32, 21, 160, 90] to have 36 channels, but got 32 channels instead
For others, the cause of the error "KSamplerAdvanced
Given groups=1, weight of size [5120, 36, 1, 2, 2], expected input[1, 32, 21, 160, 90] to have 36 channels, but got 32 channels instead" is a conflict with the custom node flow2-wan-video.
I reinstalled comfyui venv version with help of Daxamur now it works flawlessly thanks <3
CRAZYAI4U is there no fix without reinstalling?
KVTM There is actually for the above error (you just need to uninstall the flow2-wan-video custom nodes from your ComfyUI) - @CRAZYAI4U had an additional error stemming from a python environment issue, causing sageattention to break after we'd resolved the first one
Daxamur i did. i deleted the flow2 wan video node folder but the issue persists
KVTM In that case if you can post a pastebin export of the flow when erroring, and your custom_nodes list I can take a look!
KVTM did you fix the issue ? If not I could help you with an easy fix, just DM me.
CRAZYAI4U yea i had comfyui desktop version and just switched to comfyui portable. that fixed it for me