喜欢中文的你看这边:英文看完后就是中文版
WAN 2.2 5b WhiteRabbit Interp-Loop
This ready-to-run ComfyUI workflow turns one image into a short looping video with WAN 2.2 5b. Then, it cleans the loop seam so it feels natural. Optionally, you can also boost the frame rate and upscale with ESRGAN.
In other words, this is an Image to Video workflow that creates loops with WAN 2.2 5b!
Why is this so complicated?!
WAN 2.2 5b does not fully support injecting frames after the first. If you try to inject a last frame, it will create a looping animation but the last 4 frames will be "dirty" with a strange "flash" at the end of the loop.
This workflow leverages custom nodes I designed to overcome this limitation. We trim out the dirty frames and then interpolate over the seam.
Model Setup (WAN 2.2 5b)
Install these into the usual ComfyUI folders. FP16 = best quality. FP8 = faster and lighter, with some trade-offs.
Diffusion model → models/diffusion_models/
- FP16: wan2.2_ti2v_5B_fp16.safetensors
- FP8: Wan2_2-TI2V-5B_fp8_e5m2_scaled_KJ.safetensors
Text encoder → models/text_encoders/
- FP16: umt5_xxl_fp16.safetensors
- FP8: umt5_xxl_fp8_e4m3fn_scaled.safetensors
VAE → models/vae/
- wan2.2_vae.safetensors
Optional LoRA → models/lora/
- Recommended: Live Wallpaper Style
Tip: keep subfolders like models/vae/wan2.2/ so your growing collection stays tidy.
How It Works
- Seam prep: we take the very last and first frames and generate new in-betweens that bridge them smoothly. Only those new frames get appended — no duplicate of frame 1.
- Full-clip interpolation (optional): multiply in-betweens across the whole video, then resample to any FPS you want.
- Upscale (optional): add an upscaler pass before full-clip interpolation using an ESRGAN model of your choice.
- Output: saved to your ComfyUI/output/ folder, filename prefix LoopVid.
Controls You’ll Care About
Defaults are set for “safe on most GPUs.” Tweak if you have more VRAM.
Full-Clip Interpolation
- Roll & Multiply: add more in-betweens everywhere (e.g., ×3).
- Reample Framerate: convert to an exact FPS (e.g., 60). Great after Multiply, but you could use it on its own.
Other handy knobs
- Duration: WAN cost climbs past ~3s (2.2 is tuned up to ~5s).
- Working Size: long edge in pixels (shape comes from your input image).
- Steps: ~30 is WAN 2.2’s sweet spot.
- CFG: WAN default is 5, I have it bumped a little higher. Higher = more “prompt strength,” sometimes more motion.
- Schedule Shift: motion vs. stability. Higher = more motion.
- Upscale: choose model/target size; reduce tile/batch if you hit OOM.
You can find more detailed information on all these settings in the workflow itself.
Using Vision Models for Prompts (optional but handy)
If writing movement prompts feels daunting, you can use a vision model to get a great starting point. You have a few options.
Free Cloud Options
Google's Gemini or OpenAI's ChatGPT are free and will get the job done for most people.
- Upload your image and paste the prompt below.
- Copy the model’s description and paste it into this workflow’s Prompt field.
...however, these services are not exactly private and might censor lewd/NSFW requests. That's why you might prefer to explore the other two options.
Paid Cloud Options
There are many services that offer cloud model access which is a more reliable way to get uncensored access to models.
You could pay for credits on OpenRouter for example. Personally, I prefer Featherless because they charge a flat monthly fee which keeps my costs predictable, and they have a strict no log policy. If you decide you want to give them a try, you could always use my referral link which helps me out!
If you decide to go the API/Paid Cloud route, you might find my app, CloudInterrogator, useful. It's designed to make prompting cloud vision models as easy as possible and it's fully free and open source!
Local Inference Option
I know a lot of people on CivitAI are local-or-nothing types. For you, there is Ollama.
Here's the best guide I could find on setting it up. You will want to look at Google's Gemma-3 family of models and look at which size is appropriate for your card.
If you use Ollama, there's nothing stopping you from using CloudInterrogator as your access point since Ollama creates OpenAI compatible endpoints, or you could customize this workflow with Ollama nodes for ComfyUI. I don't recommend doing the latter unless you can set it to lock the prompt.
Many workflows for WAN build Gemma3/Ollama nodes into the workflow. I decided not to do that, because I think 99% of people are going to be well serviced by Gemini or ChatGPT.
Suggested prompt:
Analyze the content of this video frame and write a concise, single-paragraph description of your predictions around what movement takes place throughout the video sequence that follows.
Your description should include the details of the character and scene as a whole but only as they related to the movement that occurs in the scene. In addition, make note of the movements of the particles, blinking of the eyes if any, movement of the hair... this is a moment captured in time, and you are describing these few seconds encompassed by the image. Everything that can move, does move - even minute details of the scene.
Do not describe ‘pauses’. Don't minimize the motion with words like ‘slight’ or ‘subtle’. Do not use metaphorical language. Your description must be direct and decisive. Use simple, common language. Be specific, and describe how each detail in the scene moves, but do not be verbose; each word in your description must have purpose. Use the present tense, as if your predictions are coming true as you type them.
You will deliver one paragraph without any additional information and without any special characters that format this response, avoid ‘The image sequence depicts the character’ and describe what happens, without saying ‘the video ....’"You might also have good luck with the suggested prompt from AmazingSeek's workflow depending on the model you use or what you're looking for!
Tips & Troubleshooting
WAN Framerate: WAN 2.2 is 24 fps. On WAN 2.1, if you decide to try it, set fps to 12 instead. There is a slider for this near the model loader node. The workflow auto-calculates what to do with your framerate (for multiplication and resampling) based on this number.
Seam looks off? Try switching between Simple/Fancy seam interpolation; increase the auto-crop search in Fancy; or re-render with a slightly different prompt/CFG.
Out-of-memory (OOM)?
- Lower tile size (x and y) in the WanVideo Decode node.
- Lower Upscale tile size and/or batch size.
- Reduce Working Size or Duration.
- Enable “Use Tiled Encoder”.
AttributeError: type object 'CompiledKernel' has no attribute 'launch_enter_hook'
I'm not sure what causes this, though my assumption is it has something to do with the WAN Video Nodes. This should fix it for you:
1. Open "🧩 Manager"
2. Click "Install PIP Packages"
3. Install these two, leave the quotes out: "SageAttention", "Triton-Windows".
3.1 Obviously Triton-Windows is only for Windows users. If you get this error on Linux, I would guess the package for Triton is just "Triton".
If this doesn't fix it for you, it may be that your ComfyUI Python environment is messed up for some reason or the version of Comfy you're using doesn't work with the Manager "Install PIP Packages" module. In that case, you might find this advice from the comments section helpful:
From alex223:
"i spent almost a day, but made it work. this thing helped, but also, for some reason my embedded python missed include and libs folder, I copied them from standalone version - that was essential for triton to work. Maybe my comment will help someone."
If you're still having problems, you can leave a comment. I don't mind trying to help people troubleshoot but I don't think the issue is with my workflow or with WhiteRabbit (my custom nodes).
Acknowledgements
- It occurred to me that interpolating over a loop seam might be a good solution to the "dirty frames" problem when I was first experimenting, but it was this workflow by AmazingSeek that really made me decide to go for it.
- It appears that Ekafalain should get some credit here, too, for their seamless loop workflow on which AmazingSeek's is based.
- While I didn't end up using any of their ideas directly, I want to shout out Caravel for their excellent, multi-step process you can have a look at over here that seems to primarily target WAN 2.2 14b. The level of documentation in this workflow alone is laudable.
- My recommended vision prompt is built off of NRDX'. You can find the original workflow it's from over on his patreon. This is the guy who is training LiveWallpaper LoRA for various WAN models, too!
P.S. 💖
If this workflow helps you, I’d love to see what you make! I put a lot of hard work into making it, including designing custom nodes to bring it all together and trying to document as much as possible so it is maximally useful to you.
Links
- Have a look at the WhiteRabbit repository for node documentation and atomic workflows if you want a better idea of how to build with the custom nodes here or tweak this workflow.
- My Website & Socials: See my art, poetry, and other dev updates at artificialsweetener.ai
- Buy Me a Coffee: You can help fuel more projects like this at my Ko-fi page
This workflow is dedicated to my beloved Cubby 🥰
- Find her artwork all over the internet
- She has many excellent LoRA on CivitAI for you to explore :3
WAN 2.2 5b WhiteRabbit 插值循环
这个开箱即用的 ComfyUI 工作流可将一张图片转换为使用 WAN 2.2 5b 生成的短循环视频。随后,它会清理循环衔接处的“接缝”,让过渡更自然。可选地,你还可以提升帧率并用 ESRGAN 进行放大。
换句话说,这是一个利用 WAN 2.2 5b 生成循环效果的“图像转视频”工作流!
为什么会这么复杂?!
WAN 2.2 5b 并不完全支持在首帧之后继续注入帧。如果你尝试注入最后一帧,它虽会生成循环动画,但最后 4 帧会出现“脏帧”,在循环结束处出现奇怪的“闪烁”。
此工作流通过我设计的自定义节点来规避这一限制。我们先裁掉脏帧,然后对接缝进行插帧插值。工作流内同时提供了“简单版”和“进阶版”的裁剪/插值流程,并配有切换开关,便于你分别试用。
模型设置(WAN 2.2 5b)
按常规 ComfyUI 目录安装这些文件。FP16 = 质量最佳;FP8 = 更快更省显存,但有一定取舍。
扩散模型 → models/diffusion_models/
文本编码器 → models/text_encoders/
VAE → models/vae/
可选 LoRA → models/lora/
提示:使用诸如 models/vae/wan2.2/ 这类子文件夹,便于管理不断增长的模型集合。
工作原理
接缝准备:取最后一帧与第一帧,生成新的过渡中间帧以实现平滑衔接。只会追加这些新帧——不会重复追加第 1 帧。
全片插值(可选):在整段视频中增加倍数级的中间帧,然后重采样到任意 FPS。
放大(可选):在全片插值之前加入一次放大流程,使用你选择的 ESRGAN 模型。
输出:保存到你的 ComfyUI/output/ 文件夹,文件名前缀为 LoopVid。
你会关心的控制项
默认设置为“对多数 GPU 安全”。如果你显存更充裕,可以适当调高。
全片插值
滚动倍增 ("Roll & Multiply"):在全片范围增加更多中间帧(例如 ×3)。
重采样帧率 ("Resample Framerate"):转换到精确的 FPS(例如 60)。在倍增后使用效果更佳,但也可单独使用。
其他实用旋钮
时长 ("Duration"):超过 ~3 秒成本上涨(2.2 调校到 ~5 秒)。
工作尺寸 ("Working Size"):以长边像素为准(纵横比来自输入图)。
步数 ("Steps"):~30 是 WAN 2.2 的甜点区。
CFG:WAN 默认 5,这里略微上调。数值越高=“提示强度”更高,有时也会带来更多运动。
日程偏移(Schedule Shift):运动 vs 稳定。数值越高=运动更强。
放大 ("Upscale"):选择模型/目标尺寸;如遇 OOM,降低 tile/batch。
关于这些设置的更多细节,可在工作流中直接查看。
使用视觉模型来生成提示(可选但好用)
如果编写“运动提示”让你犯难,可以借助视觉模型获得一个很好的起点。你有多种选择。
免费云端方案
Google 的 Gemini 或 OpenAI 的 ChatGPT 是免费的,对多数人来说足够用了。
上传你的图片并粘贴下方提示词。
复制模型给出的描述,将其粘贴到本工作流的 Prompt 字段。
……不过,这些服务的私密性并不理想,并且可能会审查低俗/NSFW 类请求。这也是你或许想尝试其他两种方案的原因。
付费云端方案
有很多服务提供云端模型访问,这是获取未审查模型的更可靠方式。
例如,你可以在 OpenRouter 购买点数。就我个人而言更偏好 Featherless,因为它按月固定收费、成本可预期,而且有严格的“无日志”政策。如果你想试试,也可以使用我的推荐链接来支持我!
如果你选择 API/付费云路线,我的应用 CloudInterrogator 可能会对你有用。它旨在尽可能简化云端视觉模型的提示流程,而且完全免费开源!
本地推理方案
我知道 CivitAI 上有不少“只用本地”的用户。你可以选择 Ollama。
这里有我能找到的最佳安装指南。你可以关注 Google 的 Gemma-3 模型家族,并选择与你显卡匹配的规模。
如果使用 Ollama,你完全可以把 CloudInterrogator 当作访问入口,因为 Ollama 提供 OpenAI 兼容的端点;或者你也可以为 ComfyUI 加上 Ollama 节点来定制本工作流。除非你能把提示锁定,否则我并不推荐后者。
许多 WAN 工作流会把 Gemma3/Ollama 节点直接内置进去。我选择不这样做,因为我认为 99% 的人用 Gemini 或 ChatGPT 就已经足够。
建议的提示词:
分析该视频帧的内容,用一个简洁的单段落描述你对随后的整段视频序列中将发生哪些运动的预测。
你的描述应覆盖角色与场景的整体细节,但只限于与场景中“运动”相关的部分。另外,请记录粒子的运动、如果有的话眼睛的眨动、头发的摆动……这是一个被时间定格的瞬间,你要描述的是这张图像所涵盖的这几秒内发生的事。凡是可能运动的,都在运动——包括场景中微小的细节。
不要描述“停顿”。不要用“轻微”“细微”这类词来弱化运动。不要使用隐喻性语言。你的描述必须直接而明确。使用简单、常用的语言。要具体,说明场景中每个细节是如何运动的,但不要冗长;你写下的每个词都要有用处。使用现在时,好像你的预测在你输入时正在成真。
你将输出一个段落,不包含任何额外信息,也不要使用会改变格式的特殊字符;避免用“图像序列描绘了角色……”之类的说法,直接描述发生了什么,不要说“视频……”。根据你所用的模型或目标,你也许会发现 AmazingSeek 工作流提供的提示词同样好用!
技巧与故障排查
WAN 帧率:WAN 2.2 为 24 fps。若尝试 WAN 2.1,请将 fps 设为 12。模型加载节点附近有对应滑块。工作流会基于该数值自动计算帧率相关流程(倍增与重采样)。
接缝看起来不对?试试在“简单/进阶”接缝插值之间切换;在进阶模式中增加自动裁剪搜索范围;或用略微不同的提示/CFG 重新渲染。
显存不足?
在 WanVideo Decode 节点降低 tile 尺寸(x 和 y)。
降低放大(Upscale)的 tile 尺寸和/或批大小。
减小工作尺寸或时长。
启用“Use Tiled Encoder”。
致谢
最初试验时,我想到在循环接缝处做插值可能解决“脏帧”问题,但真正让我决定上手的是 AmazingSeek 的这个工作流。
看起来 Ekafalain 也应在此获得一些认可,AmazingSeek 的无缝循环工作流是基于其成果之上的。
虽然我最终没有直接采用他们的想法,但仍想致敬 Caravel——他们面向 WAN 2.2 14b 的多步流程非常出色,你可以在这里查看,文档水准就值得称赞。
我推荐的视觉提示是基于 NRDX 的版本改写而来。你可以在他 Patreon 上找到原始工作流。他也是为多种 WAN 模型训练 LiveWallpaper LoRA 的那位!
附言 💖
如果这个工作流对你有帮助,我很想看看你的作品!我为此投入了大量精力,包括设计自定义节点把一切串起来,并尽量详细地撰写文档,以便它对你尽可能有用。
链接
若想更好地了解如何用这些自定义节点搭建,或如何微调本工作流,请查看 WhiteRabbit 仓库中的节点文档与原子工作流。
个人网站与社交:在 artificialsweetener.ai 查看我的艺术、诗歌及开发动态
请我喝咖啡:在我的 Ko-fi 页面支持更多类似项目
本工作流献给我挚爱的 Cubby 🥰
Description
WhiteRabbit InterpLoop v1.1
Changelog
Removed Fancy Seam Interpolation. No longer necessary.
- As a result, the workflow is now much more consistent at creating looping videos.Fixed a bug where sometimes the target last frame would be injected second-to-last
- This is probably still technically possible depending on the duration hint you set. A more robust fix is in the works and requires a custom node. However, if you stick to duration steps of 0.5 it should not bug out.No longer necessary to release a hardline version thanks to changing from group bypass to group mute. Thank you Pixel_Music_Ai for the bug report that lead to this realization.
- The workflow now looks as was always intended!
Acknowledgements
Thank you to Pixel_Music_Ai for the detailed bug report and valuable feedback.
FAQ
Comments (82)
Can you please give me a hint about what I should change to make this work with 2.2 14B?
If I knew for sure I'd have made it already. I just haven't played with 14b enough to know. My understanding is that 14b uses a low noise and high noise model. You need to pipe each model into their own sampler nodes and process the frames through each one.
That's my guess!
Incredible workflow. Works flawlessly for me after following your steps. Have been using 14B but I have to say that 5B is a hidden little gem and works significantly faster while maintaining incredible accuracy and quality.
Please keep making more workflows, you're clearly talented at this.
That's very sweet of you to say. I'm not sure if I will make more workflows per-se, I am busy creating a brand new front-end for ComfyUI :]
Hi, do you think you can add Florence2 node to describe the picture input?
I am going to test this on my side.
Thank you for this workflow :)
I'm glad you like the workflow!
The main reason I don't have any kind of describe/interrogate model wired into this workflow is because of the extra resources and time it can take. If you're running the same image through a few times, tweaking settings and the prompt as you probably should do, it just doesn't make sense to regenerate the prompt over and over every time you run the workflow and every time you do it takes a lot of time for a process that already feels agonizingly long coming from image gen.
I haven't looked at Florence2 specifically, maybe it addresses some of the problems I'm talking about. I will look. It's hard to imagine, though.
Invalid T5 text encoder model, fp8 scaled is not supported by this node. How do I resolve this issue?
Forgive me, I have given you bad advice. You need this text encoder to use FP8, not the _scaled version.
https://huggingface.co/Kijai/WanVideo_comfy/blob/main/umt5-xxl-enc-fp8_e4m3fn.safetensors
Thank you for finding this bug in my instructions. I will update it ASAP so other people don't have this problem.
@ArtificialSweetener_ Super thanks for your explanation and your workflow—they're super awesome! Also, have you considered creating a workflow based on Rapid WAN 2.2?
@cybercosplay1 I don't even know what that is! I have been experimenting with WAN 14b with the Lightning LoRA.
Dude u should do youtube videos, explaining how to use this... probably u would get some money from donations ! it look amazing, but u do better videos, and mines are not that cooll !
I did my best to explain in the documentation. Videos would be nice but I am very busy!
Genning is always gonna be a trial and error process to find what works for what you're trying to do. Read the settings doc and try playing with things like Schedule Shift, CFG, and LoRA Strength if you're not getting what you want. It can also help to tweak the vision model prompt (the one you send to ChatGPT/Gemini/Gemma) to steer it towards stuff you want to see.
The prompt has a huge impact on what you'll get and you should also be looking at tweaking it to put the effects that are most important to you first in the prompt. That's why all mine start with blinking, because I wanted loops that look natural instead of a staring contest.
Hey! Thanks for providing your workflow
Is it possible to use lightx2v lora to get a loop video from 4 frames?
also my results get a lot of glitching, I see that your showcases don't, do you mind share some suggestions? I read all the notes in the workflow too
I'm not aware of lightx2v train for 5b. Maybe it exists, I just don't know. If it does, you could use it, just copy the LoRA loader, point it at the new LoRA, and connect that to the LiveWallpaper LoRA loader in the chain; the WanVideoWrapper nodes make it pretty easy to chain LoRA.
What kind of glitches do you mean? It's hard to know what you're talking about without examples. I could more easily diagnose the problem if I could see.
@ArtificialSweetener_ sure https://civitai.com/images/101361032
you can see that on her hair there is an artifact that shows for a sec and I noticed that it happens on all of my generations, not exactly hair but at some moment it glitches
@korzhiktolikbaev443 Please try with model precision set to FP16 instead of FP16_fast. In my tests, lowering precision in the loaders can cause artifacting exactly like that.
great workflow!!! incredibly useful :)
Glad you find it useful!
There seem to be a memory leak. After first 1-2 generations I flushed my memory in ComfyUI (2 icons at the top "unload model" and "free cache") and I still have 8 GB of ram taken (30GB of RAM was used when generating).
After 1-2 more generations I started to have 10 GB of memory taken (again, that's after flushing everything) and now it takes 40 GB of RAM when generating.
Also, VRAM management seems off. I offload all the models to RAM and with 800x800@12 fps this workflow takes up way too much VRAM, it basically has 0 overhead on 3080ti 12GB (it used 11.6 GB). While regular Wan 2.2 workflow (with full fp16 14b models) only takes up ~10.7 GB with 832x832@16 fps.
(I did a lot of tests with WAN and without ~1.2 GB overhead in VRAM my generation speeds tank a lot.)
So basically yes, I can't comfortably use this workflow. It even sometimes starts to store latents in shared video memory and it makes everything super slow. Default ComfyUI WAN workflow gives a error and halts generation when this happen.
I don't know what I could do differently. I've tested this on 12gb cards (3060) and didn't run into the same issues after adding the VRAM/cache clearing nodes.
My best guess is that it's a problem with WAN Video Wrapper nodes. If you know of some way to mitigate this on the graph side I'd be happy to implement.
@ArtificialSweetener_ Are you on Linux? Linux doesn't have shared VRAM feature which Windows has. So your latents PHYSICALLY can't seep into RAM.
The point is, you don't have these problems because you are on Linux, maybe a bunch of stuff behaves differently for you.
@shifty_13 I'm on Windows.
@shifty_13 Turn off Sysmem Fallback Policy in Nvidia Control Panel. It's poorly supported by Torch and well known to cause strange behavior and slowdowns. This must be what you're talking about.
You don't have to do it universally if you want it to still work for games. In that case, turn off Sysmem Fallback Policy for the Python interpreter.
I have noticed the same thing, my Shared GPU memory is very high while Dedicated GPU memory is very low when running RIFE VFI FPS Resample node. And the result is cumulative. I guess that's why you start to get problem after 1-2 more generations. BTW Are you able to solve the issue after turn off System Fallback Policy for the Python interpreter?
I'm discovering your workflow, I really like it! Good job :)
I had a lot of trouble getting it to work at first times haha... but I solved my problem with text encoders and model diffusion in fp8
I would like to know how to generate multiple videos at the same time? The batch size
Yes, I need to update to stop recommending the wrong text encoder. My apologies.
I don't know about changing batch size. Even on my 4090 I'm not sure I could batch up more than one video at a time because of VRAM constraints. I'm not even sure if WAN supports batch as a dimension! It would be in the encoder node.
@ArtificialSweetener_
Okay, no problem, because I basically use Wan2GP, but your workflow allows for a perfect loop, which is much cleaner :)
I have one problem: when I generate a video, the animation suddenly accelerates during the first second, then becomes normal again 🤔
I tested with 24fps output and the problem is almost gone, so I think the issue comes from the frame interpolation 60fps part of the workflow. Do you have any advice on how to fix or smooth this sudden speed-up at the beginning?
@Natsu24 I noticed that speed-up too, initially. I actually tried a lot of different stuff to try to fix it, and that's the whole reason WhiteRabbit has a Timing Analyzer node.
The idea I had is that RIFE interpolates with a "model assumption" that a whole clip will get interpolated with the same timings, which means linear timestep assumptions are OK. But when you're only interpolating over 2 frames (which is what we do for the loop seam), that assumption breaks and you end up with frames at the end of the loop - our interpolated frames - with different timesteps from other frames in the clip.
But what I ended up realizing was that the timing analyzer was doing more harm than good and exacerbating the issue. After I took that out of the graph, results instantly got better and I realized I'd been over-engineering. I actually didn't notice the "speedup" in any of the clips I did after that which was kind of weird since I'd seen them before. That's a big part of what v1.1 of this workflow is.
I'd be curious to see what settings you're using, especially duration, and I'd be curious to see examples of the speed-up thing happening.
@ArtificialSweetener_
I uploaded the modified workflow with my settings and the image used. I tried both versions of Rife, but had the same problem... I also tried tinkering with ChatGPT, but without success :/
https://limewire.com/d/HmlwS#fbqFkO6blJ
@Natsu24 My first thought is to try setting duration hint to 4 or 3. WAN has a tendency to loop on itself even without first/last frame conditioning like we're doing, but the loop it creates can be unpredictable, creating two "cross-over points" that happen in rapid succession and can look like a stutter. I don't think it's guaranteed to happen, but you increase the chances the longer the clip you ask for.
If you really don't want to decrease duration, you can try messing with riflex_freq_index in the WanVideoSampler. This is specifically designed to make the model resist looping which is exactly why I have it set to 0 for this workflow. You could try raising it to 2, 4, or 6 to see if it helps with longer durations. If you find a reliable relationship between duration and this frequency index value I'd love to know about it.
You could also try increasing "trim start frames" and "trim end frames" to give more of the seam over to RIFE (the interpolation model). OR, you could try decreasing these numbers, but you'll get problems if you increase trim end frames below 4.
I should also note, I do see some stutter in your loop in the fine particles, but the movement of the hair and other stuff looks pretty good. It might be that you can't get this perfect just from an automatic run and might need to use a secondary workflow to doctor it to perfect. WhiteRabbit has a few nodes for doing exactly that but it requires working over the same clip a few times to figure out what the perfect values are gonna be.
@ArtificialSweetener_
Thank you very much, I'll try that and keep you posted
I seem to be missing a lot of non-standard nodes comfy isn't familiar with. Any help on how to get these installed?
I'm a little shocked you have a Comfy installation without Manager installed.
Install ComfyUI-Manager
https://github.com/Comfy-Org/ComfyUI-Manager
It will automatically install missing nodes for you going forward, including the nodes from WhiteRabbit which this workflow relies on.
@ArtificialSweetener_ I do have Comfy manager installed. The install missing button isn't there for some reason. I have an older installation of comfy that does have it but I'm trying to keep separate installations to avoid breaks. Also tried Install Missing Custom nodes, but nothing appears in dialog window.
@robbiebunny1883 Try installing from repository using your old Manager version:
https://github.com/Artificial-Sweetener/comfyui-WhiteRabbit
But I really can't say for sure if an old version of Comfy will support WhiteRabbit. I can't think of any reason why not but it's possible there's some problem somewhere. It was built and tested with the latest Comfy version with the latest frontend.
If there are other nodes you're missing, you'll have to track em down. I use also WanVideoWrapper and KJ-Nodes in this workflow that I remember.
Your best bet is to update Comfy!
However, I keep getting a consistent error on the WanVideoSampler node, and I can’t figure out why it happens. RuntimeError: shape '[3072, 3072]' is invalid for input of size 26214400
Getting the same error, edit: you need the correct version of the lora (5B)
@Suiai13 thank you for your reply. But now, I got another error:
RIFE_VFI_Advanced: 'str' object has no attribute 'shape'
It happens inside node #565 (Interpolate Over Seam).
Does anyone know what might be causing this or how to fix it?
fp8 scaled is not supported by this node
nvm, solved
@heywhat Yeah, it's just my instructions are bad and I recommend the wrong file. I'm sorry
good workflow but will be better if the environment requirements are clearer.
here are some errors I met and solved:
1. a node called coerce to float behind scheduler shift node causes 'Compex types (LATENT/IMAGE) need to reference their width/height, e.g. a.width' error. (bypass it)
2.remember install sageattention and triton if not already
3.probably need to install Visual studio build tolls and CUDA toolkit
4.python 3.13. which is the default embedded python in windows portable version of Comfyui, doesn't work, switch back to 3.12
I have thought of some way to force installation of Triton/Sage for people. I'm not convinced CUDA toolkit is necessary but it might be. It's something I'll have to play with.
I'll look at the coerce to float node but I'm surprised. If you can send me a workflow with this problem it would help a lot.
When I generate the video, the character's eyes are blurry. The character is fine, only the eyes are not clear.
How can I fix this?
I have seen the same problem when the eyes are a small detail in the original frame you input. The way I work around this is to use detailed portraits of characters where the eyes take up a good amount of space in the composition. In the examples I put up, these are crops of high resolution portraits.
Remember that our WAN model prefers ~704p input, so it might help to pre-process your image by cropping in on a region that is about that size (for example, 1280x704). The runs where I had bad eyes were often coming from detailed, high res images that had to be scaled to match the model's constraints first.
Hope that makes sense!
I'm using this for 2D animations for a game. Very flawless, got rid of the last frames flashing, very smooth and with a background remover node it looks great on the device.
My issue is that the animations are very slow and floaty.
When using another workflow, they're much faster. Same wallpaper Lora, but a WAN 2.2 model for faster animations.
Is there a general trick to make fast animations? Something with the prompt, maybe disable interpolation or so?
I stitched this workflow together with another one for faster movement and it doesn't seem to work well anyway. The interpolation for the cut off last 4 frames isn't strong enough to overcome the missing frames for fast movement.
@hellosir Look at the strategy in v1 with the timing analyzer. That may help you in terms of looping faster animations. In terms of reducing floatyness, I recommend lowering the LoRA weight. The floaty effect is coming from the LoRA in my experience.
@ArtificialSweetener_ I've tested some more.
Lora strength does little to nothing for animation speed (nor does any other setting). I think WAN 2.2 TI2V 5b just isn't suited for fast animations. I have to use a 2.2 14b high/low specifically remixed for animation speed to get actually fast animation (that Smooth one).
I can get a result good enough with your 1.1 seam (attached to a high/low workflow) and no final interpolation (it exaggerates any imperfections). I just have to watch out what animations I make. Some of them interpolate well enough. Others you don't notice it much. Like don't make a fast looping idle, but if the character punches someone the player is distracted enough to not notice little imperfections.
Either way, your custom seam node is the best in terms of loops there is.
@hellosir Yes, I have tested 14b a little bit and noticed it often pulls out more motion than 5b. It's not always about the speed but also the ability for high frequency details to animate.
I wonder if there's a problem with my FPS resample node. Can you try multiplying the framerate without use RIFE FPS Resample (just RIFE Multiply) and see if that has any of the glitches you were talking about?
@ArtificialSweetener_ I've added a preview after every step. The roll & multiply messes up the frames. The video now starts 12+1 frames late (the start of the video is shifted. But other than that it seems fine. Just not fine for the workflow I needed, which is always starting+ending with the original image (so I can switch between animations seamlessly in Unity).
If the frame shift is intentional (perhaps it helps with interpolation), then it works fine. A simple anime wallpaper doesn't really care about which frame it starts with.
I did something silly by simply splitting off the last 4 frames, reduce gamma+highlight by 10%, color match and then merging them back in. No interpolation of any kind.
Works really well on like 9 out of 10 images (the images need enough contrast it seems).
@hellosir Did you change the base framerate value? It's a slider in the top left. That's the first thing I'd guess would be doing that to roll+multiply. If you can upload a .json someplace (pastebin works) of the workflow that has this issue I'd love to have a look and see.
Oh my god finally!!! "Why is this so complicated?!" is exactly what I've been saying for the past 2-3 weeks of trying to figure out how to fix this! I tried so many various workflows and ways, every single sheduler, models, vae. Pure insanity! Even the triple workflow from Caravel using Wan Vace was pretty hard to align both videos together without fading issue and took 10x longer to generate.
The fact you had to create custom node just to fix this issue is crazy! Thanks so much for going above and beyond! I'll have to replace all my workflows with your optimised nodes as they are indeed blazing fast.
I wonder what exactly you've optimised in your nodes to actually fix the color issue. I am surprised you're still using the WanVideoWrapper as I thought that was the entire issue.
Wan 2.1 had no issue, but Wan 2.2 went overboard, mostly with the WanVideoWrapper nodes. Kept adding color contrast over time, destroying the seamless loop illusion. I recently tried to go back to Wan 2.1 and I had some gray out color issue and was super stuck. The only okay solution I found was to use regular double KSamplers, but the motion was meh, WanVideoWrapper was best for motion, but worst for color.
Your workflow didn't show up when searching for "loop" in civitai. I just ended up going back to see the live wallpaper lora and saw your comment. Perhaps separating the title from "InterpLoop" to have "loop" on itself would help with the search.
Since I also need to animate with various loras in Wan 2.2 14b, I'll have to edit your workflow a bit. Add a High/Low + stackable loras "Power Lora Loader". Really love how you added tons of notes around for optimisations.
From trying to add the High/Low models and a double Sampler for them, it seems I'm getting an error that is commonly used for not using the correct model combination (ex i2v + t2v), but I have changed both model to i2v 14b. Not sure you have a specific node that this workflow only allow for 5b model or I am not finding the correct area that I need to modify?
"WanVideoSampler
Given groups=1, weight of size [5120, 36, 1, 2, 2], expected input[1, 16, 18, 124, 84] to have 36 channels, but got 16 channels instead"
Depends on what color issue you mean. My nodes don't do color matching but this workflow does, using KJ's color matching node iirc. That's separate from the dirty frames issue, which I do explain pretty thoroughly in the docs I think?
The Power LoRA loader doesn't work with KJ's WAN nodes iirc so you'd have to swap out the nodes that load models and do inference with the vanilla Comfy nodes for that. Not impossible but I don't know if you can inject frames in 2.2 with those.
Thanks for the SEO suggestion. You're probably right.
@Catz Make sure you're using the 2.1 VAE when you're on 14b. Make sure you're using the 14b LoRA, too. Those are the main size mismatch error creators, without actually looking too closely at your specific error I'd be willing to bet it's between those two. Let me know if either of those is ruled out.
I got another error:
RIFE_VFI_Advanced: 'str' object has no attribute 'shape'
It happens inside node #565 (Interpolate Over Seam).
Does anyone know what might be causing this or how to fix it? 🙏
Please upload your workflow to pastebin or somewhere else and link it here and I'll look at it.
@ArtificialSweetener_ Thanks. Here is the workflow. https://pastebin.com/kRJKdCbu
Hi @ArtificialSweetener_ just checking if you had a chance to look at the workflow. Still getting the same error. Thanks!
你好! 运行后提示错误 RIFE_VFI_Advanced: 'str' object has no attribute 'shape' , 检查后发现 comfyui-WhiteRabbit 插件似乎已经升级,造成无法找到
Interpolate Over Seam 节点 和 其它相关节点, 期待新的工作流版本
Does this work well with 12gb cards?
Any chance you can create a workflow for WAN 2.2 14B i2v?
I keep getting errors for the whiterabbit nodes. I followed the instructions and downloaded the folder to the proper location but I still get the same error
try this - some nodes have a requirements.txt that install dependencies
8gb vram works, it takes about 500s to get 3s 702x702 video. It looooops!
Opened this workflow and immediately got lost in the node spaghetti 😅 Where do I even start?
what do you mean bro it's really use. just go to top left, everything you need is placed there.
I have a question! Is this normal?
got prompt
Failed to validate prompt for output 521:
* (prompt):
- Required input is missing: anything
* easy cleanGpuUsed 521:
- Required input is missing: anything
Output will be ignored
Im getting a "cannot access local variable 'noise_pred_uncond' where it is not associated with a value" error. Setting the CFG value to 1 bypasses this error, anything greater than that causes it. CFG value of 1 however doesn't make good videos.
Any ideas?
I have spent at least 14 hours trying to get this working but behind every bug is yet another bug
This workflow requires sageattention, which I cannot use with my python environment (3.11) that I am using for other AI. Is there a workaround for this within comfyui?
I had exact same problem, I just changed the attention_mode to sdpa in the WanVideo model loader node, then I turned on bypass on the WanVideo Torch Compile Settings node to disable it and it works fine after that. I hope this helps
@greapergod822 Thanks for your help! I was able to get past that, but now I have this error at the Wan video sampler node:
cannot access local variable 'noise_pred_uncond' where it is not associated with a value
I'm not sure if you have an answer to this, just the only info I can find is to just update everything and that isn't working.
@Wiggler6788 I never had that problem, I did have a different error in the sampler after messing with things trying to figure out the sageattention issue, so I loaded an untouched workflow, clicked on and checked to make sure the model, text encoder, and vae are set to the ones I downloaded, I use wan2.2_ti2v_5B_fp16.safetensors, umt5_xxl_fp16.safetensors and of course the wan2.2_vae.safetensors. then I just changed the attention_mode to sdpa in the WanVideo model loader node, then I turned on bypass on the WanVideo Torch Compile Settings node and ran it and it worked perfectly, I wish I knew more to help I am just messing around with it same as you lol
I finally got it to work, but it produces a 3s video with no movement at all. It's just a freeze of the initial image, with a short fade to white at the end.
Anyone had the same issue and a fix?
Prompting seems to be very important
Yeah, I tried longer prompts, but nothing. My only success was making a star in the background move slightly - basically a white pixel :)
I moved to a 14B workflow now and borrowed the "Interpolate over Seam" part which is very cool.
@artyclaw how did your creation go?
@loneillustrator Great! I don't want to advertize another workflow here, but there's some output in my posts :)
Have you heard saying "less is more"? :)
It could be very good if it wasn't so complicated