CivArchive

    🌀 Wan2.1_14B_FusionX — Merged models for Faster, Richer Motion & Detail in as little as 8 steps!

    📢 7/1/2025 Update!

    New: FusionX Lightning Workflows

    Looking for faster video generations with WAN2.1? Check out the new FusionX_Lightning_Workflows — optimized with LightX LoRA to render videos in as little as 70 seconds (4 steps, 1024x576)!

    🧩 Available in:
    • Native • Native GGUF • Wrapper
    (VACE & Phantom coming soon)

    🎞️ Image-to-Video just got a major upgrade!!!!!!
    Better prompt adherence, more motion, and smoother dynamics.

    ⚖️ FusionX vs Lightning?
    Original = max realism.
    Lightning = speed + low VRAM, with similar quality using smart prompts.

    👉 Check it out here


    ☕ Like what I do? Support me here: Buy Me A Coffee 💜
    Every coffee helps fuel more free LoRAs & workflows!


    📢 Did you know you can now use FusionX as a LoRA instead of a full base model?
    Perfect if you want more control while sticking with your own WAN2.1 + SkyReels setup.

    🔗 Grab the FusionX LoRAs HERE
    🔗 Or Check out the Lightning Workflows HERE for a huge speed boost.


    📌 Important Details- Please read the full description below because small changes to settings will provide totally different results in a bad way! I have been testing and already found better settings so just please read below! Thank you :)

    💡Workflows can be found HERE (This is a wip and more will be added soon.)

    🛠️Updates section has been moved to the end of the description.

    A high-performance text-to-video model built on top of the base WAN 2.1 14B T2V model — carefully merged with multiple research-grade models to enhance motion quality, scene consistency, and visual detail, comparable to some of the many close source models.

    ## 📢 Join The Community!

    A friendly space to chat, share creations, and get support.

    👉 Click here to join the Discord!

    Come say hi in #welcome, check out the rules, and show off your creations! 🎨🧠

    💡 What’s Inside this base model:

    • 🧠 CausVid – Causal motion modeling for better scene flow and dramatic speed boot

    • 🎞️ AccVideo – Improves temporal alignment and realism along with speed boot

    • 🎨 MoviiGen1.1 – Brings cinematic smoothness and lighting

    • 🧬 MPS Reward LoRA – Tuned for motion dynamics and detail

    • Custom LoRAs (by me) – Focused on texture, clarity, and fine details. (These both were set to very low strengths and have a very small impact)


    🔥 Highlights:

    • 📝 Accepts standard prompt + negative prompt setup

    • 🌀 Tuned for high temporal coherence and expressive, cinematic scenes

    • 🔁 Drop-in replacement for WAN 2.1 T2V — just better

    • 🚀 Renders up to 50% faster than the base model (especially with SageAttn enabled)

    • 🧩 Fully compatible with VACE

    • 🧠 Optimized for use in ComfyUI, With both the Kijai Wan Wrapper and native nodes.


    📌 Important Details for text to video:

    • 🔧 CGF must be set to 1 — anything higher will not provide acceptable results.

    • 🔧 Shift - Results can vary based on Resolution. 1024x576 should start at 1 and if using 1080x720 start at 2. Note: For more realism lower shift values is what you need. If your looking for a more stylized look then test higher shift values between 3-9

    • Scheduler: Most of my examples used Uni_pc but you can get different results using others. Is really all about experimenting. I noticed depending on the prompt that the flowmatch_causvid works well too and helps with small details.

    📌 Important Details for image to video:

    • 🔧 CGF must be set to 1 — anything higher will not provide acceptable results.

    • 🔧 Shift - For image to video I found that 2 is best but you can experiment.

    • Scheduler: Most of my examples used dmp++_sde/beta and seems to work best but you can experiment.

    • After testing, to get more motion and reduce the slow-mo look, set your frame count to 121 and frames per second to 24. This can provide up to a 50% motion speed boost.

    📌Other Important Details:

    • Video generation works with as few as 6 steps, but 8–10 steps yield the best quality. Lower steps are great for fast drafts with huge speed gains.

    • 🧩 Best results using the Kaji Wan Wrapper custom node:
      https://github.com/kijai/ComfyUI-WanVideoWrapper

    • 🧪 Also tested with the native WAN workflow, generation time is a bit longer but results match wrapper.

    • Do not re-add CausVid, AccVideo, or MPS LoRAs — they’re already baked into the model and may cause unwanted results.

    • 🎨 You can use other LoRAs for additional styling — feel free to experiment.

    • 📽️ All demo videos were generated at 1024x576, 81 frames, using only this model — no upscaling, interpolation, or extra LoRAs.

    • 🖥️ Rendered on an RTX 5090 — each video takes around 138 seconds with the listed settings.

    • 🧠 If you run out of VRAM, enable block swapping — start at 5 blocks and adjust as needed.

    • 🚀 SageAttn was enabled, providing up to a 30% speed boost. (Wrapper only)

    • Workflows for each model can be found here: HERE

    • 🚫 Do not use teacache — it’s unnecessary due to the low step count.

    • 🔍 “Enhance a video” and “SLG” features were not tested — feel free to explore on your own. -- Edit. I did test "Enhance a video" and you can get more vibrant results with this turned on. Settings between 2-4. Experiment! SLG has not been tested much.

    • 💬 Have questions? You’re welcome to leave a message or join the community:

      • 👉 Click here to join the Discord!

    • 📝 Want better prompts? All my example video prompts were created using this custom GPT:
      🎬 WAN Cinematic Video Prompt Generator
      Try asking it to add extra visual and cinematic details — it makes a noticeable difference.


    ⚠️ Disclaimer:

    • Videos generated using this model are intended for personal, educational, or experimental use only, unless you’ve completed your own legal due diligence.

    • This model is a merge of multiple research-grade sources, and is not guaranteed to be free of copyrighted or proprietary data.

    • You are solely responsible for any content you generate and how it is used.

    • If you choose to use outputs commercially, you assume all legal liability for copyright infringement, misuse, or violation of third-party rights.

    When in doubt, consult a qualified legal advisor before monetizing or distributing any generated content.


    ### 🧠 More GGUF Variants

    - 🖼️ [FusionX Image-to-Video (GGUF)]

    - 🎥 [FusionX Text-to-Video (GGUF)]

    - 🎞️ [FusionX T2V VACE GGUF (for native)]

    - 👻 [FusionX Phantom GGUF (for native)]


    ###🧠 fp16 Versions can be found here:

    -- 🖼️fp16 FusionX Models



    📌gguf comparisons!
    I'm slowly adding to this list, but here you can see how the models compare against the main model.

    Text to video:

    --------

    🛠️Update 6/8/2025 - Image to video model is published! Settings that I use in the example videos: Steps = 10 / cfg = 1 / shift = 2 / schedular = dmp++_sde i'll post a WF soon.

    🛠️Update 6/7/2025 - Published a i2v phantom model that can take up to 4 reference images and combine them into a video. Posting workflow soon

    Phantom WF is getting uploaded soon.

    🛠️Update 6/6/2025 - Added a new gguf model! If you want the highest quality and have enough VRAM get the V1.0 model otherwise gguf is the next best thing! When using the gguf's it will take longer to generate even on an RTX 5090.

    Description

    FAQ

    Comments (32)

    9ballJun 5, 2025
    CivitAI

    Hi, What are the minimum requirements to run this?

    vrgamedevgirl
    Author
    Jun 5, 2025

    It really depends on your settings. A 3080 Ti with block swapping enabled should work well, but even lower VRAM GPUs might be able to handle it if you increase the block swapping significantly.

    What GPU are you working with? Feel free to share — I might be able to give more specific advice.

    Also, I highly recommend joining the Discord server (linked in the description). There’s a lot of helpful discussion and support from other users there!

    5327045Jun 5, 2025· 1 reaction
    CivitAI

    Been testing it,gotta say this is amazing. Is this the same as MasterModel?

    vrgamedevgirl
    Author
    Jun 5, 2025· 1 reaction

    Thank you so much!!!! And yes, the "MasterModel" was a placeholder name :) Everyone wanted to test it so I put it out fast and had not thought of a name yet. lol

    LDWorksDavidJun 6, 2025· 3 reactions
    CivitAI

    Congrats!

    vrgamedevgirl
    Author
    Jun 6, 2025

    Aww thank you :)

    LDWorksDavidJun 6, 2025

    @vrgamedevgirl BTW, this works with VACE too right?

    vrgamedevgirl
    Author
    Jun 6, 2025

    @LDWorksDavid It does :)

    Epsilon2Jun 6, 2025· 4 reactions
    CivitAI

    Hello, is there an image to video version of this model planned?

    vrgamedevgirl
    Author
    Jun 6, 2025· 8 reactions

    I can try and create a I2V version and if it works out I'll post it here! :) stay tuned!!

    Epsilon2Jun 6, 2025

    @vrgamedevgirl Thank you very much for the info!

    jonk999Jun 7, 2025· 1 reaction

    Keen for an I2V version as well. I do more I2V than T2V as prefer to work with an image I know is what I'm after.

    vrgamedevgirl
    Author
    Jun 7, 2025· 1 reaction

    @jonk999 I did play around with this. Right now, when using the new AccVid i2v model you get really good results. Adding the other lora's don't really make that much of an impact. All you really need to do is add the caudvidV2 and AccVid lora and its golden. I mean, if no one wanted to have to deal with lora's I could create the merge. I was going to try and train some i2v loras on different camera angles and character movement that I could then bake into it but it will take some time to do that. :(

    jonk999Jun 7, 2025· 1 reaction

    @vrgamedevgirl Thank you for the reply. I've played around a bit with having caudvid in a flow and from memory had issues when including other loras it seemed to lose a lot of the movements. Pretty sure I tried another flow that used both and it wasn't really much faster than a flow I was using already. I'll have to take another look as it might have been due to models being used (I only have a 3060 12GB VRAM and 32GB RAM). So one may have been using GGUF and the other a 'regular' model.
    One thing I've noticed in using the flow embedded in one of your GGUF videos (the first one of the woman in the forest at winter) is when I add a Load Lora node and a lora file I get a lot of the messages similar to: "lora key not loaded: diffusion_model.blocks.6.cross_attn.k_img.lora_B.weight", but I think the Lora still works. I got it in another flow I tried as well, but was never sure why it happened. The Lora is marked as being for T2V.

    PrecariousDuckNoiseJun 6, 2025
    CivitAI

    Could you share the workflow used in your example videos? I'm starting out using Wan for the first time and am getting different results!

    vrgamedevgirl
    Author
    Jun 6, 2025

    Can we chat via discord? Then I can share files. My discord name is: vrgamedevgirl

    syntaxdiffusionJun 6, 2025· 1 reaction
    CivitAI

    loading this up meow! great work!

    vrgamedevgirl
    Author
    Jun 6, 2025

    Thank you! Don't forget to read through the description. Lots if important details on settings. Results vary sooo much depending on settings. I am still testing and coming across better settings. LOL have already updated a few video's because I had the wrong shift value and a tiny change made a huge difference.

    pushpiblJun 6, 2025
    CivitAI

    这个是基于t2v融合是吧,那也就是可以和vace联动?有可能制作成占用空间较小的lora吗

    vrgamedevgirl
    Author
    Jun 6, 2025

    是的,这是基于 WanT 的 T2V 14B 模型,我把合并后的版本命名为 FusionX
    它确实可以和 VACE 联动!我尝试过制作一个 LoRA 版本,但没有成功,还没找到实现的方法。
    建议看一下完整说明,里面有很多详细信息!

    RealHUGEJun 6, 2025· 1 reaction
    CivitAI

    I hope for a gguf version 😁

    vrgamedevgirl
    Author
    Jun 6, 2025· 2 reactions

    Someone created an unofficial GGUF that can be found here. https://huggingface.co/lym00/Wan14BT2V_MoviiGen_AccVid_CausVid_MasterModel_GGUF

    This was before I published to CivitAI so it still had a placeholder name of just "MasterModel" but is the same model. I have not tested these so cannot guarantee results. -When I have to time to them I can post them in here I think if I get permission from who created them.

    lym0Jun 6, 2025· 1 reaction

    @vrgamedevgirl Hi, GGUF creator here, feel free to post the GGUFs. If there’s an FP32/FP16 model or a new merge (e.g., a VACE add-on patch like https://huggingface.co/QuantStack/SkyReels-V2-T2V-14B-720P-VACE-GGUF?show_file_info=SkyReels-V2-T2V-14B-720P-VACE-F16.gguf), I’d be happy to help with the GGUF quants too.

    vrgamedevgirl
    Author
    Jun 6, 2025· 2 reactions

    @lym0 I'm downloading the fp16 version to hugging face as we speak :)

    vrgamedevgirl
    Author
    Jun 6, 2025· 1 reaction

    @lym0 When it comes to VACE this one works with VACE and its really good. Unless you mean something else? let me know contact me on discord! I'm guessing you found my hugging face link from the discord server? let me know :)

    lym0Jun 7, 2025

    @vrgamedevgirl Hey yeah, I’ve done some testing with VACE compatibility and noticed a few things:

    1) Tried it with my VACE workflow (linked here), and the VACE i2v seemed to lose likeness, the output didn’t look much like the source reference image. Another redditor shared the same observation: Reddit comment.

    2) On the other hand, the GGUFs from QuantStack (skyreels-v2-vace, moviigen-vace) don’t seem to have that issue.

    3) According to the QuantStack team, they added “2 different VACE scopes” while processing the safetensors: Reddit comment.

    4) I've also compared GGUFs of yours with theirs and noticed their version includes additional tensors keys like vace_patch_embedding and 8 extra vace_blocks (metadata).

    vrgamedevgirl
    Author
    Jun 7, 2025

    @lym0 Can we chat on discord? vrgamedevgirl Send me a message! :)

    lym0Jun 7, 2025· 1 reaction

    @vrgamedevgirl sure, but kinda tied up with some errands today, so I might be a bit slow to respond.

    On another note, I just tried reaching out to the QuantStack team on Reddit, to ask how they patch the safetensors: reddit_link. Hopefully they’ll share some insights.

    RealHUGEJun 7, 2025

    @vrgamedevgirl Nice! You should include this gguf model here on civitai, many people will download it

    vrgamedevgirl
    Author
    Jun 7, 2025· 1 reaction

    @RealHUGE Its uploading as we speak! :)

    lym0Jun 7, 2025· 2 reactions

    Update: new GGUFs uploaded to https://huggingface.co/lym00/Wan14BT2VFusionX_fp16_GGUF, based on the fp16 model.

    lym0Jun 8, 2025· 1 reaction

    Re: VACE Support

    We'll upload a separate set of GGUFs specifically for VACE support, addressing the likeness issue when using the native ComfyUI node (WanVaceToVideo): https://huggingface.co/QuantStack/Wan-14B-T2V-FusionX-VACE-GGUF