Resources you need:
πFiles :
For base version
I2V Model : wan2.1_i2v_480p_14B_fp8_e4m3fn.safetensors or wan2.1_i2v_720p_14B_fp8_e4m3fn.safetensors
In models/diffusion_models
CLIP: umt5_xxl_fp8_e4m3fn_scaled.safetensors
in models/clip
For GGUF version
>24 gb Vram: Q8_0
16 gb Vram: Q5_K_M
<12 gb Vram: Q3_K_S
I2V Quant Model : wan2.1-i2v-14b-480p-QX.gguf or wan2.1-i2v-14b-720p-QX.gguf
In models/diffusion_models
Quant CLIP: umt5-xxl-encoder-QX.gguf
in models/clip
CLIP-VISION: clip_vision_h.safetensors
in models/clip_vision
VAE: wan_2.1_vae.safetensors
in models/vae
ANY upscale model:
Realistic : RealESRGAN_x4plus.pth
Anime : RealESRGAN_x4plus_anime_6B.pth
in models/upscale_models
π¦Custom Nodes :

Description
Interface adjustment :
The frames slider is replaced by a duration slider in seconds,
removed the interpolation ratio slider,
All models files are now in the main window.
Backend :
reduction of the number of custom nodes from 12 to 8,
improvement of the automatic prompt function with the replacement of 8 words like "image" or "drawing" by video to avoid making static videos,
added clip loader in GGUF version.
New "Post-production" menu.
Rollback on native upscaler.
New model optimisation :
Temporal attention for improve spatiotemporal predictive.
RifleXRoPE reduce bugs on videos longer than 5s. This allows you to increase the maximum video length from 5s to 8s.
More detail here : link




