Version 3.0 has been updated to use LTX 2.3
LTX 2.3 should provide some enhancements such as sharper video
Better audio
Etc..
The normal models are HUGE, can't be used by most people.
This workflow used GGUF for model and text endcocder, gemma.
Trying to keep footprint small, but its getting hard to do.
**************LTX 2.3 VERSION*************
https://huggingface.co/Lightricks/LTX-2.3/tree/main (upscaler)
https://huggingface.co/QuantStack/LTX-2.3-GGUF/tree/main/LTX-2.3-distilled (main gguf)
https://huggingface.co/Kijai/LTX2.3_comfy/tree/main (vae, text projector)
./models/text_encoders
gemma-3-12b-it-Q2_K.gguf (this is only 4GB for low vram)
ltx-2.3_text_projection_bf16.safetensors
./models/unet (DISTILLED version, distilled only needs 8 steps)
LTX-2.3-distilled-Q5_K_S.gguf (Distilled version provides benefits for low vram)
./models/vae (2.3)
LTX23_audio_vae_bf16.safetensors
LTX23_video_vae_bf16.safetensors
./models/latent_upscale_models (2.3)
ltx-2.3-spatial-upscaler-x2-1.0.safetensors

Models are in a Subgraph.


Description
Added 2nd workflow for SINGLE I2V video (basic)
removed "LTX2 Chunk Forward" Node, it was causing memory leaks.
cleaned up a few other minor things.
FAQ
Comments (2)
can this do over a minute without losing too much in quality or likeliness?
not possible. each iteration loses some quality. Sometimes it works better than other times depending upon how the last frame was rendered from previous clip.