Welcome to my π«π¦ Friendly LTX-2.3 T2V+I2V+Lipsync
β¨ Less mess, more magic
Unified Light - Lipsync all-in one version with single-stage video generation
Unified is cooming soon...
LTX-2.3 is a state-of-the-art DiT-based (Diffusion Transformer) audio-video foundation model developed by Lightricks. It represents a significant evolution from LTX-2, delivering enhanced audio and visual quality alongside improved prompt adherence. The model is specifically designed to generate synchronized video and audio within a single unified architecture, making it a powerful tool for multimodal content creation.
I offer my unique workflow with convenient options control and all-in-one structure (audio, t2v, i2v)
π» System requirements:
Minimum system requirements for 540p i2v and 720p t2v:
RTX 3000-s, 8GB+ VRAM, 45GB+ RAM, 8-core processor, SSD, latest ComfyUI
π Low VRAM optional optimization:
For systems with low VRAM use --reserve-vram ComfyUI parameter in run_nvidia_gpu.bat:Β
--reserve-vram 4Β (or other number in GB).
π Detailed tips and links to models in the workflow
β¨ Workflow features:
Extremely user-friendly interface
Maximum performance and optimization from 8GB of VRAM: GGUF or 8-step distilled model with fp4 or fp8 text encoder
All-in-one: i2v, t2v, and interpolation
Convenient one-click mode switching
Prompt enhancer and sampler previews
Generation time setting in seconds
Lora support (up to 3)
Detailed tips and links to all necessary models
Manual random seed for complete control over generations
π€ππΌ Thanks to Lightricks Team
Original repo β GitHub
Description
Introducing friendly workflow with the new high quality LTX-2.3 model
Single-stage version of the workflow
Use any Dev/Distilled/GGUF model, easily switch between models
Generate with or without audio (t2v and i2v)
Generate with or without the prompt enhancer
Enable/disable generation preview in the sampler
Optimized and reduced third-party nodes
Updated all relevant tooltips and model links