ComfyUI-LTXVideo
RepositoryFreeLTX-Video Support for ComfyUI
Capabilities14 decomposed
text-to-video generation with ltx-2 diffusion model
Medium confidenceGenerates video sequences from natural language prompts using the LTX-2 diffusion transformer model integrated into ComfyUI core. The system tokenizes text through a Gemma-based CLIP encoder, processes it through the DiT (Diffusion Transformer) architecture, and applies iterative denoising in latent space to produce video frames. Supports both base sampling and advanced guidance mechanisms (STG/APG) to control quality and semantic adherence during generation.
Integrates LTX-2 as a native ComfyUI core component (comfy/ldm/lightricks) with specialized samplers (LTXVBaseSampler, LTXVExtendSampler) that expose advanced diffusion control not available in standard Stable Diffusion implementations. Uses DiT architecture instead of U-Net, enabling more efficient temporal modeling across video frames.
Tighter integration with ComfyUI core than third-party video models, enabling native node-based workflow composition and direct access to model internals for advanced control; faster inference than Runway or Pika due to optimized DiT architecture.
image-to-video synthesis with temporal extension
Medium confidenceConverts a static image into a video sequence by encoding the image as the first frame and using the LTX-2 model to generate subsequent frames that maintain visual consistency and semantic coherence. The system loads the image through the VAE encoder, optionally applies IC-LoRA (in-context LoRA) for structural control, and uses specialized samplers (LTXVInContextSampler) to condition generation on the initial frame while allowing natural motion and scene evolution.
Implements in-context LoRA (IC-LoRA) conditioning system that allows structural control over generated motion without full model retraining. Uses LTXVInContextSampler to inject image conditioning at specific timesteps during diffusion, maintaining frame-level coherence while enabling motion variation.
Offers more granular control over motion generation than Runway's image-to-video through IC-LoRA conditioning; maintains better visual consistency than Pika by leveraging LTX-2's native image conditioning architecture.
two-stage upscaling workflow with quality preservation
Medium confidenceImplements a two-stage video upscaling pipeline that first generates low-resolution video with LTX-2, then applies specialized upscaling models to enhance resolution while preserving temporal coherence and semantic content. The system chains LTX-2 generation with external upscaling models (e.g., RealESRGAN, BSRGAN) through ComfyUI's node system, managing intermediate representations and quality metrics throughout the pipeline.
Implements two-stage pipeline that leverages LTX-2's fast low-resolution generation followed by specialized upscaling, enabling quality-speed tradeoffs not available in single-stage approaches. Integrates with ComfyUI's node system to enable flexible upscaling model selection and chaining.
More efficient than generating high-resolution directly; enables faster iteration and experimentation by decoupling generation from upscaling, unlike end-to-end high-resolution generation approaches.
camera control and motion specification through ic-lora
Medium confidenceEnables precise control over camera movement and object motion in generated videos through in-context LoRA (IC-LoRA) conditioning. The system allows users to specify camera trajectories (pan, zoom, rotate) and object motion paths, which are encoded as conditioning signals and injected into the diffusion process. IC-LoRA weights are loaded through LTXVQ8LoraModelLoader and applied during sampling to guide motion generation without full model retraining.
Implements IC-LoRA conditioning system that enables camera and motion control without full model retraining. Integrates with LTXVQ8LoraModelLoader to support quantized IC-LoRA weights, enabling efficient motion-controlled generation on memory-constrained systems.
More precise camera control than text-only prompts; enables reproducible camera movements across multiple generations, unlike prompt-based approaches which produce variable results.
custom node registration and workflow composition
Medium confidenceProvides a plugin architecture that registers custom nodes with ComfyUI through a dual-registration system (static mappings in __init__.py and runtime-generated nodes from nodes_registry.py). The system enables users to compose complex video generation workflows by connecting nodes in ComfyUI's visual editor, with automatic type checking and data flow validation. NODE_CLASS_MAPPINGS and NODE_DISPLAY_NAME_MAPPINGS enable ComfyUI Manager compatibility and user-friendly node discovery.
Implements dual-registration system (static NODE_CLASS_MAPPINGS + runtime nodes_registry.py) enabling both ComfyUI Manager compatibility and dynamic node generation. NODE_DISPLAY_NAME_MAPPINGS with 'LTXV' prefix provides consistent user-facing naming across all custom nodes.
More flexible than monolithic video generation tools; enables composition of arbitrary node combinations and integration with other ComfyUI extensions, unlike closed-system video generators.
gemma text encoder integration with caching
Medium confidenceIntegrates Lightricks' Gemma-based CLIP text encoder for semantic understanding of prompts, with intelligent caching to avoid redundant encoding of identical prompts. The system implements LTXVGemmaCLIPModelLoader and LTXVGemmaCLIPModelLoaderMGPU that load the encoder, cache embeddings for repeated prompts, and manage encoder lifecycle across multiple generation calls. Supports both single-GPU and multi-GPU loading strategies.
Integrates Lightricks' proprietary Gemma-based CLIP encoder with intelligent prompt embedding caching, reducing redundant encoding overhead. LTXVGemmaCLIPModelLoaderMGPU enables distributed encoder loading across GPUs for batch processing scenarios.
Better semantic understanding than generic CLIP encoders; caching mechanism reduces latency for repeated prompts compared to stateless encoding approaches.
video frame extension and temporal blending
Medium confidenceExtends existing video sequences by generating additional frames that seamlessly blend with original footage. The system uses LTXVExtendSampler to process latent representations of video clips, applies temporal blending operations (LTXVBlendLatents) to smooth transitions between original and generated frames, and supports looping generation (LTXVLoopingSampler) for continuous video synthesis. Latent normalization (LTXVNormalizeLatents) ensures consistent quality across extended sequences.
Implements specialized latent-space blending operations (LTXVBlendLatents, LTXVNormalizeLatents) that work directly on compressed video representations rather than pixel space, reducing computational cost and enabling smooth transitions. LTXVLoopingSampler provides iterative generation with automatic normalization to prevent artifact accumulation.
More efficient than pixel-space blending approaches; latent-space operations enable real-time preview and faster iteration compared to frame-by-frame interpolation methods.
structural guidance with stg and apg control systems
Medium confidenceApplies spatial and temporal guidance during video generation to improve quality and semantic adherence without retraining the model. The system implements two guidance mechanisms: STG (Spatial-Temporal Guidance) for general quality improvement and APG (Adaptive Prompt Guidance) for semantic control. Nodes (STGGuiderNode, STGGuiderAdvancedNode, MultimodalGuiderNode) inject guidance signals into the diffusion process at configurable timesteps, modulating the denoising direction toward desired outputs while maintaining diversity.
Implements dual-guidance architecture with STG for general quality improvement and APG for semantic control, allowing independent tuning of quality vs. semantic adherence. Guidance signals are injected at specific diffusion timesteps through GuiderParametersNode, enabling fine-grained control over generation trajectory without model modification.
More flexible than simple classifier-free guidance used in Stable Diffusion; provides both spatial-temporal and adaptive prompt guidance in a single framework, enabling better quality-diversity tradeoffs than single-guidance approaches.
q8 quantization for low-vram model loading
Medium confidenceReduces model memory footprint through 8-bit quantization, enabling LTX-2 inference on GPUs with limited VRAM (16GB or less). The system implements LTXVQ8LoraModelLoader and LowVRAMCheckpointLoader nodes that load model weights in quantized format, apply dynamic dequantization during inference, and optionally load LoRA adapters in quantized form. This approach trades minimal quality loss for significant memory savings (typically 40-50% reduction).
Implements Q8 quantization specifically for LTX-2 DiT architecture with dynamic dequantization during inference, maintaining quality while reducing memory footprint. LTXVQ8LoraModelLoader extends quantization to LoRA adapters, enabling full workflow quantization without separate adapter loading.
More aggressive memory optimization than standard fp16 loading while maintaining better quality than int4 quantization; specifically tuned for LTX-2's DiT architecture rather than generic quantization approaches.
multi-gpu model distribution and memory management
Medium confidenceDistributes model components across multiple GPUs to enable larger batch sizes and longer video generation on multi-GPU systems. The system implements LTXVGemmaCLIPModelLoaderMGPU and memory optimization nodes that partition the text encoder, diffusion model, and VAE across available devices, managing inter-device communication and synchronization. Automatic memory profiling (LowVRAMCheckpointLoader) detects available VRAM and adjusts model placement accordingly.
Implements GPU-aware model partitioning through LTXVGemmaCLIPModelLoaderMGPU that automatically detects available GPUs and distributes text encoder, DiT, and VAE components based on VRAM availability. Integrates with ComfyUI's device management system for seamless multi-GPU workflows.
More granular control than simple data parallelism; enables model parallelism for components that don't fit on single GPU, unlike standard ComfyUI which requires manual device specification.
latent space manipulation and normalization
Medium confidenceProvides low-level operations on compressed video representations (latent tensors) to enable advanced workflows without decoding to pixel space. The system implements nodes (LTXVSelectLatents, LTXVBlendLatents, LTXVNormalizeLatents, LTXVConcatenateLatents) that manipulate latent dimensions, blend multiple latent sequences, normalize distributions, and concatenate temporal sequences. These operations work directly in compressed space, enabling efficient composition of video generation results.
Implements comprehensive latent-space manipulation toolkit (LTXVSelectLatents, LTXVBlendLatents, LTXVNormalizeLatents, LTXVConcatenateLatents) that operates on LTX-2's specific latent format, enabling efficient video composition without pixel-space decoding. LTXVNormalizeLatents specifically addresses artifact accumulation in iterative generation.
More efficient than pixel-space video editing; enables real-time latent composition and enables workflows impossible in pixel space due to memory constraints.
vae encoding and decoding with video support
Medium confidenceConverts between pixel-space video frames and compressed latent representations using a variational autoencoder optimized for temporal coherence. The system provides VAE encoder/decoder nodes that process video sequences frame-by-frame or in temporal chunks, maintaining consistency across frames while achieving 8-16x compression. Supports both standard VAE decoding and tiled decoding for memory-constrained scenarios.
Implements VAE encoding/decoding specifically optimized for video temporal coherence, with support for both frame-by-frame and chunk-based processing. Tiled decoding option enables memory-efficient processing on systems with limited VRAM without sacrificing quality.
Better temporal consistency than generic image VAE applied frame-by-frame; tiled decoding approach more efficient than full-resolution decoding for memory-constrained systems.
tiled sampling for high-resolution video generation
Medium confidenceGenerates high-resolution videos by dividing the spatial domain into overlapping tiles, sampling each tile independently, and blending results at tile boundaries. The system implements LTXVTiledSampler that manages tile generation, overlap regions, and boundary blending to produce seamless high-resolution output without requiring proportional VRAM increases. Tile size and overlap are configurable to balance quality and memory usage.
Implements spatial tiling specifically for LTX-2's DiT architecture with configurable overlap and boundary blending. LTXVTiledSampler manages tile generation order and blending weights to minimize boundary artifacts while maintaining temporal coherence across tiles.
More efficient than post-hoc upscaling; generates high-resolution content directly from diffusion model rather than interpolating low-resolution output, enabling better detail and semantic consistency.
prompt enhancement and dynamic conditioning
Medium confidenceAugments user prompts with automatically generated enhancements and applies dynamic conditioning during generation. The system provides utility nodes that expand prompts with style descriptors, quality keywords, and temporal directives, then injects these enhanced prompts into the diffusion process at configurable timesteps. Supports both static prompt enhancement and dynamic prompt scheduling that varies conditioning over generation timesteps.
Implements prompt enhancement pipeline that augments base prompts with quality keywords and style descriptors, then applies dynamic prompt scheduling during diffusion. Supports timestep-based prompt variation enabling temporal control (e.g., 'slow motion' in early steps, 'fast motion' in later steps).
More sophisticated than simple prompt concatenation; enables temporal prompt variation and automatic quality enhancement without requiring manual prompt engineering expertise.
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with ComfyUI-LTXVideo, ranked by overlap. Discovered automatically through the match graph.
LTX-Video-ICLoRA-detailer-13b-0.9.8
text-to-video model by undefined. 37,381 downloads.
Hotshot-XL
✨ Hotshot-XL: State-of-the-art AI text-to-GIF model trained to work alongside Stable Diffusion XL
Luma Dream Machine
An AI model that makes high quality, realistic videos fast from text and images.
CogVideoX-5b
text-to-video model by undefined. 35,487 downloads.
LTX-Video
Official repository for LTX-Video
CogVideo
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
Best For
- ✓Content creators building automated video generation workflows
- ✓AI researchers experimenting with diffusion-based video synthesis
- ✓Teams prototyping video generation features in ComfyUI-based applications
- ✓Motion graphics designers creating animated assets from static artwork
- ✓Video editors needing to extend or interpolate existing footage
- ✓Developers building image-to-video features in creative applications
- ✓Teams needing high-resolution output with fast generation
- ✓Content creators prioritizing quality over generation speed
Known Limitations
- ⚠Requires significant VRAM (24GB+ recommended for full model, 16GB minimum with quantization)
- ⚠Generation speed depends on number of denoising steps and video length (typically 30-120 seconds per generation)
- ⚠Text encoder (Gemma) must be loaded separately and cached in memory
- ⚠Output resolution and frame count fixed by model architecture (typically 768x512 or similar)
- ⚠Generated motion may not perfectly match real-world physics or expected camera movements without IC-LoRA conditioning
- ⚠Temporal consistency degrades over longer sequences (typically best for 5-15 second videos)
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
Repository Details
Last commit: Apr 13, 2026
About
LTX-Video Support for ComfyUI
Categories
Alternatives to ComfyUI-LTXVideo
Are you the builder of ComfyUI-LTXVideo?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →