Capability
Transformer3d Spatiotemporal Attention With Causal Masking
15 artifacts provide this capability.
Want a personalized recommendation?
Find the best match →Top Matches
Official repository for LTX-Video
Unique: Combines 3D spatiotemporal attention with causal masking and grouped query attention, enabling efficient processing of video sequences while enforcing temporal causality and reducing memory overhead through parameter sharing across query groups
vs others: Causal 3D attention with grouped queries reduces memory by ~60% vs. full cross-attention while maintaining temporal coherence, enabling longer video generation than non-causal transformers which require bidirectional context