LTX-2 is the first DiT-based audio-video foundation model that contains all core capabilities of modern video generation in one model: synchronized audio and video, high fidelity, multiple performance modes, production-ready outputs, API access, and open access.
# Clone the repository
git clone https://github.com/Lightricks/LTX-2.git
cd LTX-2
# Set up the environment
uv sync --frozen
source .venv/bin/activate
Download the following models from the LTX-2 HuggingFace repository:
LTX-2 Model Checkpoint (choose and download one of the following)
Spatial Upscaler - Required for current two-stage pipeline implementations in this repository
Temporal Upscaler - Supported by the model and will be required for future pipeline implementations
Distilled LoRA - Required for current two-stage pipeline implementations in this repository (except DistilledPipeline and ICLoraPipeline)
Gemma Text Encoder (download all assets from the repository)
LoRAs
LTX-2-19b-IC-LoRA-Canny-Control - DownloadLTX-2-19b-IC-LoRA-Depth-Control - DownloadLTX-2-19b-IC-LoRA-Detailer - DownloadLTX-2-19b-IC-LoRA-Pose-Control - DownloadLTX-2-19b-LoRA-Camera-Control-Dolly-In - DownloadLTX-2-19b-LoRA-Camera-Control-Dolly-Left - DownloadLTX-2-19b-LoRA-Camera-Control-Dolly-Out - DownloadLTX-2-19b-LoRA-Camera-Control-Dolly-Right - DownloadLTX-2-19b-LoRA-Camera-Control-Jib-Down - DownloadLTX-2-19b-LoRA-Camera-Control-Jib-Up - DownloadLTX-2-19b-LoRA-Camera-Control-Static - Download--enable-fp8 (CLI) or fp8transformer=True (Python)uv sync --extra xformers) or Flash Attention 3 for Hopper GPUsTI2VidOneStagePipeline for faster generation when high resolution isn't requiredWhen writing prompts, focus on detailed, chronological descriptions of actions and scenes. Include specific movements, appearances, camera angles, and environmental details - all in a single flowing paragraph. Start directly with the action, and keep descriptions literal and precise. Think like a cinematographer describing a shot list. Keep within 200 words. For best results, build your prompts using this structure:
For additional guidance on writing a prompt please refer to https://ltx.video/blog/how-to-prompt-for-ltx-2
LTX-2 pipelines support automatic prompt enhancement via an enhance_prompt parameter.
To use our model with ComfyUI, please follow the instructions at https://github.com/Lightricks/ComfyUI-LTXVideo/.
This repository is organized as a monorepo with three main packages:
Each package has its own README and documentation. See the Documentation section below.
Each package includes comprehensive documentation: