asd / docs /ltx_2.md
FusionCow's picture
Add files using upload-large-folder tool
157a4c0 verified

LTX-2 / LTX-2.3

Supports LoRA training for both LTX-2 (19B) and LTX-2.3 (22B) models with the following training modes: text-to-video, joint audio-video, audio-only, IC-LoRA / video-to-video, and audio-reference IC-LoRA.

Supported Model Versions

Version Parameters Key Differences
LTX-2 (19B) 19B Single aggregate_embed, caption projection inside transformer
LTX-2.3 (22B) 22B Dual video_aggregate_embed/audio_aggregate_embed, caption projection moved to feature extractor (caption_proj_before_connector), cross-attention AdaLN (prompt_adaln), separate audio connector dimensions, BigVGAN v2 vocoder with bandwidth extension

Version choice for training is controlled by --ltx_version (default: 2.0) in ltx2_train_network.py. The trainer auto-detects the checkpoint version from metadata and warns on mismatch.

Caching scripts (ltx2_cache_latents.py, ltx2_cache_text_encoder_outputs.py) do not use --ltx_version; they work with both LTX-2 and LTX-2.3 checkpoints directly via --ltx2_checkpoint.


Table of Contents


Installation

Unless otherwise noted, command examples in this LTX-2 guide were tested on Windows 11. They should also work on Linux, but you may need small shell/path adjustments.

For a Windows-focused community setup example for this fork (tested environment and install helpers), see Discussion #19: Windows OS installation/usage helpers.

CUDA Version

The PyTorch install command must use a CUDA version compatible with your GPU. Adjust the --index-url accordingly:

# Default (most GPUs, including RTX 30xx/40xx):
pip install torch==2.8.0 torchvision==0.23.0 torchaudio==2.8.0 --index-url https://download.pytorch.org/whl/cu126

# RTX 5090 / 50xx series (Blackwell):
pip install torch==2.8.0 torchvision==0.23.0 torchaudio==2.8.0 --index-url https://download.pytorch.org/whl/cu128

Always match the CUDA version to your GPU architecture — check PyTorch's compatibility matrix for the latest supported versions.

Downloading Required Models

The trainer does not download models automatically. You must manually download the following files before caching or training.

LTX-2 Checkpoint — use as --ltx2_checkpoint:

Gemma Text Encoder — pick one:

Other Gemma 3 12B variants may work but not all have been tested.


Supported Dataset Types

Mode Dataset Type Notes
video Images Treated as 1-frame samples (F=1)
video Videos Standard video training
av Videos with audio Audio extracted from video or external audio files
audio Audio only Dataset must be audio-only; training uses audio-driven latent geometry

1. Caching Latents

This step pre-processes media files into VAE latents to speed up training.

Script: ltx2_cache_latents.py

Latent Caching Command

python ltx2_cache_latents.py ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltx-2.safetensors ^
  --device cuda ^
  --vae_dtype bf16 ^
  --ltx2_mode av ^
  --ltx2_audio_source video

Latent Caching Arguments

  • --ltx2_mode, --ltx_mode: Caching modality selector. Default is video-only (v/video). Use av to cache both *_ltx2.safetensors (video) and *_ltx2_audio.safetensors (audio) latents.
  • --ltx2_audio_source video|audio_files: Use audio from the video or from external files.
  • --ltx2_audio_dir, --ltx2_audio_ext: Optional when using --ltx2_audio_source audio_files (default extension: .wav).
  • --ltx2_checkpoint: Required for --ltx2_mode av or --ltx2_mode audio.
  • --audio_only_target_resolution: Optional square override for audio-only latent geometry. Only takes effect when --audio_only_sequence_resolution 0; otherwise the fixed sequence resolution is used instead.
  • --audio_only_target_fps: Target FPS used to derive audio-only frame counts from audio duration (default: 25).
  • --audio_video_latent_channels: Optional override for audio-only video latent channels (auto-detected from checkpoint by default).
  • --ltx2_audio_dtype: Data type for audio VAE encoding (default: float16).
  • --audio_video_latent_dtype: Optional override for audio-only video latent dtype (defaults to --ltx2_audio_dtype).
  • --vae_dtype: Data type for VAE latents (default comes from the cache script).
  • --save_dataset_manifest: Optional. Saves a cache-only dataset manifest for source-free training.
  • --precache_sample_latents: Cache I2V conditioning image latents for sample prompts, then continue with normal latent caching. Requires --sample_prompts.
  • --sample_latents_cache: Path for the I2V conditioning latents cache file (default: <cache_dir>/ltx2_sample_latents_cache.pt).
  • --reference_frames: Number of reference frames to cache for IC-LoRA / V2V (default: 1).
  • --reference_downscale: Spatial downscale factor for cached reference latents (default: 1).

Latent Cache Output Files

File Pattern Contents
*_ltx2.safetensors Video latents: latents_{F}x{H}x{W}_{dtype}. In audio-only mode, this file also stores ltx2_virtual_num_frames_int32 (used for timestep sampling) and ltx2_virtual_height_int32/ltx2_virtual_width_int32 (only used when --audio_only_sequence_resolution 0).
*_ltx2_audio.safetensors Audio latents: audio_latents_{T}x{mel_bins}x{channels}_{dtype}, audio_lengths_int32

Memory Optimization for Caching

If you encounter Out-Of-Memory (OOM) errors during caching (especially with higher resolutions like 1080p), you have two options:

Option 1: VAE temporal chunking (fewer parameters, for moderate OOM)

python ltx2_cache_latents.py ^
  ...
  --vae_chunk_size 16
  • --vae_chunk_size: Processes video in temporal chunks (e.g., 16 or 32 frames at a time). Default: None (all frames).

Option 2: VAE tiled encoding (larger VRAM savings, for severe OOM or high-resolution videos)

python ltx2_cache_latents.py ^
  ...
  --vae_spatial_tile_size 512 ^
  --vae_spatial_tile_overlap 64
  • --vae_spatial_tile_size: Splits each frame into spatial tiles of this size in pixels (e.g., 512). Must be >= 64 and divisible by 32. Default: None (disabled).
  • --vae_spatial_tile_overlap: Overlap between spatial tiles in pixels. Must be divisible by 32. Default: 64.
  • --vae_temporal_tile_size: Splits the video into temporal tiles of this many frames (e.g., 64). Must be >= 16 and divisible by 8. Default: None (disabled).
  • --vae_temporal_tile_overlap: Overlap between temporal tiles in frames. Must be divisible by 8. Default: 24.

Spatial and temporal tiling can be combined. Tiled encoding trades speed for VRAM savings.

Both options can be combined (e.g., --vae_chunk_size 16 --vae_spatial_tile_size 512).


2. Caching Text Encoder Outputs

This step pre-computes text embeddings using the Gemma text encoder.

Script: ltx2_cache_text_encoder_outputs.py

Text Encoder Caching Command

python ltx2_cache_text_encoder_outputs.py ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltx-2.safetensors ^
  --gemma_root /path/to/gemma ^
  --gemma_load_in_8bit ^
  --device cuda ^
  --mixed_precision bf16 ^
  --ltx2_mode av ^
  --batch_size 1

Text Encoder Caching Arguments

  • --gemma_root: Path to the local Gemma model folder (HuggingFace format). Required unless --gemma_safetensors is used.
  • --gemma_safetensors: Path to a single Gemma .safetensors file (e.g. an FP8 export from ComfyUI). Loads weights, config, and tokenizer from one file — no --gemma_root needed. See Loading Gemma from a Single Safetensors File below.
  • --gemma_load_in_8bit: Loads Gemma in 8-bit quantization. Cannot be combined with --gemma_safetensors.
  • --gemma_load_in_4bit: Loads Gemma in 4-bit quantization. Cannot be combined with --gemma_safetensors.
  • --gemma_bnb_4bit_quant_type nf4|fp4: Quantization type for 4-bit loading (default: nf4).
  • --gemma_bnb_4bit_disable_double_quant: Disable bitsandbytes double quantization for 4-bit loading.
  • --gemma_bnb_4bit_compute_dtype auto|fp16|bf16|fp32: Compute dtype for 4-bit operations (default: auto, uses --mixed_precision dtype).
  • --ltx2_checkpoint: Required. Use --ltx2_text_encoder_checkpoint to override for text encoder connector weights.
  • 8-bit/4-bit loading requires --device cuda.

--ltx2_mode / --ltx_mode must match the mode used during latent caching. Default is video; use av to concatenate video and audio prompt embeddings.

Text Encoder Output Files

File Pattern Contents
*_ltx2_te.safetensors video_prompt_embeds_{dtype}, audio_prompt_embeds_{dtype} (av only), prompt_attention_mask, text_{dtype}, text_mask

Loading Gemma from a Single Safetensors File

--gemma_safetensors loads Gemma from a single .safetensors file instead of a HuggingFace model directory. Weights, tokenizer (spiece_model key), and config (inferred from tensor shapes) are all read from the one file. No --gemma_root needed.

python ltx2_cache_text_encoder_outputs.py ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltx-2.safetensors ^
  --gemma_safetensors /path/to/gemma3-12b-it-fp8.safetensors ^
  --device cuda ^
  --mixed_precision bf16
  • FP8 weights (F8_E4M3 / F8_E5M2) are detected automatically and kept in FP8 on GPU (compute in bf16). By default FP8 weights are offloaded to CPU between encoding calls; set LTX2_GEMMA_SAFETENSORS_WEIGHT_OFFLOAD=0 to keep them on GPU.
  • --gemma_load_in_8bit / --gemma_load_in_4bit cannot be combined with --gemma_safetensors.
  • If the file has no spiece_model key, tokenizer extraction fails — use --gemma_root instead.
  • Works in all scripts that load Gemma: ltx2_cache_text_encoder_outputs.py, ltx2_train_network.py, ltx2_train_slider.py, ltx2_generate_video.py.

3. Training

Launch the training loop using accelerate.

Script: ltx2_train_network.py

Choosing Model Version for Training (2.0 vs 2.3)

Use this rule:

Checkpoint you train on Required training flags
LTX-2 (19B) checkpoint --ltx_version 2.0
LTX-2.3 (22B) checkpoint --ltx_version 2.3

Recommended practice:

  • Always set --ltx_version explicitly in training commands (do not rely on the default).
  • On first run, set --ltx_version_check_mode error to fail fast if the selected version does not match checkpoint metadata.
  • After validation, you can switch to --ltx_version_check_mode warn.
  • For LTX-2.3, current upstream recommendation is to train from the BF16 checkpoint. FP8-checkpoint training recipes are currently community-contributed/experimental.

When changing checkpoints (important):

  • If you change --ltx2_checkpoint (e.g., LTX-2 -> LTX-2.3, or different 2.3 variant), re-run both caches:
    • ltx2_cache_latents.py
    • ltx2_cache_text_encoder_outputs.py
  • Do not reuse old *_ltx2_te.safetensors from a different checkpoint. For LTX-2.3 audio/av training this can cause context/mask shape mismatches (for example FlashAttention varlen mask-length errors).
  • If you use --dataset_manifest, regenerate it from the recache step so training points to the new cache files.
  • If your selected checkpoint is already FP8-quantized (for example *fp8*.safetensors), still keep --fp8_base during training (usually with --mixed_precision bf16), but do not add --fp8_scaled.

Example (LTX-2.3 training):

--ltx2_checkpoint /path/to/ltx-2.3.safetensors ^
--ltx_version 2.3 ^
--ltx_version_check_mode error

Optional: Source-Free Training from Cache

If you cached with --save_dataset_manifest, you can train without source dataset paths:

accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 ltx2_train_network.py ^
  --dataset_manifest dataset_manifest.json ^
  ... (other training args)

Use --dataset_manifest instead of --dataset_config.

Standard LoRA Training

accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 ltx2_train_network.py ^
  --mixed_precision bf16 ^
  --dataset_config dataset.toml ^
  --gemma_load_in_8bit ^
  --gemma_root /path/to/gemma ^
  --separate_audio_buckets ^
  --ltx2_checkpoint /path/to/ltx-2.3.safetensors ^
  --ltx_version 2.3 ^
  --ltx_version_check_mode error ^
  --ltx2_mode av ^
  --fp8_base ^
  --fp8_scaled ^
  --blocks_to_swap 10 ^
  --sdpa ^
  --gradient_checkpointing ^
  --learning_rate 1e-4 ^
  --network_module networks.lora_ltx2 ^
  --network_dim 32 ^
  --network_alpha 32 ^
  --timestep_sampling shifted_logit_normal ^
  --sample_at_first ^
  --sample_every_n_epochs 5 ^
  --sample_prompts sampling_prompts.txt ^
  --sample_with_offloading ^
  --sample_tiled_vae ^
  --sample_vae_tile_size 512 ^
  --sample_vae_tile_overlap 64 ^
  --sample_vae_temporal_tile_size 48 ^
  --sample_vae_temporal_tile_overlap 8 ^
  --sample_merge_audio ^
  --output_dir output ^
  --output_name ltx23_lora

If the selected checkpoint is already FP8 (*fp8*.safetensors), keep --fp8_base but remove --fp8_scaled.

For LTX-2 checkpoints, replace:

  • --ltx2_checkpoint /path/to/ltx-2.3.safetensors -> --ltx2_checkpoint /path/to/ltx-2.safetensors
  • --ltx_version 2.3 -> --ltx_version 2.0

Advanced: LyCORIS/LoKR Training

LyCORIS training for LTX-2 has been reported as unstable by some users. This is currently being investigated. Use with caution and please report issues.

musubi-tuner supports advanced LoRA algorithms (LoKR, LoHA, LoCoN, etc.) via:

  • --network_args for inline key=value settings
  • --lycoris_config <path.toml> for TOML-based settings

See the LyCORIS algorithm list and guidelines for algorithm details and recommended settings.

No bundled example TOML files are shipped; provide your own config path.

# Install LyCORIS first
pip install lycoris-lora

# Example TOML (save anywhere, e.g. my_lycoris.toml)
# [network]
# base_algo = "lokr"
# base_factor = 16
#
# [network.modules."*audio*"]
# algo = "lora"
# dim = 64
# alpha = 32
#
# [network.init]
# lokr_norm = 1e-3

# LoKR example
accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 ltx2_train_network.py ^
  ... (same args as above) ^
  --network_module lycoris.kohya ^
  --lycoris_config my_lycoris.toml ^
  --output_name ltx2_lokr

# LoCoN example (inline args)
accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 ltx2_train_network.py ^
  ... (same args as above) ^
  --network_module lycoris.kohya ^
  --network_args "algo=locon" "conv_dim=8" "conv_alpha=4" ^
  --output_name ltx2_locon

Target format (--network_args)

  • Pass args as space-separated key=value pairs.
  • Each pair is one argument token: --network_args "key1=value1" "key2=value2".
  • Common LyCORIS keys: algo, factor, conv_dim, conv_alpha, dropout.
  • Use --init_lokr_norm only with LoKR (algo=lokr).
  • If both TOML and --network_args are used, --network_args can override nested TOML keys with:
    • modules.<name>.<param>=...
    • init.<param>=...

--lycoris_config requires --network_module lycoris.kohya.

Training Arguments

All training arguments can be placed in a .toml config file instead of on the command line via --config_file config.toml. See the upstream configuration files guide for format details.

Memory Optimization

Quantization Options
Method VRAM (19B, LTX-2) Weight Error (MAE) SNR Cosine Similarity
BF16 (baseline) ~38 GB 0.0011 55.6 dB 0.999999
--fp8_base --fp8_scaled ~19 GB 0.0171 (15x BF16) 32.0 dB 0.999686
--nf4_base ~10 GB 0.0678 (60x BF16) 21.2 dB 0.996188
--nf4_base --loftq_init ~10 GB 0.0654 (60x BF16) 21.5 dB 0.996437

Approximate values measured on random N(0,1) weights with shapes representative of LTX-2 transformer layers. MAE = mean absolute error between original and dequantized weights. LoftQ error is measured after adding the LoRA correction (rank 32, 2 iterations). No benchmark script is included in this repo.

NF4 has ~4x higher weight error than FP8 (cosine 0.996 vs 0.9997). The base model is frozen during LoRA training, so the quantization error is constant rather than accumulating. LoftQ initializes LoRA weights from the quantization residual via SVD.

  • --fp8_base: keep base model weights in FP8 path (~19 GB VRAM).
  • --fp8_scaled: quantize non-FP8 (fp16/bf16/fp32) checkpoints to FP8. Do not use this with already-FP8 checkpoints.
  • --nf4_base: NF4 4-bit quantization (~10 GB VRAM). Mutually exclusive with --fp8_base. See NF4 Quantization below.
  • --quantize_device cpu|cuda|gpu: Device for NF4/FP8 quantization at startup (default: cuda). cpu loads and quantizes weights on CPU, then moves to GPU. cuda loads and quantizes directly on GPU. Overrides LTX2_NF4_CALC_DEVICE / LTX2_FP8_CALC_DEVICE env vars.
Other Memory Options
Argument Description
--blocks_to_swap X Offload X transformer blocks to CPU (max 47 for 48-block model). Higher = more VRAM saved, more CPU↔GPU overhead
--use_pinned_memory_for_block_swap Use pinned memory for faster CPU↔GPU block transfers
--gradient_checkpointing Reduce VRAM by recomputing activations during backward pass
--gradient_checkpointing_cpu_offload Offload activations to CPU during gradient checkpointing
--ffn_chunk_target all, video, or audio — enable FFN chunking for selected modules
--ffn_chunk_size N Chunk size for FFN chunking (0 = disabled)
--split_attn_target none, all, self, cross, text_cross, av_cross, video, audio — split attention target modules
--split_attn_mode batch or query — split by batch dimension or query length
--split_attn_chunk_size N Chunk size for query-based split attention (0 = default 1024)
--sdpa Use PyTorch scaled dot-product attention (recommended default)
--flash_attn Use FlashAttention 2 (requires flash-attn package built for your CUDA + PyTorch)
--flash3 Use FlashAttention 3 (requires flash-attn v3 with Hopper+ GPU)

Aggressive VRAM Optimization (8-16GB GPUs)

For maximum VRAM savings on 8-16GB GPUs, use this combination of flags:

accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 ltx2_train_network.py ^
  --mixed_precision bf16 ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltx-2.safetensors ^
  --ltx2_mode av ^
  --gemma_load_in_8bit ^
  --gemma_root /path/to/gemma ^
  --fp8_base ^
  --fp8_scaled ^
  --blocks_to_swap 47 ^
  --use_pinned_memory_for_block_swap ^
  --gradient_checkpointing ^
  --gradient_checkpointing_cpu_offload ^
  --sdpa ^
  --network_module networks.lora_ltx2 ^
  --network_dim 16 ^
  --network_alpha 16 ^
  --sample_with_offloading ^
  --sample_tiled_vae ^
  --sample_vae_tile_size 512 ^
  --sample_vae_temporal_tile_size 48 ^
  --output_dir output

Tips for low-VRAM training:

  • Use --fp8_base --fp8_scaled for non-FP8 checkpoints; use --fp8_base only for already-FP8 checkpoints
  • Use --blocks_to_swap 47 (keeps only 1 block on GPU)
  • Use smaller LoRA rank (--network_dim 16 instead of 32)
  • Use smaller training resolutions (e.g., 512x320)
  • Reduce --sample_vae_temporal_tile_size to 24 or lower
  • Use --use_pinned_memory_for_block_swap - faster transfers

NF4 Quantization

NF4 (4-bit NormalFloat) quantization uses a 16-value codebook optimized for normally-distributed weights (QLoRA paper). Weights are stored as packed uint8 with per-block absmax scaling. VRAM usage is ~10 GB vs ~19 GB for FP8 and ~38 GB for BF16.

Basic usage:

accelerate launch ... ltx2_train_network.py ^
  --nf4_base ^
  --network_module networks.lora_ltx2 ^
  --network_dim 32 ^
  ...

With LoftQ initialization:

LoftQ pre-computes LoRA A/B matrices from the truncated SVD of the NF4 quantization residual (W - dequant(Q(W))). This runs once at startup and adds no runtime cost.

accelerate launch ... ltx2_train_network.py ^
  --nf4_base ^
  --loftq_init ^
  --loftq_iters 2 ^
  --network_module networks.lora_ltx2 ^
  --network_dim 32 ^
  ...
Argument Default Description
--nf4_base off Enable NF4 4-bit quantization for the base model
--nf4_block_size 32 Elements per quantization block
--loftq_init off LoftQ initialization for LoRA (requires --nf4_base)
--loftq_iters 2 Number of alternating quantize-SVD iterations
--awq_calibration off Experimental: activation-aware channel scaling before quantization
--awq_alpha 0.25 AWQ scaling strength (0 = no effect, 1 = full)
--awq_num_batches 8 Number of synthetic calibration batches for AWQ
--quantize_device cuda Device for quantization math (cpu, cuda, gpu)

Pre-quantized models (recommended):

By default, NF4 quantization runs from scratch on every startup. ltx2_quantize_model.py quantizes once and saves the result (~42% of the original file size). The training/inference code auto-detects pre-quantized checkpoints via safetensors metadata and skips re-quantization.

python src/musubi_tuner/ltx2_quantize_model.py ^
  --input_model path/to/ltx-2.3-22b-dev.safetensors ^
  --output_model path/to/ltx-2.3-22b-dev-nf4.safetensors ^
  --loftq_init --network_dim 32

Output files (kept in the same directory):

  • *-nf4.safetensors — quantized model (transformer in NF4, VAE and other components unchanged)
  • *-nf4.loftq_r32.safetensors — pre-computed LoftQ init for rank 32 (only with --loftq_init)

Then use it exactly like the original checkpoint — just point --ltx2_checkpoint at the NF4 file. --nf4_base is still required (enables the runtime dequantization patch):

accelerate launch ... ltx2_train_network.py ^
  --ltx2_checkpoint path/to/ltx-2.3-22b-dev-nf4.safetensors ^
  --nf4_base --loftq_init --network_dim 32 ^
  ...
Argument Default Description
--input_model required Path to original .safetensors checkpoint
--output_model required Path for quantized output
--nf4_block_size 32 Elements per quantization block
--calc_device cuda if available Device for quantization computation
--loftq_init off Pre-compute LoftQ initialization (requires --network_dim)
--loftq_iters 2 Number of LoftQ alternating iterations
--network_dim 0 LoRA rank for LoftQ (must match training --network_dim)
  • LoftQ is rank-specific: changing --network_dim requires re-running the quantize script with the new rank. The quantized model itself does not need to be regenerated.
  • --awq_calibration is incompatible with pre-quantized models (requires full-precision weights).
  • The quantized output is bit-for-bit identical to dynamic quantization on the same device.

Notes:

  • --nf4_base and --fp8_base are mutually exclusive.
  • --loftq_init requires --nf4_base.
  • --awq_calibration is experimental. Adds a per-layer division during forward passes. In synthetic tests, reduces activation-weighted error by ~3-5%; effect on real training quality has not been validated.
  • Compatible with --blocks_to_swap, --gradient_checkpointing, and other training options. NF4 reduces block swap transfer size (4-bit vs 16-bit per weight).
  • Quantization targets transformer block weights only. Embedding layers, norms, and projection layers remain in full precision.

Model Version

  • --ltx_version 2.0|2.3: Select target model version (default: 2.0). Controls default behavior for version-dependent settings (e.g., --shifted_logit_mode defaults to legacy for 2.0, stretched for 2.3). For LTX-2.3 checkpoints, set --ltx_version 2.3 explicitly.
  • --ltx_version_check_mode off|warn|error: How to handle mismatch between --ltx_version and checkpoint metadata (default: warn). The trainer reads checkpoint config keys (cross_attention_adaln, caption_proj_before_connector, bwe vocoder) to detect the actual version.

Audio-Video Support

  • --ltx2_mode, --ltx_mode: Training modality selector. Default is v (video). Values: video, av, audio (aliases: v, va, a).
  • --ltx2_audio_only_model: Force loading a physically audio-only transformer variant (video modules omitted). Requires --ltx2_mode audio.
  • --separate_audio_buckets: Keeps audio and non-audio items in separate batches (reduces VRAM for image/video-only batches).
  • --audio_bucket_strategy pad|truncate: Audio duration bucketing strategy. pad (default) rounds to nearest bucket boundary and pads shorter clips with loss masking. truncate floors to bucket boundary and truncates all clips to bucket length (no padding or masking needed).
  • --audio_bucket_interval: Audio bucket step size in seconds (default: 2.0). Controls how finely audio clips are grouped by duration.
  • --min_audio_batches_per_accum: Minimum number of audio-bearing microbatches per gradient accumulation window.
  • --audio_batch_probability: Probability of selecting an audio-bearing batch when both audio and non-audio batches are available.
    • --min_audio_batches_per_accum and --audio_batch_probability are mutually exclusive.
  • --caption_dropout_rate: Probability of dropping text conditioning for a sample during training (default: 0.0, disabled). When triggered, the sample's text embeddings are zeroed out and the attention mask is cleared, training the model to generate without text guidance. This enables classifier-free guidance (CFG) at inference — without it, the model has no unconditional baseline to contrast against.

Loss Function Type

--loss_type selects the element-wise loss function used for both video and audio branches. Default is mse.

--loss_type PyTorch function Per-element formula
mse (default) F.mse_loss (pred - tgt)²
mae / l1 F.l1_loss |pred - tgt|
huber / smooth_l1 F.smooth_l1_loss 0.5·(pred-tgt)²/δ when |pred-tgt| < δ, else |pred-tgt| - 0.5·δ
  • --huber_delta (float, default: 1.0): Transition point for Huber loss. Only used when --loss_type is huber or smooth_l1. Smaller values make the loss behave more like L1; larger values more like MSE.

All other training mechanics (weighting scheme, masking, audio balancing) apply on top of the chosen loss unchanged.

# L1 loss
--loss_type mae

# Huber with tighter quadratic region
--loss_type huber --huber_delta 0.1

Loss Weighting

  • --video_loss_weight: Weight for video loss (default: 1.0).
  • --audio_loss_weight: Weight for audio loss in AV mode (default: 1.0).
  • Dataset config video_loss_weight / audio_loss_weight override the corresponding CLI weight for that dataset only.
  • --audio_loss_balance_mode: Audio loss balancing strategy. Values: none (default), inv_freq, ema_mag.
  • --audio_loss_balance_min, --audio_loss_balance_max: Clamp range for effective audio weight (defaults: 0.05, 4.0).

inv_freq mode — inverse-frequency reweighting for mixed audio/non-audio training. Boosts audio loss proportionally to how rare audio batches are.

  • --audio_loss_balance_beta: EMA update rate for observed audio-batch frequency (default: 0.01).
  • --audio_loss_balance_eps: Denominator floor for inverse-frequency scaling (default: 0.05).
  • --audio_loss_balance_ema_init: Initial audio-frequency EMA value (default: 1.0).

Example:

--audio_loss_weight 1.0 ^
--audio_loss_balance_mode inv_freq ^
--audio_loss_balance_beta 0.01 ^
--audio_loss_balance_eps 0.05 ^
--audio_loss_balance_min 0.05 ^
--audio_loss_balance_max 3.0

Recommended start values:

  • --audio_loss_balance_beta 0.01 (stable EMA, slower reaction; try 0.02-0.05 for faster reaction)
  • --audio_loss_balance_eps 0.05 (safe floor; increase to 0.1 if weights spike too much)
  • --audio_loss_balance_min 0.05 --audio_loss_balance_max 3.0 (conservative clamp range)
  • --audio_loss_balance_ema_init 1.0 (no warm-start boost; use 0.5 only if you want stronger early audio emphasis)

ema_mag mode — dynamic balancing by matching audio-loss EMA magnitude to a target fraction of video-loss EMA. Bidirectional: can dampen audio (weight < 1.0) when audio loss exceeds target_ratio * video_loss, or boost it when audio loss is below that target.

  • --audio_loss_balance_target_ratio: Target audio/video loss magnitude ratio (default: 0.33 — audio loss targets ~33% of video loss).
  • --audio_loss_balance_ema_decay: EMA decay for loss magnitude tracking (default: 0.99).

Example:

--audio_loss_weight 1.0 ^
--audio_loss_balance_mode ema_mag ^
--audio_loss_balance_target_ratio 0.33 ^
--audio_loss_balance_ema_decay 0.99 ^
--audio_loss_balance_min 0.05 ^
--audio_loss_balance_max 4.0

Use ema_mag when audio and video losses have different natural magnitudes and you want automatic scaling instead of manual --audio_loss_weight tuning.

Additional Audio Training Flags

  • --independent_audio_timestep: Sample a separate timestep for audio (AV/audio modes only).
  • --audio_silence_regularizer: When AV batches are missing audio latents, use synthetic silence latents instead of skipping the audio branch.
  • --audio_silence_regularizer_weight: Loss multiplier for synthetic-silence fallback batches.
  • --audio_supervision_mode off|warn|error: AV audio-supervision monitor mode.
  • --audio_supervision_warmup_steps: Expected AV batches before supervision checks.
  • --audio_supervision_check_interval: Run supervision checks every N expected AV batches.
  • --audio_supervision_min_ratio: Minimum supervised/expected ratio required by the monitor.

Per-Module Learning Rates

Set different learning rates for audio vs. video LoRA modules. Useful when audio modules need a lower LR to stabilize AV training.

  • --audio_lr <float>: Learning rate for all audio LoRA modules (names containing audio_). Defaults to --learning_rate.
  • --lr_args <pattern=lr> ...: Regex-based per-module LR overrides. Patterns are matched against LoRA module names via re.search.

Priority (highest to lowest): --lr_args pattern match > --audio_lr catch-all > --learning_rate default.

Example:

--learning_rate 1e-4 ^
--audio_lr 1e-5 ^
--lr_args audio_attn=1e-6 video_to_audio=5e-6

Result:

  • audio_attn modules → 1e-6 (matched by --lr_args)
  • video_to_audio modules → 5e-6 (matched by --lr_args)
  • Other audio_* modules (e.g. audio_ff) → 1e-5 (--audio_lr)
  • Video modules → 1e-4 (--learning_rate)

Works with LoRA+ (loraplus_lr_ratio): the up/down split applies within each LR group. Both flags default to None and are fully backward-compatible. See LoRA+ in the upstream advanced configuration guide for setup details.

Preservation & Regularization

Optional techniques that constrain how the LoRA modifies the base model. All are disabled by default with zero overhead.

Blank Prompt Preservation — Prevents the LoRA from altering the model's blank-prompt output (used as the CFG baseline during inference):

--blank_preservation --blank_preservation_args multiplier=1.0

Differential Output Preservation (DOP) — Prevents the LoRA from altering class-prompt output, scoping the LoRA effect to the trigger word only:

--dop --dop_args class=woman multiplier=1.0

The class parameter should be a general description without your trigger word (e.g., woman, cat, landscape).

Prior Divergence — Encourages the LoRA to produce outputs that differ from the base model on training prompts, preventing weak/timid LoRAs:

--prior_divergence --prior_divergence_args multiplier=0.1

Audio DOP — Preserves the base model's audio predictions on non-audio training steps. Requires --ltx2_mode av. On each non-audio batch, constructs silence audio latents, runs the transformer with LoRA OFF and ON, and minimizes MSE on the audio branch only. Zero cost on audio batches. Mutually exclusive with --audio_silence_regularizer.

--audio_dop --audio_dop_args multiplier=0.5
Technique Extra forwards/step Extra backwards/step Recommended multiplier
--blank_preservation +2 +1 0.5 - 1.0
--dop +2 +1 0.5 - 1.0
--prior_divergence +1 0 0.05 - 0.1
--audio_dop +2 (non-audio steps only) +1 (non-audio steps only) 0.3 - 1.0

Each preservation technique adds transformer forward passes per step. Audio DOP costs apply only on non-audio steps.

CREPA (Cross-frame Representation Alignment) — Encourages temporal consistency across video frames by aligning DiT hidden states across frames via a small projector MLP. Only the projector is trained; all other modules stay frozen. CREPA uses hooks to capture intermediate features from the existing forward pass (no extra forward passes). Two modes are available: dino (based on arXiv 2506.09229, aligns to pre-cached DINOv2 features from neighboring frames) and backbone (inspired by SimpleTuner LayerSync, aligns to a deeper block of the same transformer).

Enable with --crepa. All parameters are passed via --crepa_args as key=value pairs:

accelerate launch ... ltx2_train_network.py ^
  --crepa ^
  --crepa_args mode=backbone student_block_idx=16 teacher_block_idx=32 lambda_crepa=0.1 tau=1.0 num_neighbors=2 schedule=constant warmup_steps=0 normalize=true

CLI Flags

Flag Type Description
--crepa store_true Enable CREPA regularization
--crepa_args key=value list Configuration parameters (see table below)

CREPA Parameters (--crepa_args)

Parameter Default Description
mode backbone Teacher signal source: backbone (deeper DiT block) or dino (pre-cached DINOv2 features)
dino_model dinov2_vitb14 DINOv2 variant for dino mode. Must match the model used during caching. Options: dinov2_vits14, dinov2_vitb14, dinov2_vitl14, dinov2_vitg14
student_block_idx 16 Transformer block whose hidden states are aligned (0-47 for LTX-2 48-block model)
teacher_block_idx 32 Deeper block providing the teacher signal (backbone mode only, must be > student_block_idx)
lambda_crepa 0.1 Loss weight for CREPA term. Recommended range: 0.05–0.5
tau 1.0 Temporal neighbor decay factor. Controls how much nearby frames contribute vs distant ones
num_neighbors 2 Number of neighboring frames on each side (K). Frame f aligns with frames f-K..f+K
schedule constant Lambda schedule over training: constant, linear (decay to 0), or cosine (cosine decay to 0)
warmup_steps 0 Steps before CREPA loss reaches full strength (linear ramp from 0)
max_steps 0 Total training steps for schedule computation. Auto-filled from --max_train_steps if not set
normalize true L2-normalize features before computing cosine similarity

Checkpoint & Resume

  • The projector weights (~33M params for backbone mode) are saved as crepa_projector.safetensors in the output directory alongside LoRA checkpoints.
  • When resuming training with --crepa, projector weights are automatically loaded from <output_dir>/crepa_projector.safetensors if the file exists.
  • The projector is not needed at inference — it's only used during training.

Monitoring

CREPA adds a loss/crepa metric to TensorBoard/WandB logs. A healthy CREPA loss should:

  • Start negative (cosine similarity is being maximized)
  • Gradually decrease (more negative = stronger cross-frame alignment)
  • Stabilize after warmup

Compatibility

  • Works with block swap (--blocks_to_swap) — hooks fire when each block executes regardless of CPU offloading.
  • Works with all preservation techniques (blank preservation, DOP, prior divergence).
  • Works with gradient checkpointing.
  • Projector params are included in gradient clipping alongside LoRA params.

Caching DINOv2 Features (Dino Mode)

Dino mode requires pre-cached DINOv2 features. Run this after latent caching (cache paths are derived from latent cache files). DINOv2 is not loaded during training — zero VRAM overhead.

python ltx2_cache_dino_features.py ^
  --dataset_config dataset.toml ^
  --dino_model dinov2_vitb14 ^
  --dino_batch_size 16 ^
  --device cuda ^
  --skip_existing
  • --dino_model: DINOv2 variant — dinov2_vits14 (384d), dinov2_vitb14 (768d, default), dinov2_vitl14 (1024d), dinov2_vitg14 (1536d).
  • --dino_batch_size: Frames per forward pass. Reduce if OOM (default: 16).
  • --skip_existing: Skip items that already have cached features.

Output: *_ltx2_dino.safetensors files alongside your latent caches, containing per-frame patch tokens [T, N_patches, D]. For dinov2_vitb14 at 518px input: N_patches=1369, D=768, so each frame adds ~2MB (float16). Disk usage scales linearly with frame count.

Precaching preservation prompts: Blank preservation and DOP require Gemma to encode their prompts at training startup. To avoid loading Gemma during training, precache the embeddings during the text encoder caching step:

python ltx2_cache_text_encoder_outputs.py --dataset_config ... --ltx2_checkpoint ... --gemma_root ... ^
  --precache_preservation_prompts --blank_preservation --dop --dop_class_prompt "woman"

Then add the --use_precached_preservation flag during training:

python ltx2_train_network.py ... ^
  --blank_preservation --dop --dop_args class=woman ^
  --use_precached_preservation

The cache file is saved to <cache_directory>/ltx2_preservation_cache.pt by default (same directory as your dataset cache). Use --preservation_prompts_cache <path> to override the location in either command. Prior divergence does not need precaching (it uses the training batch's own embeddings).

Self-Flow (Self-Supervised Flow Matching)

Self-Flow prevents the fine-tuned model from drifting away from the pretrained model's internal representations. It aligns student features (shallower block) against teacher features (deeper block) using cosine similarity, with dual-timestep noising to create a meaningful student-teacher gap. The default teacher_mode=base uses the frozen pretrained model as teacher by zeroing LoRA multipliers for the teacher forward pass — no extra VRAM overhead compared to EMA. An EMA-based teacher (teacher_mode=ema) is also available for LoRA-aware distillation. The optional temporal extension adds frame-neighbor and motion-delta losses that explicitly preserve temporal coherence. Based on arXiv 2603.06507.

Enable with --self_flow. All parameters are passed via --self_flow_args as key=value pairs:

# Recommended default: base-model teacher, token-level alignment only
accelerate launch ... ltx2_train_network.py ^
  --self_flow ^
  --self_flow_args teacher_mode=base student_block_ratio=0.3 teacher_block_ratio=0.7 lambda_self_flow=0.1 mask_ratio=0.1 dual_timestep=true

# With temporal consistency (hybrid = frame alignment + motion delta)
accelerate launch ... ltx2_train_network.py ^
  --self_flow ^
  --self_flow_args lambda_self_flow=0.1 temporal_mode=hybrid lambda_temporal=0.1 lambda_delta=0.05 num_neighbors=2 temporal_granularity=frame
CLI Flags
Flag Type Description
--self_flow store_true Enable Self-Flow regularization
--self_flow_args key=value list Configuration parameters (see table below)
Self-Flow Parameters (--self_flow_args)
Parameter Default Description
teacher_mode base Teacher source: base (frozen pretrained; zero LoRA multipliers during teacher pass, no extra VRAM), ema (EMA over all LoRA params), partial_ema (EMA over teacher block's LoRA params only)
student_block_idx 16 Student feature block index (0-based; overridden by student_block_ratio when set)
teacher_block_idx 32 Teacher feature block index (must be > student_block_idx; overridden by teacher_block_ratio when set)
student_block_ratio None Ratio-based student layer selection. Resolves to floor(ratio * depth). Takes priority over student_block_idx.
teacher_block_ratio None Ratio-based teacher layer selection. Resolves to ceil(ratio * depth). Takes priority over teacher_block_idx.
student_block_stochastic_range 0 Randomly vary the student capture block ±N blocks each step. 0 = fixed block. Adds regularization diversity; note a single projector is shared across all depth variants.
lambda_self_flow 0.1 Loss weight for the Self-Flow token-level representation term
mask_ratio 0.10 Token mask ratio for dual-timestep mixing. Valid range: [0.0, 0.5]
frame_level_mask false When true, mask whole latent frames instead of individual tokens. More semantically coherent masking for video.
mask_focus_loss false When true, compute the representation loss only on masked (higher-noise) tokens. Default: loss over all tokens.
max_loss 0.0 Cap Self-Flow loss magnitude by rescaling if total loss exceeds this value. 0 = disabled. Useful to prevent Self-Flow from dominating the main task loss early in training.
temporal_mode off Temporal extension mode: off, frame, delta, or hybrid
lambda_temporal 0.0 Loss weight for frame-level temporal neighbor alignment
lambda_delta 0.0 Loss weight for frame-delta alignment (motion consistency)
temporal_tau 1.0 Neighbor decay factor for frame / hybrid temporal alignment
num_neighbors 2 Number of temporal neighbors on each side used by frame / hybrid mode
temporal_granularity frame Temporal loss granularity: frame (mean-pooled per frame) or patch (preserve spatial tokens)
patch_spatial_radius 0 In temporal_granularity=patch, local spatial neighborhood radius for teacher patch matching (0 = strict same-patch only)
patch_match_mode hard Patch-neighborhood matching mode: hard (best patch in window) or soft (softmax-weighted neighborhood match)
patch_match_temperature 0.1 Soft neighborhood matching temperature when patch_match_mode=soft
delta_num_steps 1 Number of temporal delta steps included in the delta loss (1 = adjacent frames only)
motion_weighting none Temporal weighting mode: none or teacher_delta
motion_weight_strength 0.0 Strength of teacher-delta motion weighting for temporal terms
temporal_schedule constant Schedule applied to all Self-Flow lambdas (lambda_self_flow, lambda_temporal, lambda_delta): constant, linear decay, or cosine decay
temporal_warmup_steps 0 Steps to linearly ramp all lambdas up from zero to full weight
temporal_max_steps 0 Steps at which linear / cosine decay reaches zero. 0 = no decay
teacher_momentum 0.999 EMA momentum for teacher updates (ema / partial_ema modes only). Valid range: [0.0, 1.0)
teacher_update_interval 1 Update EMA teacher every N optimizer steps
projector_hidden_multiplier 1 Projector hidden width multiplier vs model inner dim
projector_lr None Optional projector-specific learning rate. Defaults to --learning_rate when unset
loss_type negative_cosine negative_cosine or one_minus_cosine
dual_timestep true Enable dual-timestep noising
tokenwise_timestep true Use per-token timesteps (otherwise per-sample averaged timestep)
offload_teacher_features false Offload cached teacher features to CPU to reduce VRAM
offload_teacher_params false Offload EMA teacher parameters to CPU (saves VRAM, slower teacher forward pass; ema / partial_ema only)
Notes
  • Supported mode: --ltx2_mode video only.
  • Cost: one extra teacher forward pass per train step. teacher_mode=base requires no extra VRAM since it reuses the existing model with LoRA multipliers zeroed.
  • Teacher modes: base gives the largest student-teacher gap (pretrained vs LoRA-finetuned); ema / partial_ema give a moving target that shrinks as training converges.
  • Temporal extension: when temporal_mode != off, Self-Flow reshapes hidden states into latent frames and adds frame-neighbor and/or frame-delta consistency losses on top of the base token alignment loss.
  • Granularity: temporal_granularity=frame uses mean-pooled per-frame features (cheaper, coarser). temporal_granularity=patch keeps spatial tokens for stronger temporal matching.
  • Local patch matching: when temporal_granularity=patch and patch_spatial_radius > 0, each student patch can align to the best teacher patch inside a local spatial window, which is more tolerant to small motion and camera drift than strict same-patch matching.
  • Soft matching: patch_match_mode=soft replaces hard local best-match selection with softmax-weighted neighborhood matching for smoother gradients.
  • Multi-step motion: delta_num_steps > 1 extends the delta loss beyond adjacent frames using exponentially decayed step weights.
  • Motion-aware weighting: motion_weighting=teacher_delta upweights temporally active teacher regions, focusing the temporal loss on moving content.
  • Scheduling: temporal_schedule, temporal_warmup_steps, and temporal_max_steps apply to all three lambdas — lambda_self_flow, lambda_temporal, and lambda_delta — uniformly.
  • State files (Accelerate *-state folder): self_flow_projector.safetensors, self_flow_teacher_ema.safetensors (EMA state only saved when teacher_mode=ema or partial_ema).
  • Resume: both state files are loaded automatically when present. Loading EMA state with teacher_mode=base emits a warning and is ignored.
  • Logged metrics: loss/self_flow, self_flow/cosine, self_flow/frame_cosine, self_flow/delta_cosine, self_flow/lambda_self_flow, self_flow/lambda_temporal, self_flow/lambda_delta, self_flow/masked_token_ratio, self_flow/tau_mean, self_flow/tau_min_mean.

Timestep Sampling

  • --timestep_sampling shifted_logit_normal: Default LTX-2 method. Uses a shifted logit-normal distribution where the shift is computed based on sequence length (frames × height × width).
  • --timestep_sampling uniform: Uniform sampling from [0, 1].
  • --logit_std: Standard deviation for the logit-normal distribution (default: 1.0). Only used with shifted_logit_normal.
  • --min_timestep / --max_timestep: Optional timestep range constraints.
  • --shifted_logit_mode legacy|stretched: Sigma sampler variant (default: auto by --ltx_version; 2.0→legacy, 2.3→stretched).
    • legacy: sigmoid(N(shift, std)). Original behavior.
    • stretched: Normalizes samples between the 0.5th and 99.9th percentiles of the distribution, reflects values below eps for numerical stability, and replaces a fraction of samples with uniform draws to prevent distribution collapse at high token counts.
  • --shifted_logit_eps: Reflection floor and uniform lower bound for stretched mode (default: 1e-3).
  • --shifted_logit_uniform_prob: Fraction of samples replaced with uniform [eps, 1] draws (default: 0.1).
  • --shifted_logit_shift: Override the auto-calculated shift value. Lower values (e.g., 0.0) produce a symmetric distribution centered on medium noise (σ≈0.5) for learning fine details. Higher values (e.g., 2.0) heavily right-skew the distribution toward high noise (σ≈0.9+) for learning global structure. If unset, it is auto-calculated from sequence length.

The shifted_logit_normal shift is linearly interpolated from 0.95 (at 1024 tokens) to 2.05 (at 4096 tokens) based on sequence length. In --ltx2_mode audio, shifted_logit_normal still needs a sequence length to compute the shift, but there is no real video spatial dimension. Using full video resolution would inflate the sequence length and skew the shift upward. Instead, --audio_only_sequence_resolution (default 64) provides a small fixed spatial footprint (4 tokens/frame), keeping the shift dominated by the temporal dimension (audio duration/FPS) which actually matters.

LoRA Targets

Use --lora_target_preset to control which layers LoRA targets:

Preset Layers Use Case
t2v (default) Attention only (to_q, to_k, to_v, to_out.0) Text-to-video, matches official LTX-2 trainer
v2v Attention + FFN Video-to-video / IC-LoRA style
audio Audio attention/FFN + audio-side cross-modal attention Audio-only training (auto-selected when --ltx2_mode audio)
audio_ref_only_ic Audio attn/FFN + bidirectional AV cross-modal Audio-reference IC-LoRA
full All linear layers All layers targeted, larger file size

All presets apply to all relevant attention types: self-attention, cross-attention, and cross-modal attention (in AV mode). Connector layers are always excluded.

To use custom layer patterns instead of a preset, use --network_args:

--network_args "include_patterns=['.*\.to_k$','.*\.to_q$','.*\.to_v$','.*\.to_out\.0$','.*\.ff\.net\.0\.proj$','.*\.ff\.net\.2$']"

Custom include_patterns override any preset. When include_patterns is set (either explicitly or via a preset), only modules matching at least one pattern are targeted (strict whitelist behavior). Use --lora_target_preset full to target all linear layers.

IC-LoRA / Video-to-Video Training

IC-LoRA (In-Context LoRA) trains the model to generate video conditioned on a reference image or video.

Reference frames are encoded as clean latent tokens (timestep=0) and concatenated with noisy target tokens during training. The model attends across both sequences, using the reference as conditioning context. At inference, the same concatenation scheme is applied. Position embeddings are computed separately for reference and target, allowing different spatial resolutions via --reference_downscale.

Step 1: Prepare Dataset

Create a video dataset with a matching reference directory. Each reference file must share the same filename stem as its corresponding training video:

videos/                    references/
  scene_001.mp4              scene_001.png     # reference for scene_001
  scene_002.mp4              scene_002.jpg
  scene_003.mp4              scene_003.mp4     # video references also work

References can be images (single frame) or videos (multiple frames).

Step 2: Dataset Config

Add reference_directory and reference_cache_directory to your TOML config:

[general]
resolution = [768, 512]
caption_extension = ".txt"
batch_size = 1
enable_bucket = true
cache_directory = "cache"
reference_cache_directory = "cache_ref"

[[datasets]]
video_directory = "videos"
reference_directory = "references"
target_frames = [1, 17, 33]
Step 3: Cache Latents

Cache both video latents and reference latents in one step:

python ltx2_cache_latents.py ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltx-2.safetensors ^
  --device cuda ^
  --vae_dtype bf16

Reference latents are automatically cached to reference_cache_directory when reference_directory is configured.

Downscaled references (--reference_downscale):

python ltx2_cache_latents.py ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltx-2.safetensors ^
  --reference_downscale 2 ^
  --device cuda

--reference_downscale 2 encodes references at half spatial resolution (e.g., 384px for 768px target). Position embeddings on the reference spatial axes are scaled by the factor so they map into the target coordinate space.

Step 4: Cache Text Encoder Outputs

Same as standard training — no special flags needed:

python ltx2_cache_text_encoder_outputs.py ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltx-2.safetensors ^
  --gemma_root /path/to/gemma ^
  --gemma_load_in_8bit ^
  --device cuda
Step 5: Train

Use --lora_target_preset v2v (targets attention + FFN layers):

accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 ltx2_train_network.py ^
  --mixed_precision bf16 ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltx-2.safetensors ^
  --fp8_base --fp8_scaled ^
  --blocks_to_swap 10 ^
  --sdpa ^
  --gradient_checkpointing ^
  --network_module networks.lora_ltx2 ^
  --network_dim 32 --network_alpha 32 ^
  --lora_target_preset v2v ^
  --ltx2_first_frame_conditioning_p 0.2 ^
  --timestep_sampling shifted_logit_normal ^
  --learning_rate 1e-4 ^
  --sample_at_first ^
  --sample_every_n_epochs 5 ^
  --sample_prompts sampling_prompts.txt ^
  --sample_include_reference ^
  --output_dir output ^
  --output_name ltx2_ic_lora

If you used --reference_downscale during caching, also pass it during training:

  --reference_downscale 2
Step 6: Sample Prompts

Use --v <path> in your sampling prompts file to specify the V2V reference for each prompt. Both images and videos are supported:

--v references/scene_001.png A woman walking through a forest --n blurry, low quality
--v references/scene_002.mp4 A cat sitting on a windowsill --n distorted

The --sample_include_reference flag shows the reference side-by-side with the generated output in validation videos.

IC-LoRA Arguments
Argument Default Description
--reference_downscale 1 Spatial downscale factor for references (1=same res, 2=half)
--reference_frames 1 Number of reference frames for V2V (images are repeated to fill this count)
--ltx2_first_frame_conditioning_p 0.1 Probability of also conditioning on the first target frame during training. No effect for single-frame (image) samples
--sample_include_reference off Show reference side-by-side with generated output in sample videos
--lora_target_preset v2v Targets attention + FFN layers (recommended for IC-LoRA)
Dataset Config Options
Option Type Description
reference_directory string Path to reference images/videos (matched by filename stem)
reference_cache_directory string Output directory for cached reference latents
Notes
  • First-frame conditioning (--ltx2_first_frame_conditioning_p): Randomly conditions on the first target frame in addition to the reference. Only applied during training; inference always denoises the full target. Has no effect for single-frame (image-only) samples — the code skips conditioning when num_frames == 1 since there are no subsequent frames to generate.
  • Multi-frame references: Supported but increase VRAM usage proportionally to the number of reference tokens.
  • Multi-subject references: The VAE compresses 8 frames into 1 temporal latent via SpaceToDepthDownsample, which pairs consecutive frames and averages their features. Subjects sharing the same 8-frame group are blended and lose individual identity. To keep N subjects separated, structure your reference video as: frame 1 = Subject A, frames 2–9 = Subject B (repeated 8×), frames 10–17 = Subject C (repeated 8×), etc. Total frames: 1 + 8×(N−1). Set --reference_frames to match. Frame 1 gets its own latent due to causal padding in the encoder; each subsequent 8-frame block produces one additional latent.
  • Video-only: IC-LoRA requires --ltx2_mode video. Audio-video mode is not supported for v2v training.
  • Downscale factor metadata: Saved in LoRA safetensors as ss_reference_downscale_factor when factor != 1.
  • Two-stage inference: Not supported with V2V; a warning is emitted and the reference is ignored.

Audio-Reference IC-LoRA

This approach is based on ID-LoRA, adapted for audio-video conditioning in the LTX-2 transformer.

Trains a LoRA using in-context audio-reference conditioning. Reference audio latents (clean, timestep=0) are concatenated with noisy target audio latents during training. Loss is computed only on the target portion. In AV mode the LoRA targets audio self/cross-attention, audio FFN, and bidirectional audio-video cross-modal attention layers; in audio-only mode the audio preset is auto-selected, which omits cross-modal layers that connect to the (dummy) video branch.

Supported modes:

  • --ltx2_mode av — full audio-video model; trains both video and audio IC-LoRA layers.
  • --ltx2_mode audio — audio-only mode; trains only audio layers (video is a dummy zero tensor). --lora_target_preset audio is auto-selected (cross-modal layers that affect the dummy video branch are omitted).
How it works
  1. Reference audio is encoded to latents and concatenated with noisy target audio along the temporal axis.
  2. Reference tokens receive timestep=0 (no noise); target tokens receive the sampled sigma.
  3. Loss is masked to exclude the reference portion — the model only learns to predict the target.
  4. Three optional attention overrides control how the reference interacts with the rest of the model:
    • Negative positions: shifts reference tokens into negative RoPE time, creating clean positional separation from target tokens.
    • A2V cross-attention mask: blocks video from attending to reference audio (video syncs with target audio only).
    • Text attention mask: blocks reference audio from attending to text (reference provides identity, not content).
Step 1: Prepare Data

Organize training videos with matching reference audio files (same filename stem):

videos/                    reference_audio/
  speaker_001.mp4            speaker_001.wav    # reference clip for speaker_001
  speaker_002.mp4            speaker_002.flac

Reference audio files are matched to training videos by filename stem.

Step 2: Dataset Config
[general]
resolution = [768, 512]
caption_extension = ".txt"
batch_size = 1
enable_bucket = true
cache_directory = "cache"
reference_audio_cache_directory = "cache_ref_audio"
separate_audio_buckets = true

[[datasets]]
video_directory = "videos"
reference_audio_directory = "reference_audio"
target_frames = [1, 17, 33]
Step 3: Cache Latents
python ltx2_cache_latents.py ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltxav-2.safetensors ^
  --ltx2_mode av ^
  --device cuda ^
  --vae_dtype bf16

For audio-only mode, replace --ltx2_mode av with --ltx2_mode audio (no video latents are cached, only audio and reference audio latents).

Reference audio latents are automatically cached to reference_audio_cache_directory.

Step 4: Cache Text Encoder Outputs

No special flags — use whichever mode you are training:

python ltx2_cache_text_encoder_outputs.py ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltxav-2.safetensors ^
  --ltx2_mode av ^
  --gemma_root /path/to/gemma ^
  --gemma_load_in_8bit ^
  --device cuda

For audio-only mode, replace --ltx2_mode av with --ltx2_mode audio.

Step 5: Train
accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 ltx2_train_network.py ^
  --mixed_precision bf16 ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltxav-2.safetensors ^
  --ltx2_mode av ^
  --fp8_base --fp8_scaled ^
  --blocks_to_swap 10 ^
  --sdpa ^
  --gradient_checkpointing ^
  --network_module networks.lora_ltx2 ^
  --network_dim 128 --network_alpha 128 ^
  --lora_target_preset audio_ref_only_ic ^
  --audio_ref_use_negative_positions ^
  --audio_ref_mask_cross_attention_to_reference ^
  --audio_ref_mask_reference_from_text_attention ^
  --ltx2_first_frame_conditioning_p 0.9 ^
  --timestep_sampling shifted_logit_normal ^
  --learning_rate 2e-4 ^
  --sample_at_first ^
  --sample_every_n_epochs 5 ^
  --sample_prompts sampling_prompts.txt ^
  --output_dir output ^
  --output_name ltx2_audio_ref_ic_lora

Audio-only mode — replace --ltx2_mode av with --ltx2_mode audio and omit --lora_target_preset (auto-selected as audio):

accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 ltx2_train_network.py ^
  --mixed_precision bf16 ^
  --dataset_config dataset.toml ^
  --ltx2_checkpoint /path/to/ltxav-2.safetensors ^
  --ltx2_mode audio ^
  --ic_lora_strategy audio_ref_only_ic ^
  --fp8_base --fp8_scaled ^
  --blocks_to_swap 10 ^
  --sdpa ^
  --gradient_checkpointing ^
  --network_module networks.lora_ltx2 ^
  --network_dim 128 --network_alpha 128 ^
  --audio_ref_use_negative_positions ^
  --audio_ref_mask_reference_from_text_attention ^
  --timestep_sampling shifted_logit_normal ^
  --learning_rate 2e-4 ^
  --sample_at_first ^
  --sample_every_n_epochs 5 ^
  --sample_prompts sampling_prompts.txt ^
  --output_dir output ^
  --output_name ltx2_audio_ref_ic_lora_audioonly
Step 6: Sample Prompts

Use --ra <path> in your sampling prompts file to specify the reference audio:

--ra reference_audio/speaker_001.wav A person speaking about nature --n blurry, low quality
--ra reference_audio/speaker_002.flac A woman laughing in a park

Reference audio latents are precached automatically when using --precache_sample_latents during latent caching.

Audio-Reference IC-LoRA Arguments
Argument Default Description
--ic_lora_strategy audio_ref_only_ic auto Activates audio-reference IC-LoRA mode (auto-inferred from --lora_target_preset audio_ref_only_ic)
--lora_target_preset audio_ref_only_ic Targets audio attn/FFN + bidirectional AV cross-modal layers
--audio_ref_use_negative_positions off Place reference audio in negative RoPE time for positional separation
--audio_ref_mask_cross_attention_to_reference off Block video from attending to reference audio tokens (AV mode only; no effect in audio-only mode)
--audio_ref_mask_reference_from_text_attention off Block reference audio from attending to text tokens
--audio_ref_identity_guidance_scale 0.0 Override CFG scale for target-audio branch during audio_ref_only_ic sampling (0 = use standard guidance scale)
Dataset Config Options
Option Type Description
reference_audio_directory string Path to reference audio files (matched by filename stem)
reference_audio_cache_directory string Output directory for cached reference audio latents
Notes
  • Checkpoint: requires an LTXAV checkpoint for both --ltx2_mode av and --ltx2_mode audio.
  • Bucket separation: separate_audio_buckets = true keeps audio/non-audio items in separate batches (avoids shape mismatches in collation).
  • Attention overrides: the three --audio_ref_* flags default to off. The reference ID-LoRA configuration enables all three.
  • LoRA rank: the reference ID-LoRA configuration uses rank 128 with alpha 128.
  • First-frame conditioning: the reference ID-LoRA configuration uses first_frame_conditioning_p = 0.9.
  • --ic_lora_strategy auto (default) infers the strategy from --lora_target_preset via infer_ic_lora_strategy_from_preset().

Sampling with Tiled VAE

The prompt file format (--sample_prompts) — including guidance scale, negative prompt, and per-prompt inference parameters — is documented in the upstream Sampling During Training guide. LTX-2 extends this with --v <path> (IC-LoRA reference) and --ra <path> (audio-reference IC-LoRA) prompt prefixes.

Argument Default Description
--height 512 Sample output height (pixels)
--width 768 Sample output width (pixels)
--sample_num_frames 45 Number of frames for sample video generation
--sample_with_offloading off Offload DiT to CPU between sampling prompts to save VRAM
--sample_tiled_vae off Enable tiled VAE decoding during sampling to reduce VRAM
--sample_vae_tile_size 512 Spatial tile size (pixels)
--sample_vae_tile_overlap 64 Spatial tile overlap (pixels)
--sample_vae_temporal_tile_size 0 Temporal tile size in frames (0 = disabled)
--sample_vae_temporal_tile_overlap 8 Temporal tile overlap (frames)
--sample_merge_audio off Merge generated audio into the output .mp4
--sample_audio_only off Generate audio-only preview outputs
--sample_disable_audio off Disable audio preview generation during sampling
--sample_audio_subprocess on Decode audio in a subprocess to avoid OOM crashes. Use --no-sample_audio_subprocess to decode in-process
--sample_disable_flash_attn off Force SDPA instead of FlashAttention during sampling
--sample_i2v_token_timestep_mask on Use I2V token timestep masking (conditioned tokens use t=0). Use --no-sample_i2v_token_timestep_mask to disable

Precached Sample Prompts

To avoid loading Gemma during training for sample generation, you can precache the prompt embeddings:

  1. During text encoder caching, add --precache_sample_prompts --sample_prompts sampling_prompts.txt to also cache the sample prompt embeddings.
  2. During training, add --use_precached_sample_prompts (or --precache_sample_prompts) to load embeddings from cache instead of running Gemma.
  • --sample_prompts_cache: Path to the precached embeddings file. Defaults to <cache_directory>/ltx2_sample_prompts_cache.pt.

For IC-LoRA / V2V training, you can also precache the conditioning image latents during latent caching (see Latent Caching Arguments):

  1. During latent caching, add --precache_sample_latents --sample_prompts sampling_prompts.txt.
  2. During training, add --use_precached_sample_latents to load conditioning latents from cache instead of loading the VAE encoder.
  • --sample_latents_cache: Path to the precached latents file. Defaults to <cache_directory>/ltx2_sample_latents_cache.pt.

Two-Stage Sampling (WIP)

This feature is work in progress and disabled by default. Two-stage inference generates at half resolution, then upsamples and refines.

Argument Default Description
--sample_two_stage off Enable two-stage inference during sampling
--spatial_upsampler_path Path to spatial upsampler model. Required when --sample_two_stage is set
--distilled_lora_path Path to distilled LoRA for stage 2 refinement. Optional
--sample_stage2_steps 3 Number of denoising steps for stage 2

Checkpoint Output Format

Saved LoRA checkpoints are converted to ComfyUI format by default. Both the original musubi-tuner format and the ComfyUI format are kept.

Flag Behavior
(default) Saves both *.safetensors (original) and *.comfy.safetensors (ComfyUI).
--no_save_original_lora Deletes the original after conversion, keeping only *.comfy.safetensors.
--no_convert_to_comfy Saves only the original *.safetensors (no conversion).
--save_checkpoint_metadata Saves a .json sidecar file alongside each checkpoint with loss, lr, step, and epoch.

Important: Training can only be resumed from the original (non-comfy) checkpoint format. If you plan to use --resume, do not use --no_save_original_lora.

Checkpoint rotation (--save_last_n_epochs) cleans up old ComfyUI checkpoints alongside originals. HuggingFace upload (--huggingface_repo_id) uploads both formats by default. Use --no_save_original_lora to upload only the ComfyUI checkpoint.

Resuming Training

Requires --save_state to be enabled. State directories contain optimizer, scheduler, and RNG states.

Flag Description
--resume <path> Resume from a specific state directory
--autoresume Automatically resume from the latest state in output_dir. Ignored if --resume is specified. Starts from scratch if no state is found

Merge LTX-2 LoRAs

Use the dedicated LTX-2 LoRA merger to combine multiple LoRA files into a single LoRA checkpoint.

Script: ltx2_merge_lora.py

Example Command (Windows)

python ltx2_merge_lora.py ^
  --lora_weight path/to/lora_A.safetensors path/to/lora_B.safetensors ^
  --lora_multiplier 1.0 1.0 ^
  --save_merged_lora path/to/merged_lora.safetensors

LoRA Merge Arguments

  • --lora_weight: Input LoRA paths to merge in order (required).
  • --lora_multiplier: Per-LoRA multipliers aligned with --lora_weight. Use one value to apply the same multiplier to all inputs.
  • --save_merged_lora: Output merged LoRA path (required).
  • --merge_method concat|orthogonal: Merge method (default: concat). concat keeps all ranks by concatenation. orthogonal uses SVD refactorization to merge exactly 2 LoRAs with orthogonal projection.
  • --orthogonal_k_fraction: Fraction of top singular directions projected out bilaterally before combining (default: 0.5, range [0, 1]). Only used with --merge_method orthogonal.
  • --orthogonal_rank_mode sum|max|min: Target rank mode for orthogonal merge (default: sum).
  • --dtype auto|float32|float16|bfloat16: Output tensor dtype. auto promotes from input dtypes.
  • --emit_alpha: Force writing <module>.alpha keys in output.

Notes

  • This merger is intended for LTX-2 LoRA formats used in this repo, including Comfy-style lora_A/lora_B weights.
  • It handles different ranks and partial module overlap across input LoRAs.
  • Orthogonal merge requires exactly 2 input LoRAs.

Dataset Configuration

The dataset config is a TOML file with [general] defaults and [[datasets]] entries. Common options shared across all musubi-tuner architectures — including frame_extraction modes, JSONL metadata format, control image support, and resolution bucketing — are documented in the upstream Dataset Configuration guide. The options below are LTX-2-specific or supplement upstream defaults.

Video Dataset Options

Option Type Default Description
video_directory string Path to video/image directory
video_jsonl_file string Path to JSONL metadata file
resolution int or [int, int] [960, 544] Target resolution
target_frames [int] [1] List of target frame counts
frame_extraction string "head" Frame extraction mode
max_frames int 129 Maximum number of frames
source_fps float auto-detected Source video FPS. Auto-detected from video container metadata when not set. Use this to override auto-detection.
target_fps float 25.0 Target training FPS. Frames are resampled to this rate. When audio is present and the source video has a different FPS, the audio waveform is automatically time-stretched (pitch-preserving) to match the target video duration.
batch_size int 1 Batch size
num_repeats int 1 Dataset repetitions
enable_bucket bool false Enable resolution bucketing
bucket_no_upscale bool false Prevent upscaling when bucketing (only downscale to fit)
cache_directory string Latent cache output directory
reference_directory string Reference images/videos for IC-LoRA (matched by filename)
reference_cache_directory string Output directory for cached reference latents (IC-LoRA)
reference_audio_directory string Reference audio files for audio-reference IC-LoRA (matched by filename stem)
reference_audio_cache_directory string Output directory for cached reference audio latents
separate_audio_buckets bool false Keep audio/non-audio items in separate batches

Audio Dataset Options

Audio-only datasets use audio_directory instead of video_directory.

Option Type Default Description
audio_directory string Path to audio file directory
audio_jsonl_file string Path to JSONL metadata file
audio_bucket_strategy string "pad" "pad" (round-to-nearest, pad + mask) or "truncate" (floor, clip to bucket length)
audio_bucket_interval float 2.0 Bucket step size in seconds
batch_size int 1 Batch size
num_repeats int 1 Dataset repetitions
cache_directory string Latent cache output directory

Example TOML

[general]
resolution = [768, 512]
caption_extension = ".txt"
batch_size = 1
enable_bucket = true
cache_directory = "cache"

[[datasets]]
video_directory = "videos"
target_frames = [1, 17, 33, 49]
target_fps = 25    # optional, defaults to 25

Frame Rate (FPS) Handling

During latent caching, the source FPS is auto-detected from each video's container metadata and frames are resampled to target_fps (default: 25). The model receives video at the configured temporal rate regardless of the source material.

How It Works

  1. For each video file, the source FPS is read from the container metadata (average_rate or base_rate).
  2. If abs(ceil(source_fps) - target_fps) > 1, frames are resampled (dropped) to match target_fps.
  3. If the difference is within this threshold (e.g., 23.976 → ceil=24 vs target 25, diff=1), no resampling is done — this avoids spurious frame drops from NTSC rounding (23.976, 29.97, 59.94, etc.).
  4. If audio is present (--ltx2_mode av), the audio waveform is automatically time-stretched (pitch-preserving) to match the resampled video duration.

Common Scenarios

Default — no FPS config needed:

[[datasets]]
video_directory = "videos"
target_frames = [1, 17, 33, 49]
# source_fps: auto-detected per video
# target_fps: defaults to 25

A 60fps video is resampled to 25fps. A 30fps video is resampled to 25fps. A 25fps video is passed through as-is. Mixed-FPS datasets work correctly — each video is resampled independently.

Training at a non-standard frame rate (e.g., 60fps):

[[datasets]]
video_directory = "videos_60fps"
target_frames = [1, 17, 33, 49]
target_fps = 60

Set target_fps to the desired training rate. Videos at 60fps (or 59.94fps) pass through without resampling. Videos at other frame rates are resampled to 60fps.

Overriding auto-detection (e.g., variable frame rate videos):

[[datasets]]
video_directory = "videos"
target_frames = [1, 17, 33, 49]
source_fps = 30

If auto-detection gives wrong results (common with variable frame rate / VFR recordings from phones), set source_fps explicitly. This applies to all videos in that dataset entry, so group videos by FPS into separate [[datasets]] blocks if needed.

Image directories:

Image directories have no FPS metadata. No resampling is applied — all images are loaded as individual frames regardless of target_fps.

Log Messages

During latent caching, log messages confirm what's happening for each video:

Auto-detected source FPS: 60.00 for my_video.mp4
Resampling my_video.mp4: 60.00 FPS -> 25.00 FPS

If you see no "Resampling" line for a video, it means source and target FPS were close enough (within 1 FPS after rounding up the source) and all frames were kept as-is. If you see unexpected frame counts in your cached latents, check these log lines first.

Quick Reference

Your situation What to set What happens
Mixed FPS dataset, want 25fps training Nothing (defaults work) Each video auto-detected, resampled to 25fps
All videos are 25fps Nothing Auto-detected as 25fps, no resampling
All videos are 60fps, want 60fps training target_fps = 60 Auto-detected as 60fps, no resampling
All videos are 60fps, want 25fps training Nothing Auto-detected as 60fps, resampled to 25fps
VFR videos with wrong detection source_fps = 30 (your actual FPS) Overrides auto-detection
Image directory Nothing No FPS concept, all images loaded

Validation Datasets

You can configure a separate validation dataset to track validation loss (val_loss) during training. This helps detect overfitting and compare training runs. Validation datasets use exactly the same schema as training datasets — any format that works for [[datasets]] works for [[validation_datasets]].

Configuration

Add a [[validation_datasets]] section to your existing TOML config file:

[general]
resolution = [768, 512]
caption_extension = ".txt"
batch_size = 1
enable_bucket = true

# Training data
[[datasets]]
video_directory = "videos/train"
cache_directory = "cache/train"
target_frames = [1, 17, 33, 49]

# Validation data
[[validation_datasets]]
video_directory = "videos/val"
cache_directory = "cache/val"
target_frames = [1, 17, 33, 49]

The cache_directory for validation must be different from the training cache directory.

Caching

Validation datasets are automatically picked up by the caching scripts — no extra flags needed. Run the same caching commands you use for training:

python ltx2_cache_latents.py --dataset_config dataset.toml --ltx2_checkpoint /path/to/ltx-2.safetensors --ltx2_mode av ...
python ltx2_cache_text_encoder_outputs.py --dataset_config dataset.toml --ltx2_checkpoint /path/to/ltx-2.safetensors --ltx2_mode av ...

Both scripts detect the [[validation_datasets]] section and cache latents/text embeddings for validation data alongside training data.

Training Arguments

Argument Type Default Description
--validate_every_n_steps int None Run validation every N training steps
--validate_every_n_epochs int None Run validation every N epochs

At least one of these must be set for validation to run. If neither is set, validation is skipped even if [[validation_datasets]] is configured.

Example

accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 ltx2_train_network.py ^
  --dataset_config dataset.toml ^
  --validate_every_n_steps 100 ^
  ... (other training args)

How It Works

  1. A separate validation dataloader is created with batch_size=1 and shuffle=False (deterministic order).
  2. At the configured interval, the model switches to eval mode and runs inference on all validation samples with torch.no_grad().
  3. The average MSE loss is computed and logged as val_loss to TensorBoard/WandB.
  4. The model is restored to training mode and training continues.

Tips

  • Keep validation sets small. Aim for 5-20% of your main dataset size. Validation runs on every sample each time, so 10-50 clips is usually enough. Large validation sets slow down training.
  • Use held-out data. Validation data should be different from the training set for meaningful overfitting detection. In extreme cases, using a small subset of the training data is acceptable — it will still help catch divergence, but won't reliably detect overfitting.
  • Monitor the gap. If val_loss starts increasing while training loss keeps decreasing, you're overfitting — consider stopping or reducing the learning rate.
  • Same preprocessing. Validation data goes through the same frame extraction, FPS resampling, and resolution bucketing as training data.

Directory Structure

Raw Dataset Layout (Example)

dataset_root/
  videos/
    000001.mp4
    000001.txt       # caption
    000002.mp4
    000002.txt

Cache Directory Layout (After Caching)

cache_directory/
  000001_1024x0576_ltx2.safetensors       # video latents
  000001_ltx2_te.safetensors              # text encoder outputs
  000001_ltx2_audio.safetensors           # audio latents (av mode only)
  000001_1024x0576_ltx2_dino.safetensors  # DINOv2 features (CREPA dino mode only)

reference_cache_directory/                  # IC-LoRA only
  000001_1024x0576_ltx2.safetensors       # reference latents

Troubleshooting

Error Cause Solution
Missing cache keys during training Caching incomplete Run both ltx2_cache_latents.py and ltx2_cache_text_encoder_outputs.py
FlashAttention varlen mask-length mismatch (for example expects mask length 1920, got 1024) after checkpoint switch Stale text cache from a different checkpoint/version/mode Re-run ltx2_cache_text_encoder_outputs.py with the same --ltx2_checkpoint and --ltx2_mode as training. Remove old *_ltx2_te.safetensors if needed.
Samples become progressively noisier/degraded when training with --flash_attn FlashAttention install/runtime mismatch (CUDA/PyTorch/flash-attn build) Switch to --sdpa to confirm baseline stability. If SDPA is stable, reinstall FlashAttention for your exact CUDA + PyTorch versions and retry.
Training fails after changing --ltx2_checkpoint even though args look correct Reused latent/text caches generated from a different checkpoint Re-run both caches (ltx2_cache_latents.py and ltx2_cache_text_encoder_outputs.py) and regenerate --dataset_manifest before training.
OOM appears after removing --fp8_base while using an FP8 checkpoint Base model no longer uses FP8 loading path, so VRAM increases sharply Keep --fp8_base enabled for FP8 checkpoints (typically with --mixed_precision bf16)
Error when combining --fp8_scaled with an FP8 checkpoint --fp8_scaled is for quantizing non-FP8 checkpoints Remove --fp8_scaled; keep --fp8_base
Missing *_ltx2_audio.safetensors Audio caching skipped Re-run latent caching with --ltx2_mode av
Gemma connector weights missing Incorrect checkpoint Ensure --ltx2_checkpoint (or --ltx2_text_encoder_checkpoint) contains Gemma connector weights
Gemma OOM Model too large Use --gemma_load_in_8bit or --gemma_load_in_4bit with --device cuda, or use --gemma_safetensors with an FP8 file
Audio caching fails torchaudio missing Install torchaudio before running ltx2_cache_latents.py
Sampling OOM VAE decode too large Enable --sample_tiled_vae or reduce --sample_vae_temporal_tile_size
Crash with block swap (esp. RTX 5090) --use_pinned_memory_for_block_swap bug Remove --use_pinned_memory_for_block_swap from training arguments
stack expects each tensor to be equal size during AV training Mixed audio/non-audio videos in the same batch — text embeddings are 2×caption_channels for AV items vs 1×caption_channels for video-only (e.g., 7680 vs 3840 for LTX-2.3), and torch.stack fails Add --separate_audio_buckets to training args. Required when your dataset mixes videos with and without audio at batch_size > 1. At batch_size=1 it has no effect.
Wrong frame count in cached latents Auto-detected FPS incorrect (e.g., VFR video) Set source_fps explicitly in TOML config to override auto-detection
Too few frames from high-FPS video FPS resampling working correctly (e.g., 60fps→25fps = 42% of frames) Expected behavior. Set target_fps = 60 if you want to keep all frames
Audio/video out of sync after caching Source FPS mismatch causing wrong time-stretch Check "Auto-detected source FPS" log line; set source_fps explicitly if wrong
Voice/audio learning slow when mixing images with videos in AV mode Image batches produce zero audio training signal — audio branch is skipped entirely. Dilutes audio learning proportionally to image step fraction Use video-only datasets for AV training when voice quality matters
No audio during sampling in video training mode ltx2_mode is set to v/video Expected behavior. Train in AV mode (--ltx2_mode av or audio) to generate audio during sampling
Cannot resume training from checkpoint Using a *.comfy.safetensors checkpoint with --resume Training can only be resumed from the original (non-comfy) LoRA format. Use the *.safetensors file without the .comfy extension. If you used --no_save_original_lora, you must retrain from scratch.
CUDA errors or crashes on RTX 5090 / 50xx GPUs CUDA 12.6 (cu126) not supported on Windows for Blackwell GPUs Use CUDA 12.8: pip install torch==2.8.0 ... --index-url https://download.pytorch.org/whl/cu128. See CUDA Version
ValueError: Gemma safetensors is missing required language-model tensors with missing_buffers mentioning full_attention_inv_freq or sliding_attention_inv_freq transformers>=5.0 renamed Gemma3 rotary embedding buffers (rotary_emb.inv_freqrotary_emb.full_attention_inv_freq / sliding_attention_inv_freq). The derivable-buffer suffix check expects .inv_freq and does not match the new _inv_freq suffix. The safetensors file is correct — rotary buffers are non-persistent and computed from config at init time. pip install transformers==4.56.1 (pinned in pyproject.toml), or reinstall all deps with pip install -e .
loss_a too low but loss_v still high (audio overfitting) Audio latent space converges faster than video; audio gradients dominate shared weights Lower --audio_loss_weight (e.g., 0.3), or use --audio_loss_balance_mode ema_mag to auto-dampen audio when it exceeds target_ratio × video_loss. Reduce audio learning rate with --audio_lr 1e-6 or fine-grained --lr_args audio_attn=1e-6 audio_ff=1e-6. Disable --audio_dop / --audio_silence_regularizer if active — they add more audio signal.
loss_a absent or not dropping in mixed dataset (audio starvation) Audio batches too rare — non-audio steps outnumber audio steps, audio branch gets insufficient supervision Increase num_repeats on audio datasets (target 30-50% audio steps). Add --audio_loss_balance_mode inv_freq to auto-boost audio weight. Use --audio_dop or --audio_silence_regularizer to provide audio signal on non-audio steps. Check caching summary for failed > 0.

Audio/Voice Training with Mixed Datasets

In mixed datasets, audio batches are a minority. Non-audio steps update shared transformer weights without audio supervision, causing audio drift. Mitigations:

1. Oversample audio items — set num_repeats so audio batches are 30-50% of steps:

[[datasets]]
video_directory = "audio_video_clips"
num_repeats = 5

2. --audio_dop — preserves base model audio predictions on non-audio steps. Runs LoRA OFF/ON forwards on silence audio latents, MSE on audio branch only. +2 fwd / +1 bwd per non-audio step, zero cost on audio batches. Logged as loss/audio_dop. Mutually exclusive with --audio_silence_regularizer.

--audio_dop --audio_dop_args multiplier=0.5

3. --dop — preserves all base model outputs (video + audio) using a class prompt. Broader than audio DOP but applies to every step. +2 fwd / +1 bwd per step.

--dop --dop_args multiplier=0.5

4. --audio_silence_regularizer — converts non-audio batches to audio batches with silence target. Cheaper than DOP (+0 extra forwards) but silence is an approximation.

--audio_silence_regularizer --audio_silence_regularizer_weight 0.5

5. Inverse-frequency loss balancing — auto-boosts audio loss weight proportional to audio batch rarity:

--audio_loss_balance_mode inv_freq --audio_loss_balance_min 0.05 --audio_loss_balance_max 4.0

Alternative: --audio_loss_balance_mode ema_mag matches audio loss magnitude to a target fraction of video loss.

6. Diagnostics

  • If failed > 0 in latent caching summary, audio extraction is broken for those items
  • After mode switch (video→AV), re-run both latent and text encoder caching without --skip_existing
  • loss_a dropping = audio learning; absent/zero = no audio batches forming; degrades over time = forgetting

Technical Notes

  • Float32 AdaLN: The transformer applies Adaptive Layer Norm (AdaLN) shift/scale operations in float32, then casts back to the working dtype. This prevents overflow that can occur when bf16 scale values multiply bf16 hidden states. The fix is always active and requires no flags.
  • Loss dtype: The LTX-2 training path computes the task loss (MSE, L1, Huber) in trainer.dit_dtype (typically bf16 with --mixed_precision bf16). Internal regularization losses (motion preservation, CREPA, Self-Flow) always use MSE and are unaffected by --loss_type.

4. Slider LoRA Training

Slider LoRA training is based on the ai-toolkit implementation by ostris, adapted for LTX-2.

Slider LoRAs learn a controllable direction in model output space (e.g., "detailed" vs "blurry"). At inference, you scale the LoRA multiplier to control the effect strength and direction: +1.0 enhances, -1.0 erases, 0.0 is the base model, and values like +2.0 or -0.5 work too.

Script: ltx2_train_slider.py

Two modes are available:

Mode Input Use Case
text Prompt pairs only (no dataset) Sliders from text prompt pairs, no images needed
reference Pre-cached latent pairs Sliders from paired positive/negative image samples

4a. Text-Only Mode

Learns a slider direction from positive/negative prompt pairs. No images or dataset config needed.

Slider Config (ltx2_slider.toml)

mode = "text"
guidance_strength = 1.0
sample_slider_range = [-2.0, -1.0, 0.0, 1.0, 2.0]

[[targets]]
positive = "extremely detailed, sharp, high resolution, 8k"
negative = "blurry, out of focus, low quality, soft"
target_class = ""      # empty = affect all content
weight = 1.0
  • guidance_strength: Scales the directional offset applied to targets. Higher values = stronger direction signal but may overshoot.
  • target_class: The conditioning prompt used during training passes. Empty string means the slider affects all content regardless of prompt. Set to e.g. "a portrait" to restrict the slider's effect to a specific subject.
  • weight: Per-target loss weight. Controls relative emphasis when training multiple directions simultaneously.
  • sample_slider_range: Multiplier values used for preview samples during training.

Multiple [[targets]] blocks can be defined to train several directions at once (e.g., detail + lighting).

Example Command (Text-Only)

accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 ltx2_train_slider.py ^
  --mixed_precision bf16 ^
  --ltx2_checkpoint /path/to/ltx-2.safetensors ^
  --gemma_root /path/to/gemma ^
  --gemma_load_in_8bit ^
  --fp8_base --fp8_scaled ^
  --gradient_checkpointing ^
  --blocks_to_swap 10 ^
  --network_module networks.lora_ltx2 ^
  --network_dim 16 --network_alpha 16 ^
  --lora_target_preset t2v ^
  --learning_rate 1e-4 ^
  --optimizer_type AdamW8bit ^
  --lr_scheduler constant_with_warmup --lr_warmup_steps 20 ^
  --max_train_steps 500 ^
  --output_dir output --output_name detail_slider ^
  --slider_config ltx2_slider.toml ^
  --latent_frames 1 ^
  --latent_height 512 --latent_width 768

Text-Only Arguments

Argument Default Description
--slider_config (required) Path to slider TOML config file
--latent_frames 1 Number of latent frames (1=image, >1=video)
--latent_height 512 Pixel height for synthetic latents
--latent_width 768 Pixel width for synthetic latents
--guidance_strength (from TOML) Override guidance_strength from config
--sample_slider_range (from TOML) Override as comma-separated values, e.g. "-2,-1,0,1,2"

All standard training arguments (--fp8_base, --blocks_to_swap, --gradient_checkpointing, etc.) work the same as regular training. --dataset_config is not needed for text-only mode.

4b. Reference Mode

Learns a slider direction from paired positive/negative images (or videos). Requires pre-cached latents.

Step 1: Prepare Paired Data

Create two directories with matching filenames — one with positive-attribute images, one with negative:

positive_images/        negative_images/
  img_001.png             img_001.png      # same subject, different attribute
  img_002.png             img_002.png
  img_003.png             img_003.png

Each positive image must have a corresponding negative image with the same filename. The images should depict the same subject but differ in the target attribute (e.g., smiling vs neutral face, detailed vs blurry).

Step 2: Cache Latents and Text

Create a dataset config for each directory and run the caching scripts. Both directories can share the same text captions (since the direction comes from the images, not the text).

# Cache positive latents
python ltx2_cache_latents.py --dataset_config positive_dataset.toml --ltx2_checkpoint /path/to/ltx-2.safetensors

# Cache negative latents
python ltx2_cache_latents.py --dataset_config negative_dataset.toml --ltx2_checkpoint /path/to/ltx-2.safetensors

# Cache text (once, for either directory — text is shared)
python ltx2_cache_text_encoder_outputs.py --dataset_config positive_dataset.toml --ltx2_checkpoint /path/to/ltx-2.safetensors --gemma_root /path/to/gemma

Step 3: Configure and Train

Slider Config (ltx2_slider_reference.toml)
mode = "reference"
pos_cache_dir = "path/to/positive/cache"
neg_cache_dir = "path/to/negative/cache"
text_cache_dir = "path/to/positive/cache"   # can be same as pos (text is shared)
sample_slider_range = [-2.0, -1.0, 0.0, 1.0, 2.0]
  • pos_cache_dir: Directory containing cached positive latents (output of ltx2_cache_latents.py)
  • neg_cache_dir: Directory containing cached negative latents
  • text_cache_dir: Directory containing cached text encoder outputs. Defaults to pos_cache_dir if omitted.

Pairs are matched by filename: for each {name}_{W}x{H}_ltx2.safetensors in the positive directory, a matching file must exist in the negative directory. Unmatched files are skipped with a warning.

Example Command (Reference)
accelerate launch --num_cpu_threads_per_process 1 --mixed_precision bf16 ltx2_train_slider.py ^
  --mixed_precision bf16 ^
  --ltx2_checkpoint /path/to/ltx-2.safetensors ^
  --fp8_base --fp8_scaled ^
  --gradient_checkpointing ^
  --blocks_to_swap 10 ^
  --network_module networks.lora_ltx2 ^
  --network_dim 16 --network_alpha 16 ^
  --lora_target_preset t2v ^
  --learning_rate 1e-4 ^
  --optimizer_type AdamW8bit ^
  --lr_scheduler constant_with_warmup --lr_warmup_steps 20 ^
  --max_train_steps 500 ^
  --output_dir output --output_name smile_slider ^
  --slider_config ltx2_slider_reference.toml

Note: --gemma_root is not needed for reference mode (text embeddings are loaded from cache). --dataset_config, --latent_frames/height/width are also not used.

Slider Tips

  • Start small: --network_dim 8 or 16 with --max_train_steps 200-500 is usually sufficient.
  • Monitor loss: Loss should decrease steadily. If it diverges, reduce --learning_rate.
  • Preview samples: Add --sample_prompts sampling_prompts.txt --sample_every_n_steps 50 to generate previews at each slider strength during training. Requires --gemma_root for text encoding.
  • Guidance strength: For text-only mode, the default is 1.0. Values of 2.0-3.0 increase direction strength but may reduce convergence stability.
  • Multiple targets: Text-only mode supports multiple [[targets]] blocks. Each step randomly selects one target, so all directions get trained evenly.
  • Inference: Use the trained LoRA with any multiplier value. Positive multipliers enhance the positive attribute, negative multipliers enhance the negative attribute. Values beyond [-1, +1] extrapolate the effect.

References

Upstream musubi-tuner Documentation

  • Dataset Configuration — TOML format, frame_extraction modes, JSONL metadata, control images, resolution bucketing
  • Sampling During Training — Prompt file format, per-prompt guidance scale, negative prompts, sampling CLI flags
  • Advanced Configuration--config_file TOML training configuration, --network_args format, LoRA+, TensorBoard/WandB logging, PyTorch Dynamo, timestep bucketing, Schedule-Free optimizer
  • LyCORIS Algorithm List and Guidelines — LoKR, LoHA, LoCoN and other algorithm details (used via pip install lycoris-lora)
  • Tools — Post-hoc EMA LoRA merging, image captioning with Qwen2.5-VL

Research

  • ID-LoRA — In-context identity LoRA; the audio-reference IC-LoRA implementation in this trainer is based on this approach
  • CREPA (arXiv 2506.09229) — Cross-frame Representation Alignment; basis for --crepa dino mode (DINOv2 teacher from neighboring frames)
  • Self-Flow (arXiv 2603.06507) — Self-supervised flow matching regularization; basis for --self_flow

Official LTX Resources

  • LTX-2 — Official Lightricks LTX-2/2.3 repository; contains the well-structured ltx-trainer and ltx-pipelines packages that served as the upstream source and reference for this implementation
  • LTX-Video — Official Lightricks model repository (inference, ComfyUI nodes, model weights)
  • LTX Documentation — Unified docs hub: open-source model, API reference, ComfyUI integration, LoRA usage, and LTX-2 trainer guide

Alternative Trainers

  • ai-toolkit (ostris) — General diffusion fine-tuning toolkit with LTX-2 LoRA support; slider LoRA training is based on its implementation
  • SimpleTuner — Multi-model fine-tuning framework with LTX-2 support; --crepa backbone mode is inspired by its LayerSync regularizer
  • DiffSynth-Studio — ModelScope diffusion synthesis framework with LTX-2/2.3 support

Community Resources

  • awesome-ltx2 — Curated list of LTX-2 resources, tools, models, and guides
  • Banodoco Discord — Active AI video generation community; discussions on LTX-2 training, workflows, and research
  • Windows Installation Guide — Windows-specific setup (Python 3.12, CUDA, Flash Attention 2), dependencies, troubleshooting
  • LTX-2 Training Optimizers — Optimizer comparison for LTX-2 training: AdamW, Prodigy, Muon, CAME, and recommended settings
  • LTX-2 Audio Dataset Builder — Tool to automate audio dataset creation: transforms raw audio into clean, captioned segments optimized for LTX-2 audio-only training

Tutorials & Guides

Cloud Platforms