.. _parallelisms: Parallelisms ============ NeMo uses native PyTorch parallelism primitives for distributed training, enabling efficient multi-GPU and multi-node model training for Speech AI workloads. DDP (all collections) --------------------- Distributed Data Parallelism (DDP) is the default strategy for all NeMo collections (ASR, TTS, Audio, SpeechLM2). It replicates the entire model on every GPU, runs each GPU on a different data shard, and synchronizes parameter gradients via all-reduce after each backward pass. **When to use:** DDP works well when the full model fits in a single GPU's memory. This covers the vast majority of ASR, TTS, and Audio training workloads. DDP is enabled by default in NeMo. You can configure it explicitly in YAML: .. code-block:: yaml trainer: strategy: _target_: lightning.pytorch.strategies.DDPStrategy gradient_as_bucket_view: true find_unused_parameters: true Or in Python: .. code-block:: python from lightning.pytorch.strategies import DDPStrategy trainer = pl.Trainer( strategy=DDPStrategy(gradient_as_bucket_view=True, find_unused_parameters=True), devices=8, accelerator="gpu", ) ModelParallelStrategy (SpeechLM2) --------------------------------- For SpeechLM2 models (e.g. SALM / Canary-Qwen), the backbone LLM can be too large for a single GPU. PyTorch Lightning's ``ModelParallelStrategy`` enables FSDP2, Tensor Parallelism (TP), and Sequence Parallelism (SP) using PyTorch-native DTensor. **When to use:** When training or fine-tuning SpeechLM2 models whose LLM backbone does not fit in a single GPU's memory, or when you want to scale training to many GPUs more efficiently than DDP allows. **Requirements:** Each model must implement a ``configure_model()`` method that defines how its layers are sharded (FSDP2) and parallelized (TP / SP). The SpeechLM2 models (SALM, DuplexEARTTS) already implement this. You cannot simply switch an arbitrary model from DDP to ``ModelParallelStrategy`` without providing this implementation. Concepts ^^^^^^^^ **FSDP2 (Fully Sharded Data Parallelism):** Shards model parameters, gradients, and optimizer states across GPUs in the data-parallel dimension. Dramatically reduces per-GPU memory -- enabling training of models that would not fit with DDP. Controlled via the ``data_parallel_size`` argument. **Tensor Parallelism (TP):** Splits individual weight matrices across GPUs. For example, a large linear layer's weight is partitioned column-wise or row-wise so each GPU holds only a slice. Controlled via the ``tensor_parallel_size`` argument. The model must define a TP sharding plan (which layers are split and how). SpeechLM2 models automatically use the HuggingFace TP plan for the backbone LLM when available. **Sequence Parallelism (SP):** Distributes activation memory along the sequence dimension across the TP group. SP is typically enabled alongside TP and reduces activation memory further. Configuration ^^^^^^^^^^^^^ To enable ``ModelParallelStrategy`` for SpeechLM2, replace the DDP strategy block in the trainer config. The product of ``data_parallel_size`` and ``tensor_parallel_size`` must equal the total number of GPUs (``devices * num_nodes``). In YAML (with Hydra): .. code-block:: yaml trainer: devices: 8 num_nodes: 1 accelerator: gpu precision: bf16-true strategy: _target_: lightning.pytorch.strategies.ModelParallelStrategy data_parallel_size: 4 # FSDP2: shard across 4 GPUs tensor_parallel_size: 2 # TP: split layers across 2 GPUs In Python: .. code-block:: python from lightning.pytorch.strategies import ModelParallelStrategy trainer = pl.Trainer( strategy=ModelParallelStrategy( data_parallel_size=4, tensor_parallel_size=2, ), devices=8, accelerator="gpu", precision="bf16-true", use_distributed_sampler=False, ) .. note:: When using ``ModelParallelStrategy``, set ``use_distributed_sampler=False`` in the trainer. NeMo's data modules handle distributed sampling internally. Example: SALM with FSDP2 only (no TP) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ The simplest ``ModelParallelStrategy`` setup uses FSDP2 alone. This requires no TP plan and works when individual layers fit in GPU memory: .. code-block:: yaml trainer: devices: 8 strategy: _target_: lightning.pytorch.strategies.ModelParallelStrategy data_parallel_size: 8 tensor_parallel_size: 1 Example: SALM with TP + FSDP2 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ For larger LLM backbones, combine TP with FSDP2. Here, 2-way TP splits each layer across 2 GPUs within a node, and 4-way FSDP2 shards the model across 4 such groups: .. code-block:: yaml trainer: devices: 8 strategy: _target_: lightning.pytorch.strategies.ModelParallelStrategy data_parallel_size: 4 tensor_parallel_size: 2 See the SpeechLM2 example configs in ``examples/speechlm2/conf/`` for complete training configurations including data and optimizer settings.