Transformers documentation

DeepSpeed ZeRO

You are viewing main version, which requires installation from source. If you'd like regular pip install, checkout the latest stable version (v5.5.4).
Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

DeepSpeed ZeRO

DeepSpeed ZeRO (Zero Redundancy Optimizer) eliminates memory redundancy across distributed training by sharding optimizer states, gradients, and parameters across GPUs. ZeRO has three stages, each sharding more state than the last. DeepSpeed also supports offloading to CPU or NVMe memory for further savings. Every additional stage and offload level reduces peak memory, at the cost of more inter-GPU communication.

                  params        grads       opt states
                ┌──────────┐ ┌──────────┐ ┌──────────┐
ZeRO-1          │██████████│ │██████████│ │███░░░░░░░│  GPU 0
                │██████████│ │██████████│ │░░░███░░░░│  GPU 1
                │██████████│ │██████████│ │░░░░░░████│  GPU 2
                └──────────┘ └──────────┘ └──────────┘
                ┌──────────┐ ┌──────────┐ ┌──────────┐
ZeRO-2          │██████████│ │███░░░░░░░│ │███░░░░░░░│  GPU 0
                │██████████│ │░░░███░░░░│ │░░░███░░░░│  GPU 1
                │██████████│ │░░░░░░████│ │░░░░░░████│  GPU 2
                └──────────┘ └──────────┘ └──────────┘
                ┌──────────┐ ┌──────────┐ ┌──────────┐
ZeRO-3          │███░░░░░░░│ │███░░░░░░░│ │███░░░░░░░│  GPU 0
                │░░░███░░░░│ │░░░███░░░░│ │░░░███░░░░│  GPU 1
                │░░░░░░████│ │░░░░░░████│ │░░░░░░████│  GPU 2
                └──────────┘ └──────────┘ └──────────┘
  █ resident    ░ held on another GPU

ZeRO-2 shards gradients and optimizer states with lower communication overhead than ZeRO-3. Use ZeRO-3 only when your model doesn’t fit across GPUs with ZeRO-2.

Installation

Install DeepSpeed from PyPI, or install Transformers with the deepspeed extra.

pip install deepspeed
pip install transformers[deepspeed]

If you run into CUDA-related install errors, check the DeepSpeed CUDA docs. Installing from source is the more reliable option because it matches your exact hardware and includes features not yet available in the PyPI release.

Configure

Trainer integrates DeepSpeed through the TrainingArguments.deepspeed argument, which accepts a JSON config file. Alternatively, use an Accelerate config file instead of TrainingArguments.

TrainingArguments
Accelerate config file

Use "auto" in your config for values you want DeepSpeed to fill from TrainingArguments. If you want to explicitly specify a value, make sure you use the same value for both the DeepSpeed argument and TrainingArguments.

See the DeepSpeed Configuration JSON reference for a complete list of DeepSpeed config options.

"train_micro_batch_size_per_gpu": "auto",  // ← per_device_train_batch_size in TrainingArguments
"gradient_accumulation_steps": "auto",     // ← gradient_accumulation_steps in TrainingArguments
"optimizer.params.lr": "auto",             // ← learning_rate in TrainingArguments
"fp16.enabled": "auto",                    // ← fp16 flag in TrainingArguments

Pass the config to the deepspeed argument.

from transformers import TrainingArguments

args = TrainingArguments(
    deepspeed="path/to/deepspeed_config.json",
    ...
)
# DeepSpeed launcher
deepspeed --num_gpus 4 train.py

# torchrun
torchrun --nproc_per_node 4 train.py

# Accelerate
accelerate launch --num_processes 4 train.py

ZeRO stages

Select a ZeRO stage config to use as a starting point.

ZeRO-1
ZeRO-2
ZeRO-3
{
    "bf16": { "enabled": "auto" },
    "zero_optimization": { "stage": 1 },
    "gradient_clipping": "auto",
    "train_micro_batch_size_per_gpu": "auto",
    "train_batch_size": "auto",
    "gradient_accumulation_steps": "auto"
}

The following fields are important for customizing training.

  • zero_optimization sets the ZeRO stage.

    { "zero_optimization": { "stage": 3 } }
  • Set the batch size and gradient accumulation arguments to "auto". If you manually set these to values that disagree with TrainingArguments, training continues silently with the wrong values.

    {
        "train_micro_batch_size_per_gpu": "auto",
        "train_batch_size": "auto",
        "gradient_accumulation_steps": "auto",
        "gradient_clipping": "auto"
    }
  • bf16 sets the training precision. Set it to "auto" so it mirrors the bf16 flag in TrainingArguments.

    { "bf16": { "enabled": "auto" } }
  • stage3_gather_16bit_weights_on_model_save performs an all-gather across all GPUs before saving, reconstructing the full tensors from their shards. This is a ZeRO-3 argument.

    {
        "zero_optimization": {
            "stage": 3,
            "stage3_gather_16bit_weights_on_model_save": true,
        }
    }
  • Set overlap_comm to true to hide all-reduce latency behind the backward pass. allgather_bucket_size and reduce_bucket_size trade communication speed for GPU memory. Lower values use less memory but slow communication.

    {
        "zero_optimization": {
            "stage": 2,
            "overlap_comm": true,
            "allgather_bucket_size": 2e8,
            "reduce_bucket_size": 2e8,
            "contiguous_gradients": true
        }
    }
  • offload_optimizer offloads the optimizer to CPU memory. To save even more memory, also offload model parameters with offload_param (ZeRO-3 only). Set pin_memory to true to speed up CPU-GPU transfers, but this locks RAM that is unavailable to other processes.

    {
        "zero_optimization": {
            "stage": 3,
            "offload_optimizer": { "device": "cpu", "pin_memory": true },
            "offload_param":     { "device": "cpu", "pin_memory": true }
        }
    }
  • optimizer and scheduler default to the optimizer and scheduler configured in TrainingArguments. Set to "auto" so DeepSpeed reads the values from TrainingArguments unless you need a DeepSpeed-native optimizer like LAMB.

    {
        "optimizer": {
            "type": "AdamW",
            "params": { "lr": "auto", "betas": "auto", "eps": "auto", "weight_decay": "auto" }
        },
        "scheduler": {
            "type": "WarmupDecayLR",
            "params": { "total_num_steps": "auto", "warmup_min_lr": "auto", "warmup_max_lr": "auto", "warmup_num_steps": "auto" }
        }
    }

    If you’re offloading the optimizer, set zero_force_ds_cpu_optimizer to false to use DeepSpeed’s CPU Adam optimizer.

    {
        "zero_force_ds_cpu_optimizer": false
    }

Checkpoints

DeepSpeed saves checkpoints in a sharded format that can’t be loaded directly with from_pretrained(). Set load_best_model_at_end() to True to have Trainer track and reload the best checkpoint at the end of training.

from transformers import TrainingArguments, Trainer

args = TrainingArguments(
    deepspeed="ds_config_zero3.json",
    load_best_model_at_end=True,
    ...
)
# after training, save a normal transformers checkpoint
trainer.save_model("./best-model")

Setting save_only_model=True skips saving the full optimizer state, which means you can’t reload the best model at the end of training. Also set stage3_gather_16bit_weights_on_model_save: true to reconstruct full weights from their shards. This is required for saving a consolidated 16-bit model artifact or 16-bit state dict with ZeRO-3. Transformers raises an error when save_only_model=True is combined with load_best_model_at_end=True.

For resuming across different parallelism configurations, see DeepSpeed’s Universal Checkpointing guide.

Next steps

Update on GitHub