| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import os |
| from contextlib import contextmanager |
|
|
| import torch |
|
|
| from ..commands.config.default import write_basic_config |
| from ..state import AcceleratorState |
| from .dataclasses import DistributedType |
| from .imports import is_deepspeed_available, is_tpu_available |
|
|
|
|
| if is_deepspeed_available(): |
| from deepspeed import DeepSpeedEngine |
|
|
| if is_tpu_available(check_device=False): |
| import torch_xla.core.xla_model as xm |
|
|
|
|
| def extract_model_from_parallel(model, keep_fp32_wrapper: bool = False): |
| """ |
| Extract a model from its distributed containers. |
| |
| Args: |
| model (`torch.nn.Module`): |
| The model to extract. |
| keep_fp32_wrapper (`bool`, *optional*): |
| Whether to remove mixed precision hooks from the model. |
| |
| Returns: |
| `torch.nn.Module`: The extracted model. |
| """ |
| options = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) |
| if is_deepspeed_available(): |
| options += (DeepSpeedEngine,) |
|
|
| while isinstance(model, options): |
| model = model.module |
|
|
| if not keep_fp32_wrapper: |
| forward = getattr(model, "forward") |
| original_forward = model.__dict__.pop("_original_forward", None) |
| if original_forward is not None: |
| while hasattr(forward, "__wrapped__"): |
| forward = forward.__wrapped__ |
| if forward == original_forward: |
| break |
| model.forward = forward |
| return model |
|
|
|
|
| def wait_for_everyone(): |
| """ |
| Introduces a blocking point in the script, making sure all processes have reached this point before continuing. |
| |
| <Tip warning={true}> |
| |
| Make sure all processes will reach this instruction otherwise one of your processes will hang forever. |
| |
| </Tip> |
| """ |
| if ( |
| AcceleratorState().distributed_type == DistributedType.MULTI_GPU |
| or AcceleratorState().distributed_type == DistributedType.MULTI_CPU |
| or AcceleratorState().distributed_type == DistributedType.DEEPSPEED |
| or AcceleratorState().distributed_type == DistributedType.FSDP |
| ): |
| torch.distributed.barrier() |
| elif AcceleratorState().distributed_type == DistributedType.TPU: |
| xm.rendezvous("accelerate.utils.wait_for_everyone") |
|
|
|
|
| def save(obj, f): |
| """ |
| Save the data to disk. Use in place of `torch.save()`. |
| |
| Args: |
| obj: The data to save |
| f: The file (or file-like object) to use to save the data |
| """ |
| if AcceleratorState().distributed_type == DistributedType.TPU: |
| xm.save(obj, f) |
| elif AcceleratorState().local_process_index == 0: |
| torch.save(obj, f) |
|
|
|
|
| @contextmanager |
| def patch_environment(**kwargs): |
| """ |
| A context manager that will add each keyword argument passed to `os.environ` and remove them when exiting. |
| |
| Will convert the values in `kwargs` to strings and upper-case all the keys. |
| """ |
| for key, value in kwargs.items(): |
| os.environ[key.upper()] = str(value) |
|
|
| yield |
|
|
| for key in kwargs: |
| del os.environ[key.upper()] |
|
|
|
|
| def get_pretty_name(obj): |
| """ |
| Gets a pretty name from `obj`. |
| """ |
| if not hasattr(obj, "__qualname__") and not hasattr(obj, "__name__"): |
| obj = getattr(obj, "__class__", obj) |
| if hasattr(obj, "__qualname__"): |
| return obj.__qualname__ |
| if hasattr(obj, "__name__"): |
| return obj.__name__ |
| return str(obj) |
|
|