| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """ |
| A collection of generic interfaces for MONAI transforms. |
| """ |
|
|
| from __future__ import annotations |
|
|
| import logging |
| from abc import ABC, abstractmethod |
| from collections.abc import Callable, Generator, Hashable, Iterable, Mapping |
| from typing import Any, TypeVar |
|
|
| import numpy as np |
| import torch |
|
|
| from monai import config, transforms |
| from monai.config import KeysCollection |
| from monai.data.meta_tensor import MetaTensor |
| from monai.transforms.traits import LazyTrait, RandomizableTrait, ThreadUnsafe |
| from monai.utils import MAX_SEED, ensure_tuple, first |
| from monai.utils.enums import TransformBackends |
| from monai.utils.misc import MONAIEnvVars |
|
|
| __all__ = [ |
| "ThreadUnsafe", |
| "apply_transform", |
| "Randomizable", |
| "LazyTransform", |
| "RandomizableTransform", |
| "Transform", |
| "MapTransform", |
| ] |
|
|
| ReturnType = TypeVar("ReturnType") |
|
|
|
|
| def _apply_transform( |
| transform: Callable[..., ReturnType], |
| data: Any, |
| unpack_parameters: bool = False, |
| lazy: bool | None = False, |
| overrides: dict | None = None, |
| logger_name: bool | str = False, |
| ) -> ReturnType: |
| """ |
| Perform a transform 'transform' on 'data', according to the other parameters specified. |
| |
| If `data` is a tuple and `unpack_parameters` is True, each parameter of `data` is unpacked |
| as arguments to `transform`. Otherwise `data` is considered as single argument to `transform`. |
| |
| If 'lazy' is True, this method first checks whether it can execute this method lazily. If it |
| can't, it will ensure that all pending lazy transforms on 'data' are applied before applying |
| this 'transform' to it. If 'lazy' is True, and 'overrides' are provided, those overrides will |
| be applied to the pending operations on 'data'. See ``Compose`` for more details on lazy |
| resampling, which is an experimental feature for 1.2. |
| |
| Please note, this class is function is designed to be called by ``apply_transform``. |
| In general, you should not need to make specific use of it unless you are implementing |
| pipeline execution mechanisms. |
| |
| Args: |
| transform: a callable to be used to transform `data`. |
| data: the tensorlike or dictionary of tensorlikes to be executed on |
| unpack_parameters: whether to unpack parameters for `transform`. Defaults to False. |
| lazy: whether to enable lazy evaluation for lazy transforms. If False, transforms will be |
| carried out on a transform by transform basis. If True, all lazy transforms will |
| be executed by accumulating changes and resampling as few times as possible. |
| See the :ref:`Lazy Resampling topic<lazy_resampling> for more information about lazy resampling. |
| overrides: this optional parameter allows you to specify a dictionary of parameters that should be overridden |
| when executing a pipeline. These each parameter that is compatible with a given transform is then applied |
| to that transform before it is executed. Note that overrides are currently only applied when |
| :ref:`Lazy Resampling<lazy_resampling>` is enabled for the pipeline or a given transform. If lazy is False |
| they are ignored. Currently supported args are: |
| {``"mode"``, ``"padding_mode"``, ``"dtype"``, ``"align_corners"``, ``"resample_mode"``, ``device``}. |
| logger_name: this optional parameter allows you to specify a logger by name for logging of pipeline execution. |
| Setting this to False disables logging. Setting it to True enables logging to the default loggers. |
| Setting a string overrides the logger name to which logging is performed. |
| |
| Returns: |
| ReturnType: The return type of `transform`. |
| """ |
| from monai.transforms.lazy.functional import apply_pending_transforms_in_order |
|
|
| data = apply_pending_transforms_in_order(transform, data, lazy, overrides, logger_name) |
|
|
| if isinstance(data, tuple) and unpack_parameters: |
| return transform(*data, lazy=lazy) if isinstance(transform, LazyTrait) else transform(*data) |
|
|
| return transform(data, lazy=lazy) if isinstance(transform, LazyTrait) else transform(data) |
|
|
|
|
| def apply_transform( |
| transform: Callable[..., ReturnType], |
| data: Any, |
| map_items: bool = True, |
| unpack_items: bool = False, |
| log_stats: bool | str = False, |
| lazy: bool | None = None, |
| overrides: dict | None = None, |
| ) -> list[ReturnType] | ReturnType: |
| """ |
| Transform `data` with `transform`. |
| |
| If `data` is a list or tuple and `map_data` is True, each item of `data` will be transformed |
| and this method returns a list of outcomes. |
| otherwise transform will be applied once with `data` as the argument. |
| |
| Args: |
| transform: a callable to be used to transform `data`. |
| data: an object to be transformed. |
| map_items: whether to apply transform to each item in `data`, |
| if `data` is a list or tuple. Defaults to True. |
| unpack_items: whether to unpack parameters using `*`. Defaults to False. |
| log_stats: log errors when they occur in the processing pipeline. By default, this is set to False, which |
| disables the logger for processing pipeline errors. Setting it to None or True will enable logging to the |
| default logger name. Setting it to a string specifies the logger to which errors should be logged. |
| lazy: whether to execute in lazy mode or not. See the :ref:`Lazy Resampling topic<lazy_resampling> for more |
| information about lazy resampling. Defaults to None. |
| overrides: optional overrides to apply to transform parameters. This parameter is ignored unless transforms |
| are being executed lazily. See the :ref:`Lazy Resampling topic<lazy_resampling> for more details and |
| examples of its usage. |
| |
| Raises: |
| Exception: When ``transform`` raises an exception. |
| |
| Returns: |
| Union[List[ReturnType], ReturnType]: The return type of `transform` or a list thereof. |
| """ |
| try: |
| if isinstance(data, (list, tuple)) and map_items: |
| return [_apply_transform(transform, item, unpack_items, lazy, overrides, log_stats) for item in data] |
| return _apply_transform(transform, data, unpack_items, lazy, overrides, log_stats) |
| except Exception as e: |
| |
| |
| if MONAIEnvVars.debug(): |
| raise |
| if log_stats is not False and not isinstance(transform, transforms.compose.Compose): |
| |
| if isinstance(log_stats, str): |
| datastats = transforms.utility.array.DataStats(data_shape=False, value_range=False, name=log_stats) |
| else: |
| datastats = transforms.utility.array.DataStats(data_shape=False, value_range=False) |
| logger = logging.getLogger(datastats._logger_name) |
| logger.error(f"\n=== Transform input info -- {type(transform).__name__} ===") |
| if isinstance(data, (list, tuple)): |
| data = data[0] |
|
|
| def _log_stats(data, prefix: str | None = "Data"): |
| if isinstance(data, (np.ndarray, torch.Tensor)): |
| |
| datastats(img=data, data_shape=True, value_range=True, prefix=prefix) |
| else: |
| |
| datastats(img=data, data_value=True, prefix=prefix) |
|
|
| if isinstance(data, dict): |
| for k, v in data.items(): |
| _log_stats(data=v, prefix=k) |
| else: |
| _log_stats(data=data) |
| raise RuntimeError(f"applying transform {transform}") from e |
|
|
|
|
| class Randomizable(ThreadUnsafe, RandomizableTrait): |
| """ |
| An interface for handling random state locally, currently based on a class |
| variable `R`, which is an instance of `np.random.RandomState`. This |
| provides the flexibility of component-specific determinism without |
| affecting the global states. It is recommended to use this API with |
| :py:class:`monai.data.DataLoader` for deterministic behaviour of the |
| preprocessing pipelines. This API is not thread-safe. Additionally, |
| deepcopying instance of this class often causes insufficient randomness as |
| the random states will be duplicated. |
| """ |
|
|
| R: np.random.RandomState = np.random.RandomState() |
|
|
| def set_random_state(self, seed: int | None = None, state: np.random.RandomState | None = None) -> Randomizable: |
| """ |
| Set the random state locally, to control the randomness, the derived |
| classes should use :py:attr:`self.R` instead of `np.random` to introduce random |
| factors. |
| |
| Args: |
| seed: set the random state with an integer seed. |
| state: set the random state with a `np.random.RandomState` object. |
| |
| Raises: |
| TypeError: When ``state`` is not an ``Optional[np.random.RandomState]``. |
| |
| Returns: |
| a Randomizable instance. |
| |
| """ |
| if seed is not None: |
| _seed = id(seed) if not isinstance(seed, (int, np.integer)) else seed |
| _seed = _seed % MAX_SEED |
| self.R = np.random.RandomState(_seed) |
| return self |
|
|
| if state is not None: |
| if not isinstance(state, np.random.RandomState): |
| raise TypeError(f"state must be None or a np.random.RandomState but is {type(state).__name__}.") |
| self.R = state |
| return self |
|
|
| self.R = np.random.RandomState() |
| return self |
|
|
| def randomize(self, data: Any) -> None: |
| """ |
| Within this method, :py:attr:`self.R` should be used, instead of `np.random`, to introduce random factors. |
| |
| all :py:attr:`self.R` calls happen here so that we have a better chance to |
| identify errors of sync the random state. |
| |
| This method can generate the random factors based on properties of the input data. |
| |
| Raises: |
| NotImplementedError: When the subclass does not override this method. |
| """ |
| raise NotImplementedError(f"Subclass {self.__class__.__name__} must implement this method.") |
|
|
|
|
| class Transform(ABC): |
| """ |
| An abstract class of a ``Transform``. |
| A transform is callable that processes ``data``. |
| |
| It could be stateful and may modify ``data`` in place, |
| the implementation should be aware of: |
| |
| #. thread safety when mutating its own states. |
| When used from a multi-process context, transform's instance variables are read-only. |
| thread-unsafe transforms should inherit :py:class:`monai.transforms.ThreadUnsafe`. |
| #. ``data`` content unused by this transform may still be used in the |
| subsequent transforms in a composed transform. |
| #. storing too much information in ``data`` may cause some memory issue or IPC sync issue, |
| especially in the multi-processing environment of PyTorch DataLoader. |
| |
| See Also |
| |
| :py:class:`monai.transforms.Compose` |
| """ |
|
|
| |
| |
| |
| |
| |
| |
| |
| backend: list[TransformBackends] = [] |
|
|
| @abstractmethod |
| def __call__(self, data: Any): |
| """ |
| ``data`` is an element which often comes from an iteration over an |
| iterable, such as :py:class:`torch.utils.data.Dataset`. This method should |
| return an updated version of ``data``. |
| To simplify the input validations, most of the transforms assume that |
| |
| - ``data`` is a Numpy ndarray, PyTorch Tensor or string, |
| - the data shape can be: |
| |
| #. string data without shape, `LoadImage` transform expects file paths, |
| #. most of the pre-/post-processing transforms expect: ``(num_channels, spatial_dim_1[, spatial_dim_2, ...])``, |
| except for example: `AddChannel` expects (spatial_dim_1[, spatial_dim_2, ...]) |
| |
| - the channel dimension is often not omitted even if number of channels is one. |
| |
| This method can optionally take additional arguments to help execute transformation operation. |
| |
| Raises: |
| NotImplementedError: When the subclass does not override this method. |
| |
| """ |
| raise NotImplementedError(f"Subclass {self.__class__.__name__} must implement this method.") |
|
|
|
|
| class LazyTransform(Transform, LazyTrait): |
| """ |
| An implementation of functionality for lazy transforms that can be subclassed by array and |
| dictionary transforms to simplify implementation of new lazy transforms. |
| """ |
|
|
| def __init__(self, lazy: bool | None = False): |
| if lazy is not None: |
| if not isinstance(lazy, bool): |
| raise TypeError(f"lazy must be a bool but is of type {type(lazy)}") |
| self._lazy = lazy |
|
|
| @property |
| def lazy(self): |
| return self._lazy |
|
|
| @lazy.setter |
| def lazy(self, lazy: bool | None): |
| if lazy is not None: |
| if not isinstance(lazy, bool): |
| raise TypeError(f"lazy must be a bool but is of type {type(lazy)}") |
| self._lazy = lazy |
|
|
| @property |
| def requires_current_data(self): |
| return False |
|
|
|
|
| class RandomizableTransform(Randomizable, Transform): |
| """ |
| An interface for handling random state locally, currently based on a class variable `R`, |
| which is an instance of `np.random.RandomState`. |
| This class introduces a randomized flag `_do_transform`, is mainly for randomized data augmentation transforms. |
| For example: |
| |
| .. code-block:: python |
| |
| from monai.transforms import RandomizableTransform |
| |
| class RandShiftIntensity100(RandomizableTransform): |
| def randomize(self): |
| super().randomize(None) |
| self._offset = self.R.uniform(low=0, high=100) |
| |
| def __call__(self, img): |
| self.randomize() |
| if not self._do_transform: |
| return img |
| return img + self._offset |
| |
| transform = RandShiftIntensity() |
| transform.set_random_state(seed=0) |
| print(transform(10)) |
| |
| """ |
|
|
| def __init__(self, prob: float = 1.0, do_transform: bool = True): |
| self._do_transform = do_transform |
| self.prob = min(max(prob, 0.0), 1.0) |
|
|
| def randomize(self, data: Any) -> None: |
| """ |
| Within this method, :py:attr:`self.R` should be used, instead of `np.random`, to introduce random factors. |
| |
| all :py:attr:`self.R` calls happen here so that we have a better chance to |
| identify errors of sync the random state. |
| |
| This method can generate the random factors based on properties of the input data. |
| """ |
| self._do_transform = self.R.rand() < self.prob |
|
|
|
|
| class MapTransform(Transform): |
| """ |
| A subclass of :py:class:`monai.transforms.Transform` with an assumption |
| that the ``data`` input of ``self.__call__`` is a MutableMapping such as ``dict``. |
| |
| The ``keys`` parameter will be used to get and set the actual data |
| item to transform. That is, the callable of this transform should |
| follow the pattern: |
| |
| .. code-block:: python |
| |
| def __call__(self, data): |
| for key in self.keys: |
| if key in data: |
| # update output data with some_transform_function(data[key]). |
| else: |
| # raise exception unless allow_missing_keys==True. |
| return data |
| |
| Raises: |
| ValueError: When ``keys`` is an empty iterable. |
| TypeError: When ``keys`` type is not in ``Union[Hashable, Iterable[Hashable]]``. |
| |
| """ |
|
|
| def __new__(cls, *args, **kwargs): |
| if config.USE_META_DICT: |
| |
| cls.__call__ = transforms.attach_hook(cls.__call__, MapTransform.call_update, "post") |
|
|
| if hasattr(cls, "inverse"): |
| |
| cls.inverse: Any = transforms.attach_hook(cls.inverse, transforms.InvertibleTransform.inverse_update) |
| return Transform.__new__(cls) |
|
|
| def __init__(self, keys: KeysCollection, allow_missing_keys: bool = False) -> None: |
| super().__init__() |
| self.keys: tuple[Hashable, ...] = ensure_tuple(keys) |
| self.allow_missing_keys = allow_missing_keys |
| if not self.keys: |
| raise ValueError("keys must be non empty.") |
| for key in self.keys: |
| if not isinstance(key, Hashable): |
| raise TypeError(f"keys must be one of (Hashable, Iterable[Hashable]) but is {type(keys).__name__}.") |
|
|
| def call_update(self, data): |
| """ |
| This function is to be called after every `self.__call__(data)`, |
| update `data[key_transforms]` and `data[key_meta_dict]` using the content from MetaTensor `data[key]`, |
| for MetaTensor backward compatibility 0.9.0. |
| """ |
| if not isinstance(data, (list, tuple, Mapping)): |
| return data |
| is_dict = False |
| if isinstance(data, Mapping): |
| data, is_dict = [data], True |
| if not data or not isinstance(data[0], Mapping): |
| return data[0] if is_dict else data |
| list_d = [dict(x) for x in data] |
| for idx, dict_i in enumerate(list_d): |
| for k in dict_i: |
| if not isinstance(dict_i[k], MetaTensor): |
| continue |
| list_d[idx] = transforms.sync_meta_info(k, dict_i, t=not isinstance(self, transforms.InvertD)) |
| return list_d[0] if is_dict else list_d |
|
|
| @abstractmethod |
| def __call__(self, data): |
| """ |
| ``data`` often comes from an iteration over an iterable, |
| such as :py:class:`torch.utils.data.Dataset`. |
| |
| To simplify the input validations, this method assumes: |
| |
| - ``data`` is a Python dictionary, |
| - ``data[key]`` is a Numpy ndarray, PyTorch Tensor or string, where ``key`` is an element |
| of ``self.keys``, the data shape can be: |
| |
| #. string data without shape, `LoadImaged` transform expects file paths, |
| #. most of the pre-/post-processing transforms expect: ``(num_channels, spatial_dim_1[, spatial_dim_2, ...])``, |
| except for example: `AddChanneld` expects (spatial_dim_1[, spatial_dim_2, ...]) |
| |
| - the channel dimension is often not omitted even if number of channels is one. |
| |
| Raises: |
| NotImplementedError: When the subclass does not override this method. |
| |
| returns: |
| An updated dictionary version of ``data`` by applying the transform. |
| |
| """ |
| raise NotImplementedError(f"Subclass {self.__class__.__name__} must implement this method.") |
|
|
| def key_iterator(self, data: Mapping[Hashable, Any], *extra_iterables: Iterable | None) -> Generator: |
| """ |
| Iterate across keys and optionally extra iterables. If key is missing, exception is raised if |
| `allow_missing_keys==False` (default). If `allow_missing_keys==True`, key is skipped. |
| |
| Args: |
| data: data that the transform will be applied to |
| extra_iterables: anything else to be iterated through |
| """ |
| |
| ex_iters = extra_iterables or [[None] * len(self.keys)] |
|
|
| |
| _ex_iters: list[Any] |
| for key, *_ex_iters in zip(self.keys, *ex_iters): |
| |
| if key in data: |
| yield (key,) + tuple(_ex_iters) if extra_iterables else key |
| elif not self.allow_missing_keys: |
| raise KeyError( |
| f"Key `{key}` of transform `{self.__class__.__name__}` was missing in the data" |
| " and allow_missing_keys==False." |
| ) |
|
|
| def first_key(self, data: dict[Hashable, Any]): |
| """ |
| Get the first available key of `self.keys` in the input `data` dictionary. |
| If no available key, return an empty tuple `()`. |
| |
| Args: |
| data: data that the transform will be applied to. |
| |
| """ |
| return first(self.key_iterator(data), ()) |
|
|