| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| import warnings |
| from collections.abc import Hashable, Mapping |
| from contextlib import contextmanager |
| from typing import Any |
|
|
| import torch |
|
|
| from monai import transforms |
| from monai.data.meta_obj import MetaObj, get_track_meta |
| from monai.data.meta_tensor import MetaTensor |
| from monai.data.utils import to_affine_nd |
| from monai.transforms.traits import InvertibleTrait |
| from monai.transforms.transform import Transform |
| from monai.utils import ( |
| LazyAttr, |
| MetaKeys, |
| TraceKeys, |
| TraceStatusKeys, |
| convert_to_dst_type, |
| convert_to_numpy, |
| convert_to_tensor, |
| ) |
| from monai.utils.misc import MONAIEnvVars |
|
|
| __all__ = ["TraceableTransform", "InvertibleTransform"] |
|
|
|
|
| class TraceableTransform(Transform): |
| """ |
| Maintains a stack of applied transforms to data. |
| |
| Data can be one of two types: |
| 1. A `MetaTensor` (this is the preferred data type). |
| 2. A dictionary of data containing arrays/tensors and auxiliary metadata. In |
| this case, a key must be supplied (this dictionary-based approach is deprecated). |
| |
| If `data` is of type `MetaTensor`, then the applied transform will be added to ``data.applied_operations``. |
| |
| If `data` is a dictionary, then one of two things can happen: |
| 1. If data[key] is a `MetaTensor`, the applied transform will be added to ``data[key].applied_operations``. |
| 2. Else, the applied transform will be appended to an adjacent list using |
| `trace_key`. If, for example, the key is `image`, then the transform |
| will be appended to `image_transforms` (this dictionary-based approach is deprecated). |
| |
| Hopefully it is clear that there are three total possibilities: |
| 1. data is `MetaTensor` |
| 2. data is dictionary, data[key] is `MetaTensor` |
| 3. data is dictionary, data[key] is not `MetaTensor` (this is a deprecated approach). |
| |
| The ``__call__`` method of this transform class must be implemented so |
| that the transformation information is stored during the data transformation. |
| |
| The information in the stack of applied transforms must be compatible with the |
| default collate, by only storing strings, numbers and arrays. |
| |
| `tracing` could be enabled by `self.set_tracing` or setting |
| `MONAI_TRACE_TRANSFORM` when initializing the class. |
| """ |
|
|
| tracing = MONAIEnvVars.trace_transform() != "0" |
|
|
| def set_tracing(self, tracing: bool) -> None: |
| """Set whether to trace transforms.""" |
| self.tracing = tracing |
|
|
| @staticmethod |
| def trace_key(key: Hashable = None): |
| """The key to store the stack of applied transforms.""" |
| if key is None: |
| return f"{TraceKeys.KEY_SUFFIX}" |
| return f"{key}{TraceKeys.KEY_SUFFIX}" |
|
|
| @staticmethod |
| def transform_info_keys(): |
| """The keys to store necessary info of an applied transform.""" |
| return (TraceKeys.CLASS_NAME, TraceKeys.ID, TraceKeys.TRACING, TraceKeys.DO_TRANSFORM) |
|
|
| def get_transform_info(self) -> dict: |
| """ |
| Return a dictionary with the relevant information pertaining to an applied transform. |
| """ |
| vals = ( |
| self.__class__.__name__, |
| id(self), |
| self.tracing, |
| self._do_transform if hasattr(self, "_do_transform") else True, |
| ) |
| return dict(zip(self.transform_info_keys(), vals)) |
|
|
| def push_transform(self, data, *args, **kwargs): |
| """ |
| Push to a stack of applied transforms of ``data``. |
| |
| Args: |
| data: dictionary of data or `MetaTensor`. |
| args: additional positional arguments to track_transform_meta. |
| kwargs: additional keyword arguments to track_transform_meta, |
| set ``replace=True`` (default False) to rewrite the last transform infor in |
| applied_operation/pending_operation based on ``self.get_transform_info()``. |
| """ |
| lazy_eval = kwargs.get("lazy", False) |
| transform_info = self.get_transform_info() |
| do_transform = transform_info.get(TraceKeys.DO_TRANSFORM, True) |
| kwargs = kwargs or {} |
| replace = kwargs.pop("replace", False) |
| if replace and get_track_meta() and isinstance(data, MetaTensor): |
| if not lazy_eval: |
| xform = self.pop_transform(data, check=False) if do_transform else {} |
| meta_obj = self.push_transform(data, orig_size=xform.get(TraceKeys.ORIG_SIZE), extra_info=xform) |
| return data.copy_meta_from(meta_obj) |
| if do_transform: |
| xform = data.pending_operations.pop() |
| extra = xform.copy() |
| xform.update(transform_info) |
| else: |
| xform, extra = transform_info, {} |
| meta_obj = self.push_transform(data, transform_info=xform, lazy=True, extra_info=extra) |
| return data.copy_meta_from(meta_obj) |
| kwargs["lazy"] = lazy_eval |
| if "transform_info" in kwargs and isinstance(kwargs["transform_info"], dict): |
| kwargs["transform_info"].update(transform_info) |
| else: |
| kwargs["transform_info"] = transform_info |
| meta_obj = TraceableTransform.track_transform_meta(data, *args, **kwargs) |
| return data.copy_meta_from(meta_obj) if isinstance(data, MetaTensor) else data |
|
|
| @classmethod |
| def track_transform_meta( |
| cls, |
| data, |
| key: Hashable = None, |
| sp_size=None, |
| affine=None, |
| extra_info: dict | None = None, |
| orig_size: tuple | None = None, |
| transform_info=None, |
| lazy=False, |
| ): |
| """ |
| Update a stack of applied/pending transforms metadata of ``data``. |
| |
| Args: |
| data: dictionary of data or `MetaTensor`. |
| key: if data is a dictionary, data[key] will be modified. |
| sp_size: the expected output spatial size when the transform is applied. |
| it can be tensor or numpy, but will be converted to a list of integers. |
| affine: the affine representation of the (spatial) transform in the image space. |
| When the transform is applied, meta_tensor.affine will be updated to ``meta_tensor.affine @ affine``. |
| extra_info: if desired, any extra information pertaining to the applied |
| transform can be stored in this dictionary. These are often needed for |
| computing the inverse transformation. |
| orig_size: sometimes during the inverse it is useful to know what the size |
| of the original image was, in which case it can be supplied here. |
| transform_info: info from self.get_transform_info(). |
| lazy: whether to push the transform to pending_operations or applied_operations. |
| |
| Returns: |
| |
| For backward compatibility, if ``data`` is a dictionary, it returns the dictionary with |
| updated ``data[key]``. Otherwise, this function returns a MetaObj with updated transform metadata. |
| """ |
| data_t = data[key] if key is not None else data |
| out_obj = MetaObj() |
| |
| if isinstance(data_t, MetaTensor): |
| out_obj.copy_meta_from(data_t, keys=out_obj.__dict__.keys()) |
|
|
| if lazy and (not get_track_meta()): |
| warnings.warn("metadata is not tracked, please call 'set_track_meta(True)' if doing lazy evaluation.") |
|
|
| if not lazy and affine is not None and isinstance(data_t, MetaTensor): |
| |
| orig_affine = data_t.peek_pending_affine() |
| orig_affine = convert_to_dst_type(orig_affine, affine, dtype=torch.float64)[0] |
| try: |
| affine = orig_affine @ to_affine_nd(len(orig_affine) - 1, affine, dtype=torch.float64) |
| except RuntimeError as e: |
| if orig_affine.ndim > 2: |
| if data_t.is_batch: |
| msg = "Transform applied to batched tensor, should be applied to instances only" |
| else: |
| msg = "Mismatch affine matrix, ensured that the batch dimension is not included in the calculation." |
| raise RuntimeError(msg) from e |
| else: |
| raise |
| out_obj.meta[MetaKeys.AFFINE] = convert_to_tensor(affine, device=torch.device("cpu"), dtype=torch.float64) |
|
|
| if not (get_track_meta() and transform_info and transform_info.get(TraceKeys.TRACING)): |
| if isinstance(data, Mapping): |
| if not isinstance(data, dict): |
| data = dict(data) |
| data[key] = data_t.copy_meta_from(out_obj) if isinstance(data_t, MetaTensor) else data_t |
| return data |
| return out_obj |
|
|
| info = transform_info.copy() |
| |
| if orig_size is not None: |
| info[TraceKeys.ORIG_SIZE] = orig_size |
| elif isinstance(data_t, MetaTensor): |
| info[TraceKeys.ORIG_SIZE] = data_t.peek_pending_shape() |
| elif hasattr(data_t, "shape"): |
| info[TraceKeys.ORIG_SIZE] = data_t.shape[1:] |
|
|
| |
| info[TraceKeys.LAZY] = lazy |
|
|
| |
| if extra_info is not None: |
| extra_info.pop(LazyAttr.SHAPE, None) |
| extra_info.pop(LazyAttr.AFFINE, None) |
| info[TraceKeys.EXTRA_INFO] = extra_info |
|
|
| |
| if lazy: |
| if sp_size is None: |
| if LazyAttr.SHAPE not in info: |
| info[LazyAttr.SHAPE] = info.get(TraceKeys.ORIG_SIZE, []) |
| else: |
| info[LazyAttr.SHAPE] = sp_size |
| info[LazyAttr.SHAPE] = tuple(convert_to_numpy(info[LazyAttr.SHAPE], wrap_sequence=True).tolist()) |
| if affine is None: |
| if LazyAttr.AFFINE not in info: |
| info[LazyAttr.AFFINE] = MetaTensor.get_default_affine() |
| else: |
| info[LazyAttr.AFFINE] = affine |
| info[LazyAttr.AFFINE] = convert_to_tensor(info[LazyAttr.AFFINE], device=torch.device("cpu")) |
| out_obj.push_pending_operation(info) |
| else: |
| if out_obj.pending_operations: |
| transform_name = info.get(TraceKeys.CLASS_NAME, "") if isinstance(info, dict) else "" |
| msg = ( |
| f"Transform {transform_name} has been applied to a MetaTensor with pending operations: " |
| f"{[x.get(TraceKeys.CLASS_NAME) for x in out_obj.pending_operations]}" |
| ) |
| if key is not None: |
| msg += f" for key {key}" |
|
|
| pend = out_obj.pending_operations[-1] |
| statuses = pend.get(TraceKeys.STATUSES, dict()) |
| messages = statuses.get(TraceStatusKeys.PENDING_DURING_APPLY, list()) |
| messages.append(msg) |
| statuses[TraceStatusKeys.PENDING_DURING_APPLY] = messages |
| info[TraceKeys.STATUSES] = statuses |
| out_obj.push_applied_operation(info) |
| if isinstance(data, Mapping): |
| if not isinstance(data, dict): |
| data = dict(data) |
| if isinstance(data_t, MetaTensor): |
| data[key] = data_t.copy_meta_from(out_obj) |
| else: |
| x_k = TraceableTransform.trace_key(key) |
| if x_k not in data: |
| data[x_k] = [] |
| data[x_k].append(info) |
| return data |
| return out_obj |
|
|
| def check_transforms_match(self, transform: Mapping) -> None: |
| """Check transforms are of same instance.""" |
| xform_id = transform.get(TraceKeys.ID, "") |
| if xform_id == id(self): |
| return |
| |
| if xform_id == TraceKeys.NONE: |
| return |
| xform_name = transform.get(TraceKeys.CLASS_NAME, "") |
| warning_msg = transform.get(TraceKeys.EXTRA_INFO, {}).get("warn") |
| if warning_msg: |
| warnings.warn(warning_msg) |
| |
| if torch.multiprocessing.get_start_method() in ("spawn", None) and xform_name == self.__class__.__name__: |
| return |
| raise RuntimeError( |
| f"Error {self.__class__.__name__} getting the most recently " |
| f"applied invertible transform {xform_name} {xform_id} != {id(self)}." |
| ) |
|
|
| def get_most_recent_transform(self, data, key: Hashable = None, check: bool = True, pop: bool = False): |
| """ |
| Get most recent transform for the stack. |
| |
| Args: |
| data: dictionary of data or `MetaTensor`. |
| key: if data is a dictionary, data[key] will be modified. |
| check: if true, check that `self` is the same type as the most recently-applied transform. |
| pop: if true, remove the transform as it is returned. |
| |
| Returns: |
| Dictionary of most recently applied transform |
| |
| Raises: |
| - RuntimeError: data is neither `MetaTensor` nor dictionary |
| """ |
| if not self.tracing: |
| raise RuntimeError("Transform Tracing must be enabled to get the most recent transform.") |
| if isinstance(data, MetaTensor): |
| all_transforms = data.applied_operations |
| elif isinstance(data, Mapping): |
| if key in data and isinstance(data[key], MetaTensor): |
| all_transforms = data[key].applied_operations |
| else: |
| all_transforms = data.get(self.trace_key(key), MetaTensor.get_default_applied_operations()) |
| else: |
| raise ValueError(f"`data` should be either `MetaTensor` or dictionary, got {type(data)}.") |
| if check: |
| self.check_transforms_match(all_transforms[-1]) |
| return all_transforms.pop() if pop else all_transforms[-1] |
|
|
| def pop_transform(self, data, key: Hashable = None, check: bool = True): |
| """ |
| Return and pop the most recent transform. |
| |
| Args: |
| data: dictionary of data or `MetaTensor` |
| key: if data is a dictionary, data[key] will be modified |
| check: if true, check that `self` is the same type as the most recently-applied transform. |
| |
| Returns: |
| Dictionary of most recently applied transform |
| |
| Raises: |
| - RuntimeError: data is neither `MetaTensor` nor dictionary |
| """ |
| return self.get_most_recent_transform(data, key, check, pop=True) |
|
|
| @contextmanager |
| def trace_transform(self, to_trace: bool): |
| """Temporarily set the tracing status of a transform with a context manager.""" |
| prev = self.tracing |
| self.tracing = to_trace |
| yield |
| self.tracing = prev |
|
|
|
|
| class InvertibleTransform(TraceableTransform, InvertibleTrait): |
| """Classes for invertible transforms. |
| |
| This class exists so that an ``invert`` method can be implemented. This allows, for |
| example, images to be cropped, rotated, padded, etc., during training and inference, |
| and after be returned to their original size before saving to file for comparison in |
| an external viewer. |
| |
| When the ``inverse`` method is called: |
| |
| - the inverse is called on each key individually, which allows for |
| different parameters being passed to each label (e.g., different |
| interpolation for image and label). |
| |
| - the inverse transforms are applied in a last-in-first-out order. As |
| the inverse is applied, its entry is removed from the list detailing |
| the applied transformations. That is to say that during the forward |
| pass, the list of applied transforms grows, and then during the |
| inverse it shrinks back down to an empty list. |
| |
| We currently check that the ``id()`` of the transform is the same in the forward and |
| inverse directions. This is a useful check to ensure that the inverses are being |
| processed in the correct order. |
| |
| Note to developers: When converting a transform to an invertible transform, you need to: |
| |
| #. Inherit from this class. |
| #. In ``__call__``, add a call to ``push_transform``. |
| #. Any extra information that might be needed for the inverse can be included with the |
| dictionary ``extra_info``. This dictionary should have the same keys regardless of |
| whether ``do_transform`` was `True` or `False` and can only contain objects that are |
| accepted in pytorch data loader's collate function (e.g., `None` is not allowed). |
| #. Implement an ``inverse`` method. Make sure that after performing the inverse, |
| ``pop_transform`` is called. |
| |
| """ |
|
|
| def inverse_update(self, data): |
| """ |
| This function is to be called before every `self.inverse(data)`, |
| update each MetaTensor `data[key]` using `data[key_transforms]` and `data[key_meta_dict]`, |
| for MetaTensor backward compatibility 0.9.0. |
| """ |
| if not isinstance(data, dict) or not isinstance(self, transforms.MapTransform): |
| return data |
| d = dict(data) |
| for k in self.key_iterator(data): |
| transform_key = transforms.TraceableTransform.trace_key(k) |
| if transform_key not in data or not data[transform_key]: |
| continue |
| d = transforms.sync_meta_info(k, data, t=False) |
| return d |
|
|
| def inverse(self, data: Any) -> Any: |
| """ |
| Inverse of ``__call__``. |
| |
| Raises: |
| NotImplementedError: When the subclass does not override this method. |
| |
| """ |
| raise NotImplementedError(f"Subclass {self.__class__.__name__} must implement this method.") |
|
|