| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| from collections.abc import Callable, Mapping, Sequence |
| from typing import Any |
|
|
| import numpy as np |
| import torch |
| import torch.nn as nn |
|
|
| from monai.data.meta_tensor import MetaTensor |
| from monai.networks.utils import eval_mode |
| from monai.transforms import Compose, GaussianSmooth, Lambda, ScaleIntensity, SpatialCrop |
| from monai.utils import ensure_tuple_rep |
|
|
|
|
| class OcclusionSensitivity: |
| """ |
| This class computes the occlusion sensitivity for a model's prediction of a given image. By occlusion sensitivity, |
| we mean how the probability of a given prediction changes as the occluded section of an image changes. This can be |
| useful to understand why a network is making certain decisions. |
| |
| As important parts of the image are occluded, the probability of classifying the image correctly will decrease. |
| Hence, more negative values imply the corresponding occluded volume was more important in the decision process. |
| |
| Two ``torch.Tensor`` will be returned by the ``__call__`` method: an occlusion map and an image of the most probable |
| class. Both images will be cropped if a bounding box used, but voxel sizes will always match the input. |
| |
| The occlusion map shows the inference probabilities when the corresponding part of the image is occluded. Hence, |
| more -ve values imply that region was important in the decision process. The map will have shape ``BCHW(D)N``, |
| where ``N`` is the number of classes to be inferred by the network. Hence, the occlusion for class ``i`` can |
| be seen with ``map[...,i]``. |
| |
| The most probable class is an image of the probable class when the corresponding part of the image is occluded |
| (equivalent to ``occ_map.argmax(dim=-1)``). |
| |
| See: R. R. Selvaraju et al. Grad-CAM: Visual Explanations from Deep Networks via |
| Gradient-based Localization. https://doi.org/10.1109/ICCV.2017.74. |
| |
| Examples: |
| |
| .. code-block:: python |
| |
| # densenet 2d |
| from monai.networks.nets import DenseNet121 |
| from monai.visualize import OcclusionSensitivity |
| import torch |
| |
| model_2d = DenseNet121(spatial_dims=2, in_channels=1, out_channels=3) |
| occ_sens = OcclusionSensitivity(nn_module=model_2d) |
| occ_map, most_probable_class = occ_sens(x=torch.rand((1, 1, 48, 64)), b_box=[2, 40, 1, 62]) |
| |
| # densenet 3d |
| from monai.networks.nets import DenseNet |
| from monai.visualize import OcclusionSensitivity |
| |
| model_3d = DenseNet(spatial_dims=3, in_channels=1, out_channels=3, init_features=2, growth_rate=2, block_config=(6,)) |
| occ_sens = OcclusionSensitivity(nn_module=model_3d, n_batch=10) |
| occ_map, most_probable_class = occ_sens(torch.rand(1, 1, 6, 6, 6), b_box=[1, 3, -1, -1, -1, -1]) |
| |
| See Also: |
| |
| - :py:class:`monai.visualize.occlusion_sensitivity.OcclusionSensitivity.` |
| """ |
|
|
| def __init__( |
| self, |
| nn_module: nn.Module, |
| mask_size: int | Sequence = 16, |
| n_batch: int = 16, |
| verbose: bool = True, |
| mode: str | float | Callable = "gaussian", |
| overlap: float = 0.25, |
| activate: bool | Callable = True, |
| ) -> None: |
| """ |
| Occlusion sensitivity constructor. |
| |
| Args: |
| nn_module: Classification model to use for inference |
| mask_size: Size of box to be occluded, centred on the central voxel. If a single number |
| is given, this is used for all dimensions. If a sequence is given, this is used for each dimension |
| individually. |
| n_batch: Number of images in a batch for inference. |
| verbose: Use progress bar (if ``tqdm`` available). |
| mode: what should the occluded region be replaced with? If a float is given, that value will be used |
| throughout the occlusion. Else, ``gaussian``, ``mean_img`` and ``mean_patch`` can be supplied: |
| |
| * ``gaussian``: occluded region is multiplied by 1 - gaussian kernel. In this fashion, the occlusion |
| will be 0 at the center and will be unchanged towards the edges, varying smoothly between. When |
| gaussian is used, a weighted average will be used to combine overlapping regions. This will be |
| done using the gaussian (not 1-gaussian) as occluded regions count more. |
| * ``mean_patch``: occluded region will be replaced with the mean of occluded region. |
| * ``mean_img``: occluded region will be replaced with the mean of the whole image. |
| |
| overlap: overlap between inferred regions. Should be in range 0<=x<1. |
| activate: if ``True``, do softmax activation if num_channels > 1 else do ``sigmoid``. If ``False``, don't do any |
| activation. If ``callable``, use callable on inferred outputs. |
| |
| """ |
| self.nn_module = nn_module |
| self.mask_size = mask_size |
| self.n_batch = n_batch |
| self.verbose = verbose |
| self.overlap = overlap |
| self.activate = activate |
| |
| if isinstance(mode, str) and mode not in ("gaussian", "mean_patch", "mean_img"): |
| raise NotImplementedError |
| self.mode = mode |
|
|
| @staticmethod |
| def constant_occlusion(x: torch.Tensor, val: float, mask_size: Sequence) -> tuple[float, torch.Tensor]: |
| """Occlude with a constant occlusion. Multiplicative is zero, additive is constant value.""" |
| ones = torch.ones((*x.shape[:2], *mask_size), device=x.device, dtype=x.dtype) |
| return 0, ones * val |
|
|
| @staticmethod |
| def gaussian_occlusion(x: torch.Tensor, mask_size: Sequence, sigma: float = 0.25) -> tuple[torch.Tensor, float]: |
| """ |
| For Gaussian occlusion, Multiplicative is 1-Gaussian, additive is zero. |
| Default sigma of 0.25 empirically shown to give reasonable kernel, see here: |
| https://github.com/Project-MONAI/MONAI/pull/5230#discussion_r984520714. |
| """ |
| kernel = torch.zeros((x.shape[1], *mask_size), device=x.device, dtype=x.dtype) |
| spatial_shape = kernel.shape[1:] |
| |
| center = [slice(None)] + [slice(s // 2, s // 2 + 1) for s in spatial_shape] |
| |
| kernel[center] = 1.0 |
| |
| |
| gaussian = Compose( |
| [GaussianSmooth(sigma=[b * sigma for b in spatial_shape]), Lambda(lambda x: -x), ScaleIntensity()] |
| ) |
| |
| mul: torch.Tensor = gaussian(kernel)[None] |
|
|
| return mul, 0 |
|
|
| @staticmethod |
| def predictor( |
| cropped_grid: torch.Tensor, |
| nn_module: nn.Module, |
| x: torch.Tensor, |
| mul: torch.Tensor | float, |
| add: torch.Tensor | float, |
| mask_size: Sequence, |
| occ_mode: str, |
| activate: bool | Callable, |
| module_kwargs: Mapping[str, Any], |
| ) -> torch.Tensor: |
| """ |
| Predictor function to be passed to the sliding window inferer. Takes a cropped meshgrid, |
| referring to the coordinates in the input image. We use the index of the top-left corner |
| in combination ``mask_size`` to figure out which region of the image is to be occluded. The |
| occlusion is performed on the original image, ``x``, using ``cropped_region * mul + add``. ``mul`` |
| and ``add`` are sometimes pre-computed (e.g., a constant Gaussian blur), or they are |
| sometimes calculated on the fly (e.g., the mean of the occluded patch). For this reason |
| ``occ_mode`` is given. Lastly, ``activate`` is used to activate after each call of the model. |
| |
| Args: |
| cropped_grid: subsection of the meshgrid, where each voxel refers to the coordinate of |
| the input image. The meshgrid is created by the ``OcclusionSensitivity`` class, and |
| the generation of the subset is determined by ``sliding_window_inference``. |
| nn_module: module to call on data. |
| x: the image that was originally passed into ``OcclusionSensitivity.__call__``. |
| mul: occluded region will be multiplied by this. Can be ``torch.Tensor`` or ``float``. |
| add: after multiplication, this is added to the occluded region. Can be ``torch.Tensor`` or ``float``. |
| mask_size: Size of box to be occluded, centred on the central voxel. Should be |
| a sequence, one value for each spatial dimension. |
| occ_mode: might be used to calculate ``mul`` and ``add`` on the fly. |
| activate: if ``True``, do softmax activation if num_channels > 1 else do ``sigmoid``. If ``False``, don't do any |
| activation. If ``callable``, use callable on inferred outputs. |
| module_kwargs: kwargs to be passed onto module when inferring |
| """ |
| n_batch = cropped_grid.shape[0] |
| sd = cropped_grid.ndim - 2 |
| |
| im = torch.repeat_interleave(x, n_batch, 0) |
| |
| corner_coord_slices = [slice(None)] * 2 + [slice(1)] * sd |
| top_corners = cropped_grid[corner_coord_slices] |
|
|
| |
| for b, t in enumerate(top_corners): |
| |
| slices = [slice(b, b + 1), slice(None)] + [slice(int(j), int(j) + m) for j, m in zip(t, mask_size)] |
| to_occlude = im[slices] |
| if occ_mode == "mean_patch": |
| add, mul = OcclusionSensitivity.constant_occlusion(x, to_occlude.mean().item(), mask_size) |
|
|
| if callable(occ_mode): |
| to_occlude = occ_mode(x, to_occlude) |
| else: |
| to_occlude = to_occlude * mul + add |
| if add is None or mul is None: |
| raise RuntimeError("Shouldn't be here, something's gone wrong...") |
| im[slices] = to_occlude |
| |
| out: torch.Tensor = nn_module(im, **module_kwargs) |
|
|
| |
| if callable(activate): |
| out = activate(out) |
| |
| elif activate: |
| out = out.sigmoid() if x.shape[1] == 1 else out.softmax(1) |
|
|
| |
| |
| for m in mask_size: |
| out = torch.repeat_interleave(out.unsqueeze(-1), m, dim=-1) |
|
|
| return out |
|
|
| @staticmethod |
| def crop_meshgrid( |
| grid: MetaTensor, b_box: Sequence, mask_size: Sequence |
| ) -> tuple[MetaTensor, SpatialCrop, Sequence]: |
| """Crop the meshgrid so we only perform occlusion sensitivity on a subsection of the image.""" |
| |
| mask_edge = [(m - 1) // 2 for m in mask_size] |
| bbox_min = [max(b - m, 0) for b, m in zip(b_box[::2], mask_edge)] |
| bbox_max = [] |
| for b, m, s in zip(b_box[1::2], mask_edge, grid.shape[2:]): |
| |
| if b == -1: |
| bbox_max.append(s) |
| |
| else: |
| bbox_max.append(min(b + m, s)) |
| |
| |
| |
| slices = [slice(s, e) for s, e in zip(bbox_min, bbox_max)] |
| cropper = SpatialCrop(roi_slices=slices) |
| cropped: MetaTensor = cropper(grid[0])[None] |
| mask_size = list(mask_size) |
| for i, s in enumerate(cropped.shape[2:]): |
| mask_size[i] = min(s, mask_size[i]) |
| return cropped, cropper, mask_size |
|
|
| def __call__( |
| self, x: torch.Tensor, b_box: Sequence | None = None, **kwargs: Any |
| ) -> tuple[torch.Tensor, torch.Tensor]: |
| """ |
| Args: |
| x: Image to use for inference. Should be a tensor consisting of 1 batch. |
| b_box: Bounding box on which to perform the analysis. The output image will be limited to this size. |
| There should be a minimum and maximum for all spatial dimensions: ``[min1, max1, min2, max2,...]``. |
| * By default, the whole image will be used. Decreasing the size will speed the analysis up, which might |
| be useful for larger images. |
| * Min and max are inclusive, so ``[0, 63, ...]`` will have size ``(64, ...)``. |
| * Use -ve to use ``min=0`` and ``max=im.shape[x]-1`` for xth dimension. |
| * N.B.: we add half of the mask size to the bounding box to ensure that the region of interest has a |
| sufficiently large area surrounding it. |
| kwargs: any extra arguments to be passed on to the module as part of its `__call__`. |
| |
| Returns: |
| * Occlusion map: |
| * Shows the inference probabilities when the corresponding part of the image is occluded. |
| Hence, more -ve values imply that region was important in the decision process. |
| * The map will have shape ``BCHW(D)N``, where N is the number of classes to be inferred by the |
| network. Hence, the occlusion for class ``i`` can be seen with ``map[...,i]``. |
| * If `per_channel==False`, output ``C`` will equal 1: ``B1HW(D)N`` |
| * Most probable class: |
| * The most probable class when the corresponding part of the image is occluded (``argmax(dim=-1)``). |
| Both images will be cropped if a bounding box used, but voxel sizes will always match the input. |
| """ |
| if x.shape[0] > 1: |
| raise ValueError("Expected batch size of 1.") |
|
|
| sd = x.ndim - 2 |
| mask_size: Sequence = ensure_tuple_rep(self.mask_size, sd) |
|
|
| |
| grid: MetaTensor = MetaTensor( |
| np.stack(np.meshgrid(*[np.arange(0, i) for i in x.shape[2:]], indexing="ij"))[None], |
| device=x.device, |
| dtype=x.dtype, |
| ) |
| |
| if b_box is not None: |
| grid, cropper, mask_size = self.crop_meshgrid(grid, b_box, mask_size) |
|
|
| |
| if any(m > g for g, m in zip(grid.shape[2:], mask_size)): |
| raise ValueError(f"Image (spatial shape) {grid.shape[2:]} should be bigger than mask {mask_size}.") |
|
|
| |
| add: float | torch.Tensor | None |
| mul: float | torch.Tensor | None |
| |
| if isinstance(self.mode, float): |
| mul, add = self.constant_occlusion(x, self.mode, mask_size) |
| |
| elif self.mode == "mean_img": |
| mul, add = self.constant_occlusion(x, x.mean().item(), mask_size) |
| |
| elif self.mode == "gaussian": |
| mul, add = self.gaussian_occlusion(x, mask_size) |
| |
| else: |
| add, mul = None, None |
|
|
| with eval_mode(self.nn_module): |
| |
| from monai.inferers import sliding_window_inference |
|
|
| sensitivity_im: MetaTensor = sliding_window_inference( |
| grid, |
| roi_size=mask_size, |
| sw_batch_size=self.n_batch, |
| predictor=OcclusionSensitivity.predictor, |
| overlap=self.overlap, |
| mode="gaussian" if self.mode == "gaussian" else "constant", |
| progress=self.verbose, |
| nn_module=self.nn_module, |
| x=x, |
| add=add, |
| mul=mul, |
| mask_size=mask_size, |
| occ_mode=self.mode, |
| activate=self.activate, |
| module_kwargs=kwargs, |
| ) |
|
|
| if b_box is not None: |
| |
| sensitivity_im = cropper.inverse(sensitivity_im[0])[None] |
| |
| bbox_min = [max(b, 0) for b in b_box[::2]] |
| bbox_max = [b if b > 0 else s for b, s in zip(b_box[1::2], x.shape[2:])] |
| cropper = SpatialCrop(roi_start=bbox_min, roi_end=bbox_max) |
| sensitivity_im = cropper(sensitivity_im[0])[None] |
|
|
| |
| most_probable_class = sensitivity_im.argmax(dim=1, keepdim=True) |
| return sensitivity_im, most_probable_class |
|
|