| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| from functools import partial |
| from typing import Any, Callable |
|
|
| import torch |
|
|
| from monai.networks.utils import replace_modules_temp |
| from monai.utils.module import optional_import |
| from monai.visualize.class_activation_maps import ModelWithHooks |
|
|
| trange, has_trange = optional_import("tqdm", name="trange") |
|
|
| __all__ = ["VanillaGrad", "SmoothGrad", "GuidedBackpropGrad", "GuidedBackpropSmoothGrad"] |
|
|
|
|
| class _AutoGradReLU(torch.autograd.Function): |
|
|
| @staticmethod |
| def forward(ctx, x): |
| pos_mask = (x > 0).type_as(x) |
| output = torch.mul(x, pos_mask) |
| ctx.save_for_backward(x, output) |
| return output |
|
|
| @staticmethod |
| def backward(ctx, grad_output): |
| x, _ = ctx.saved_tensors |
| pos_mask_1 = (x > 0).type_as(grad_output) |
| pos_mask_2 = (grad_output > 0).type_as(grad_output) |
| y = torch.mul(grad_output, pos_mask_1) |
| grad_input = torch.mul(y, pos_mask_2) |
| return grad_input |
|
|
|
|
| class _GradReLU(torch.nn.Module): |
| """ |
| A customized ReLU with the backward pass imputed for guided backpropagation (https://arxiv.org/abs/1412.6806). |
| """ |
|
|
| def forward(self, x: torch.Tensor) -> torch.Tensor: |
| out: torch.Tensor = _AutoGradReLU.apply(x) |
| return out |
|
|
|
|
| class VanillaGrad: |
| """ |
| Given an input image ``x``, calling this class will perform the forward pass, then set to zero |
| all activations except one (defined by ``index``) and propagate back to the image to achieve a gradient-based |
| saliency map. |
| |
| If ``index`` is None, argmax of the output logits will be used. |
| |
| See also: |
| |
| - Simonyan et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps |
| (https://arxiv.org/abs/1312.6034) |
| """ |
|
|
| def __init__(self, model: torch.nn.Module) -> None: |
| if not isinstance(model, ModelWithHooks): |
| self._model = ModelWithHooks(model, target_layer_names=(), register_backward=True) |
| else: |
| self._model = model |
|
|
| @property |
| def model(self): |
| return self._model.model |
|
|
| @model.setter |
| def model(self, m): |
| if not isinstance(m, ModelWithHooks): |
| self._model.model = m |
| else: |
| self._model = m |
|
|
| def get_grad( |
| self, x: torch.Tensor, index: torch.Tensor | int | None, retain_graph: bool = True, **kwargs: Any |
| ) -> torch.Tensor: |
| if x.shape[0] != 1: |
| raise ValueError("expect batch size of 1") |
| x.requires_grad = True |
|
|
| self._model(x, class_idx=index, retain_graph=retain_graph, **kwargs) |
| grad: torch.Tensor = x.grad.detach() |
| return grad |
|
|
| def __call__(self, x: torch.Tensor, index: torch.Tensor | int | None = None, **kwargs: Any) -> torch.Tensor: |
| return self.get_grad(x, index, **kwargs) |
|
|
|
|
| class SmoothGrad(VanillaGrad): |
| """ |
| Compute averaged sensitivity map based on ``n_samples`` (Gaussian additive) of noisy versions |
| of the input image ``x``. |
| |
| See also: |
| |
| - Smilkov et al. SmoothGrad: removing noise by adding noise https://arxiv.org/abs/1706.03825 |
| """ |
|
|
| def __init__( |
| self, |
| model: torch.nn.Module, |
| stdev_spread: float = 0.15, |
| n_samples: int = 25, |
| magnitude: bool = True, |
| verbose: bool = True, |
| ) -> None: |
| super().__init__(model) |
| self.stdev_spread = stdev_spread |
| self.n_samples = n_samples |
| self.magnitude = magnitude |
| self.range: Callable |
| if verbose and has_trange: |
| self.range = partial(trange, desc=f"Computing {self.__class__.__name__}") |
| else: |
| self.range = range |
|
|
| def __call__(self, x: torch.Tensor, index: torch.Tensor | int | None = None, **kwargs: Any) -> torch.Tensor: |
| stdev = (self.stdev_spread * (x.max() - x.min())).item() |
| total_gradients = torch.zeros_like(x) |
| for _ in self.range(self.n_samples): |
| |
| noise = torch.normal(0, stdev, size=x.shape, dtype=torch.float32, device=x.device) |
| x_plus_noise = x + noise |
| x_plus_noise = x_plus_noise.detach() |
|
|
| |
| grad = self.get_grad(x_plus_noise, index, **kwargs) |
| total_gradients += (grad * grad) if self.magnitude else grad |
|
|
| |
| if self.magnitude: |
| total_gradients = total_gradients**0.5 |
|
|
| return total_gradients / self.n_samples |
|
|
|
|
| class GuidedBackpropGrad(VanillaGrad): |
| """ |
| Based on Springenberg and Dosovitskiy et al. https://arxiv.org/abs/1412.6806, |
| compute gradient-based saliency maps by backpropagating positive gradients and inputs (see ``_AutoGradReLU``). |
| |
| See also: |
| |
| - Springenberg and Dosovitskiy et al. Striving for Simplicity: The All Convolutional Net |
| (https://arxiv.org/abs/1412.6806) |
| """ |
|
|
| def __call__(self, x: torch.Tensor, index: torch.Tensor | int | None = None, **kwargs: Any) -> torch.Tensor: |
| with replace_modules_temp(self.model, "relu", _GradReLU(), strict_match=False): |
| return super().__call__(x, index, **kwargs) |
|
|
|
|
| class GuidedBackpropSmoothGrad(SmoothGrad): |
| """ |
| Compute gradient-based saliency maps based on both ``GuidedBackpropGrad`` and ``SmoothGrad``. |
| """ |
|
|
| def __call__(self, x: torch.Tensor, index: torch.Tensor | int | None = None, **kwargs: Any) -> torch.Tensor: |
| with replace_modules_temp(self.model, "relu", _GradReLU(), strict_match=False): |
| return super().__call__(x, index, **kwargs) |
|
|