| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """ |
| A collection of dictionary-based wrappers around the "vanilla" transforms for model output tensors |
| defined in :py:class:`monai.transforms.utility.array`. |
| |
| Class names are ended with 'd' to denote dictionary-based transforms. |
| """ |
|
|
| from __future__ import annotations |
|
|
| import warnings |
| from collections.abc import Callable, Hashable, Iterable, Mapping, Sequence |
| from copy import deepcopy |
| from typing import Any |
|
|
| import numpy as np |
| import torch |
|
|
| from monai import config |
| from monai.config.type_definitions import KeysCollection, NdarrayOrTensor, PathLike |
| from monai.data.csv_saver import CSVSaver |
| from monai.data.meta_tensor import MetaTensor |
| from monai.transforms.inverse import InvertibleTransform |
| from monai.transforms.post.array import ( |
| Activations, |
| AsDiscrete, |
| DistanceTransformEDT, |
| FillHoles, |
| KeepLargestConnectedComponent, |
| LabelFilter, |
| LabelToContour, |
| MeanEnsemble, |
| ProbNMS, |
| RemoveSmallObjects, |
| SobelGradients, |
| VoteEnsemble, |
| ) |
| from monai.transforms.transform import MapTransform |
| from monai.transforms.utility.array import ToTensor |
| from monai.transforms.utils import allow_missing_keys_mode, convert_applied_interp_mode |
| from monai.utils import PostFix, convert_to_tensor, ensure_tuple, ensure_tuple_rep |
|
|
| __all__ = [ |
| "ActivationsD", |
| "ActivationsDict", |
| "Activationsd", |
| "AsDiscreteD", |
| "AsDiscreteDict", |
| "AsDiscreted", |
| "Ensembled", |
| "EnsembleD", |
| "EnsembleDict", |
| "FillHolesD", |
| "FillHolesDict", |
| "FillHolesd", |
| "InvertD", |
| "InvertDict", |
| "Invertd", |
| "KeepLargestConnectedComponentD", |
| "KeepLargestConnectedComponentDict", |
| "KeepLargestConnectedComponentd", |
| "RemoveSmallObjectsD", |
| "RemoveSmallObjectsDict", |
| "RemoveSmallObjectsd", |
| "LabelFilterD", |
| "LabelFilterDict", |
| "LabelFilterd", |
| "LabelToContourD", |
| "LabelToContourDict", |
| "LabelToContourd", |
| "MeanEnsembleD", |
| "MeanEnsembleDict", |
| "MeanEnsembled", |
| "ProbNMSD", |
| "ProbNMSDict", |
| "ProbNMSd", |
| "SaveClassificationD", |
| "SaveClassificationDict", |
| "SaveClassificationd", |
| "SobelGradientsD", |
| "SobelGradientsDict", |
| "SobelGradientsd", |
| "VoteEnsembleD", |
| "VoteEnsembleDict", |
| "VoteEnsembled", |
| "DistanceTransformEDTd", |
| "DistanceTransformEDTD", |
| "DistanceTransformEDTDict", |
| ] |
|
|
| DEFAULT_POST_FIX = PostFix.meta() |
|
|
|
|
| class Activationsd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.AddActivations`. |
| Add activation layers to the input data specified by `keys`. |
| """ |
|
|
| backend = Activations.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| sigmoid: Sequence[bool] | bool = False, |
| softmax: Sequence[bool] | bool = False, |
| other: Sequence[Callable] | Callable | None = None, |
| allow_missing_keys: bool = False, |
| **kwargs, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to model output and label. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| sigmoid: whether to execute sigmoid function on model output before transform. |
| it also can be a sequence of bool, each element corresponds to a key in ``keys``. |
| softmax: whether to execute softmax function on model output before transform. |
| it also can be a sequence of bool, each element corresponds to a key in ``keys``. |
| other: callable function to execute other activation layers, |
| for example: `other = torch.tanh`. it also can be a sequence of Callable, each |
| element corresponds to a key in ``keys``. |
| allow_missing_keys: don't raise exception if key is missing. |
| kwargs: additional parameters to `torch.softmax` (used when ``softmax=True``). |
| Defaults to ``dim=0``, unrecognized parameters will be ignored. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.sigmoid = ensure_tuple_rep(sigmoid, len(self.keys)) |
| self.softmax = ensure_tuple_rep(softmax, len(self.keys)) |
| self.other = ensure_tuple_rep(other, len(self.keys)) |
| self.converter = Activations() |
| self.converter.kwargs = kwargs |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key, sigmoid, softmax, other in self.key_iterator(d, self.sigmoid, self.softmax, self.other): |
| d[key] = self.converter(d[key], sigmoid, softmax, other) |
| return d |
|
|
|
|
| class AsDiscreted(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.AsDiscrete`. |
| """ |
|
|
| backend = AsDiscrete.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| argmax: Sequence[bool] | bool = False, |
| to_onehot: Sequence[int | None] | int | None = None, |
| threshold: Sequence[float | None] | float | None = None, |
| rounding: Sequence[str | None] | str | None = None, |
| allow_missing_keys: bool = False, |
| **kwargs, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to model output and label. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| argmax: whether to execute argmax function on input data before transform. |
| it also can be a sequence of bool, each element corresponds to a key in ``keys``. |
| to_onehot: if not None, convert input data into the one-hot format with specified number of classes. |
| defaults to ``None``. it also can be a sequence, each element corresponds to a key in ``keys``. |
| threshold: if not None, threshold the float values to int number 0 or 1 with specified threshold value. |
| defaults to ``None``. it also can be a sequence, each element corresponds to a key in ``keys``. |
| rounding: if not None, round the data according to the specified option, |
| available options: ["torchrounding"]. it also can be a sequence of str or None, |
| each element corresponds to a key in ``keys``. |
| allow_missing_keys: don't raise exception if key is missing. |
| kwargs: additional parameters to ``AsDiscrete``. |
| ``dim``, ``keepdim``, ``dtype`` are supported, unrecognized parameters will be ignored. |
| These default to ``0``, ``True``, ``torch.float`` respectively. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.argmax = ensure_tuple_rep(argmax, len(self.keys)) |
| self.to_onehot = [] |
| for flag in ensure_tuple_rep(to_onehot, len(self.keys)): |
| if isinstance(flag, bool): |
| raise ValueError("`to_onehot=True/False` is deprecated, please use `to_onehot=num_classes` instead.") |
| self.to_onehot.append(flag) |
|
|
| self.threshold = [] |
| for flag in ensure_tuple_rep(threshold, len(self.keys)): |
| if isinstance(flag, bool): |
| raise ValueError("`threshold_values=True/False` is deprecated, please use `threshold=value` instead.") |
| self.threshold.append(flag) |
|
|
| self.rounding = ensure_tuple_rep(rounding, len(self.keys)) |
| self.converter = AsDiscrete() |
| self.converter.kwargs = kwargs |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key, argmax, to_onehot, threshold, rounding in self.key_iterator( |
| d, self.argmax, self.to_onehot, self.threshold, self.rounding |
| ): |
| d[key] = self.converter(d[key], argmax, to_onehot, threshold, rounding) |
| return d |
|
|
|
|
| class KeepLargestConnectedComponentd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.KeepLargestConnectedComponent`. |
| """ |
|
|
| backend = KeepLargestConnectedComponent.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| applied_labels: Sequence[int] | int | None = None, |
| is_onehot: bool | None = None, |
| independent: bool = True, |
| connectivity: int | None = None, |
| num_components: int = 1, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| applied_labels: Labels for applying the connected component analysis on. |
| If given, voxels whose value is in this list will be analyzed. |
| If `None`, all non-zero values will be analyzed. |
| is_onehot: if `True`, treat the input data as OneHot format data, otherwise, not OneHot format data. |
| default to None, which treats multi-channel data as OneHot and single channel data as not OneHot. |
| independent: whether to treat ``applied_labels`` as a union of foreground labels. |
| If ``True``, the connected component analysis will be performed on each foreground label independently |
| and return the intersection of the largest components. |
| If ``False``, the analysis will be performed on the union of foreground labels. |
| default is `True`. |
| connectivity: Maximum number of orthogonal hops to consider a pixel/voxel as a neighbor. |
| Accepted values are ranging from 1 to input.ndim. If ``None``, a full |
| connectivity of ``input.ndim`` is used. for more details: |
| https://scikit-image.org/docs/dev/api/skimage.measure.html#skimage.measure.label. |
| num_components: The number of largest components to preserve. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.converter = KeepLargestConnectedComponent( |
| applied_labels=applied_labels, |
| is_onehot=is_onehot, |
| independent=independent, |
| connectivity=connectivity, |
| num_components=num_components, |
| ) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class RemoveSmallObjectsd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.RemoveSmallObjectsd`. |
| |
| Args: |
| min_size: objects smaller than this size (in number of voxels; or surface area/volume value |
| in whatever units your image is if by_measure is True) are removed. |
| connectivity: Maximum number of orthogonal hops to consider a pixel/voxel as a neighbor. |
| Accepted values are ranging from 1 to input.ndim. If ``None``, a full |
| connectivity of ``input.ndim`` is used. For more details refer to linked scikit-image |
| documentation. |
| independent_channels: Whether or not to consider channels as independent. If true, then |
| conjoining islands from different labels will be removed if they are below the threshold. |
| If false, the overall size islands made from all non-background voxels will be used. |
| by_measure: Whether the specified min_size is in number of voxels. if this is True then min_size |
| represents a surface area or volume value of whatever units your image is in (mm^3, cm^2, etc.) |
| default is False. e.g. if min_size is 3, by_measure is True and the units of your data is mm, |
| objects smaller than 3mm^3 are removed. |
| pixdim: the pixdim of the input image. if a single number, this is used for all axes. |
| If a sequence of numbers, the length of the sequence must be equal to the image dimensions. |
| """ |
|
|
| backend = RemoveSmallObjects.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| min_size: int = 64, |
| connectivity: int = 1, |
| independent_channels: bool = True, |
| by_measure: bool = False, |
| pixdim: Sequence[float] | float | np.ndarray | None = None, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.converter = RemoveSmallObjects(min_size, connectivity, independent_channels, by_measure, pixdim) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class LabelFilterd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.LabelFilter`. |
| """ |
|
|
| backend = LabelFilter.backend |
|
|
| def __init__( |
| self, keys: KeysCollection, applied_labels: Sequence[int] | int, allow_missing_keys: bool = False |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| applied_labels: Label(s) to filter on. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.converter = LabelFilter(applied_labels) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class FillHolesd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.FillHoles`. |
| """ |
|
|
| backend = FillHoles.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| applied_labels: Iterable[int] | int | None = None, |
| connectivity: int | None = None, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Initialize the connectivity and limit the labels for which holes are filled. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| applied_labels (Optional[Union[Iterable[int], int]], optional): Labels for which to fill holes. Defaults to None, |
| that is filling holes for all labels. |
| connectivity (int, optional): Maximum number of orthogonal hops to consider a pixel/voxel as a neighbor. |
| Accepted values are ranging from 1 to input.ndim. Defaults to a full |
| connectivity of ``input.ndim``. |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.converter = FillHoles(applied_labels=applied_labels, connectivity=connectivity) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class LabelToContourd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.LabelToContour`. |
| """ |
|
|
| backend = LabelToContour.backend |
|
|
| def __init__(self, keys: KeysCollection, kernel_type: str = "Laplace", allow_missing_keys: bool = False) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| kernel_type: the method applied to do edge detection, default is "Laplace". |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.converter = LabelToContour(kernel_type=kernel_type) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class Ensembled(MapTransform): |
| """ |
| Base class of dictionary-based ensemble transforms. |
| |
| """ |
|
|
| backend = list(set(VoteEnsemble.backend) & set(MeanEnsemble.backend)) |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| ensemble: Callable[[Sequence[NdarrayOrTensor] | NdarrayOrTensor], NdarrayOrTensor], |
| output_key: str | None = None, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be stack and execute ensemble. |
| if only 1 key provided, suppose it's a PyTorch Tensor with data stacked on dimension `E`. |
| output_key: the key to store ensemble result in the dictionary. |
| ensemble: callable method to execute ensemble on specified data. |
| if only 1 key provided in `keys`, `output_key` can be None and use `keys` as default. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| Raises: |
| TypeError: When ``ensemble`` is not ``callable``. |
| ValueError: When ``len(keys) > 1`` and ``output_key=None``. Incompatible values. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| if not callable(ensemble): |
| raise TypeError(f"ensemble must be callable but is {type(ensemble).__name__}.") |
| self.ensemble = ensemble |
| if len(self.keys) > 1 and output_key is None: |
| raise ValueError("Incompatible values: len(self.keys) > 1 and output_key=None.") |
| self.output_key = output_key if output_key is not None else self.keys[0] |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| items: list[NdarrayOrTensor] | NdarrayOrTensor |
| if len(self.keys) == 1 and self.keys[0] in d: |
| items = d[self.keys[0]] |
| else: |
| items = [d[key] for key in self.key_iterator(d)] |
|
|
| if len(items) > 0: |
| d[self.output_key] = self.ensemble(items) |
|
|
| return d |
|
|
|
|
| class MeanEnsembled(Ensembled): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.MeanEnsemble`. |
| """ |
|
|
| backend = MeanEnsemble.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| output_key: str | None = None, |
| weights: Sequence[float] | NdarrayOrTensor | None = None, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be stack and execute ensemble. |
| if only 1 key provided, suppose it's a PyTorch Tensor with data stacked on dimension `E`. |
| output_key: the key to store ensemble result in the dictionary. |
| if only 1 key provided in `keys`, `output_key` can be None and use `keys` as default. |
| weights: can be a list or tuple of numbers for input data with shape: [E, C, H, W[, D]]. |
| or a Numpy ndarray or a PyTorch Tensor data. |
| the `weights` will be added to input data from highest dimension, for example: |
| 1. if the `weights` only has 1 dimension, it will be added to the `E` dimension of input data. |
| 2. if the `weights` has 2 dimensions, it will be added to `E` and `C` dimensions. |
| it's a typical practice to add weights for different classes: |
| to ensemble 3 segmentation model outputs, every output has 4 channels(classes), |
| so the input data shape can be: [3, 4, H, W, D]. |
| and add different `weights` for different classes, so the `weights` shape can be: [3, 4]. |
| for example: `weights = [[1, 2, 3, 4], [4, 3, 2, 1], [1, 1, 1, 1]]`. |
| |
| """ |
| ensemble = MeanEnsemble(weights=weights) |
| super().__init__(keys, ensemble, output_key) |
|
|
|
|
| class VoteEnsembled(Ensembled): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.VoteEnsemble`. |
| """ |
|
|
| backend = VoteEnsemble.backend |
|
|
| def __init__(self, keys: KeysCollection, output_key: str | None = None, num_classes: int | None = None) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be stack and execute ensemble. |
| if only 1 key provided, suppose it's a PyTorch Tensor with data stacked on dimension `E`. |
| output_key: the key to store ensemble result in the dictionary. |
| if only 1 key provided in `keys`, `output_key` can be None and use `keys` as default. |
| num_classes: if the input is single channel data instead of One-Hot, we can't get class number |
| from channel, need to explicitly specify the number of classes to vote. |
| |
| """ |
| ensemble = VoteEnsemble(num_classes=num_classes) |
| super().__init__(keys, ensemble, output_key) |
|
|
|
|
| class ProbNMSd(MapTransform): |
| """ |
| Performs probability based non-maximum suppression (NMS) on the probabilities map via |
| iteratively selecting the coordinate with highest probability and then move it as well |
| as its surrounding values. The remove range is determined by the parameter `box_size`. |
| If multiple coordinates have the same highest probability, only one of them will be |
| selected. |
| |
| Args: |
| spatial_dims: number of spatial dimensions of the input probabilities map. |
| Defaults to 2. |
| sigma: the standard deviation for gaussian filter. |
| It could be a single value, or `spatial_dims` number of values. Defaults to 0.0. |
| prob_threshold: the probability threshold, the function will stop searching if |
| the highest probability is no larger than the threshold. The value should be |
| no less than 0.0. Defaults to 0.5. |
| box_size: the box size (in pixel) to be removed around the pixel with the maximum probability. |
| It can be an integer that defines the size of a square or cube, |
| or a list containing different values for each dimensions. Defaults to 48. |
| |
| Return: |
| a list of selected lists, where inner lists contain probability and coordinates. |
| For example, for 3D input, the inner lists are in the form of [probability, x, y, z]. |
| |
| Raises: |
| ValueError: When ``prob_threshold`` is less than 0.0. |
| ValueError: When ``box_size`` is a list or tuple, and its length is not equal to `spatial_dims`. |
| ValueError: When ``box_size`` has a less than 1 value. |
| |
| """ |
|
|
| backend = ProbNMS.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| spatial_dims: int = 2, |
| sigma: Sequence[float] | float | Sequence[torch.Tensor] | torch.Tensor = 0.0, |
| prob_threshold: float = 0.5, |
| box_size: int | Sequence[int] = 48, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.prob_nms = ProbNMS( |
| spatial_dims=spatial_dims, sigma=sigma, prob_threshold=prob_threshold, box_size=box_size |
| ) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]): |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.prob_nms(d[key]) |
| return d |
|
|
|
|
| class Invertd(MapTransform): |
| """ |
| Utility transform to invert the previously applied transforms. |
| |
| Taking the ``transform`` previously applied on ``orig_keys``, this ``Invertd`` will apply the inverse of it |
| to the data stored at ``keys``. |
| |
| ``Invertd``'s output will also include a copy of the metadata |
| dictionary (originally from ``orig_meta_keys`` or the metadata of ``orig_keys``), |
| with the relevant fields inverted and stored at ``meta_keys``. |
| |
| A typical usage is to apply the inverse of the preprocessing (``transform=preprocessings``) on |
| input ``orig_keys=image`` to the model predictions ``keys=pred``. |
| |
| A detailed usage example is available in the tutorial: |
| https://github.com/Project-MONAI/tutorials/blob/master/3d_segmentation/torch/unet_inference_dict.py |
| |
| Note: |
| |
| - The output of the inverted data and metadata will be stored at ``keys`` and ``meta_keys`` respectively. |
| - To correctly invert the transforms, the information of the previously applied transforms should be |
| available at ``{orig_keys}_transforms``, and the original metadata at ``orig_meta_keys``. |
| (``meta_key_postfix`` is an optional string to conveniently construct "meta_keys" and/or "orig_meta_keys".) |
| see also: :py:class:`monai.transforms.TraceableTransform`. |
| - The transform will not change the content in ``orig_keys`` and ``orig_meta_key``. |
| These keys are only used to represent the data status of ``key`` before inverting. |
| |
| """ |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| transform: InvertibleTransform, |
| orig_keys: KeysCollection | None = None, |
| meta_keys: KeysCollection | None = None, |
| orig_meta_keys: KeysCollection | None = None, |
| meta_key_postfix: str = DEFAULT_POST_FIX, |
| nearest_interp: bool | Sequence[bool] = True, |
| to_tensor: bool | Sequence[bool] = True, |
| device: str | torch.device | Sequence[str | torch.device] | None = None, |
| post_func: Callable | Sequence[Callable] | None = None, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: the key of expected data in the dict, the inverse of ``transforms`` will be applied on it in-place. |
| It also can be a list of keys, will apply the inverse transform respectively. |
| transform: the transform applied to ``orig_key``, its inverse will be applied on ``key``. |
| orig_keys: the key of the original input data in the dict. These keys default to `self.keys` if not set. |
| the transform trace information of ``transforms`` should be stored at ``{orig_keys}_transforms``. |
| It can also be a list of keys, each matches the ``keys``. |
| meta_keys: The key to output the inverted metadata dictionary. |
| The metadata is a dictionary optionally containing: filename, original_shape. |
| It can be a sequence of strings, maps to ``keys``. |
| If None, will try to create a metadata dict with the default key: `{key}_{meta_key_postfix}`. |
| orig_meta_keys: the key of the metadata of original input data. |
| The metadata is a dictionary optionally containing: filename, original_shape. |
| It can be a sequence of strings, maps to the `keys`. |
| If None, will try to create a metadata dict with the default key: `{orig_key}_{meta_key_postfix}`. |
| This metadata dict will also be included in the inverted dict, stored in `meta_keys`. |
| meta_key_postfix: if `orig_meta_keys` is None, use `{orig_key}_{meta_key_postfix}` to fetch the |
| metadata from dict, if `meta_keys` is None, use `{key}_{meta_key_postfix}`. Default: ``"meta_dict"``. |
| nearest_interp: whether to use `nearest` interpolation mode when inverting the spatial transforms, |
| default to `True`. If `False`, use the same interpolation mode as the original transform. |
| It also can be a list of bool, each matches to the `keys` data. |
| to_tensor: whether to convert the inverted data into PyTorch Tensor first, default to `True`. |
| It also can be a list of bool, each matches to the `keys` data. |
| device: if converted to Tensor, move the inverted results to target device before `post_func`, |
| default to None, it also can be a list of string or `torch.device`, each matches to the `keys` data. |
| post_func: post processing for the inverted data, should be a callable function. |
| It also can be a list of callable, each matches to the `keys` data. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| if not isinstance(transform, InvertibleTransform): |
| raise ValueError("transform is not invertible, can't invert transform for the data.") |
| self.transform = transform |
| self.orig_keys = ensure_tuple_rep(orig_keys, len(self.keys)) if orig_keys is not None else self.keys |
| self.meta_keys = ensure_tuple_rep(None, len(self.keys)) if meta_keys is None else ensure_tuple(meta_keys) |
| if len(self.keys) != len(self.meta_keys): |
| raise ValueError("meta_keys should have the same length as keys.") |
| self.orig_meta_keys = ensure_tuple_rep(orig_meta_keys, len(self.keys)) |
| self.meta_key_postfix = ensure_tuple_rep(meta_key_postfix, len(self.keys)) |
| self.nearest_interp = ensure_tuple_rep(nearest_interp, len(self.keys)) |
| self.to_tensor = ensure_tuple_rep(to_tensor, len(self.keys)) |
| self.device = ensure_tuple_rep(device, len(self.keys)) |
| self.post_func = ensure_tuple_rep(post_func, len(self.keys)) |
| self._totensor = ToTensor() |
|
|
| def __call__(self, data: Mapping[Hashable, Any]) -> dict[Hashable, Any]: |
| d = dict(data) |
| for ( |
| key, |
| orig_key, |
| meta_key, |
| orig_meta_key, |
| meta_key_postfix, |
| nearest_interp, |
| to_tensor, |
| device, |
| post_func, |
| ) in self.key_iterator( |
| d, |
| self.orig_keys, |
| self.meta_keys, |
| self.orig_meta_keys, |
| self.meta_key_postfix, |
| self.nearest_interp, |
| self.to_tensor, |
| self.device, |
| self.post_func, |
| ): |
| if isinstance(d[key], MetaTensor): |
| if orig_key not in d: |
| warnings.warn(f"transform info of `{orig_key}` is not available in MetaTensor {key}.") |
| continue |
| else: |
| transform_key = InvertibleTransform.trace_key(orig_key) |
| if transform_key not in d: |
| warnings.warn(f"transform info of `{orig_key}` is not available or no InvertibleTransform applied.") |
| continue |
|
|
| orig_meta_key = orig_meta_key or f"{orig_key}_{meta_key_postfix}" |
| if orig_key in d and isinstance(d[orig_key], MetaTensor): |
| transform_info = d[orig_key].applied_operations |
| meta_info = d[orig_key].meta |
| else: |
| transform_info = d[InvertibleTransform.trace_key(orig_key)] |
| meta_info = d.get(orig_meta_key, {}) |
| if nearest_interp: |
| transform_info = convert_applied_interp_mode( |
| trans_info=transform_info, mode="nearest", align_corners=None |
| ) |
|
|
| inputs = d[key] |
| if isinstance(inputs, torch.Tensor): |
| inputs = inputs.detach() |
|
|
| if not isinstance(inputs, MetaTensor): |
| inputs = convert_to_tensor(inputs, track_meta=True) |
| inputs.applied_operations = deepcopy(transform_info) |
| inputs.meta = deepcopy(meta_info) |
|
|
| |
| input_dict = {orig_key: inputs} |
| if config.USE_META_DICT: |
| input_dict[InvertibleTransform.trace_key(orig_key)] = transform_info |
| input_dict[PostFix.meta(orig_key)] = meta_info |
| with allow_missing_keys_mode(self.transform): |
| inverted = self.transform.inverse(input_dict) |
|
|
| |
| inverted_data = inverted[orig_key] |
| if to_tensor and not isinstance(inverted_data, MetaTensor): |
| inverted_data = self._totensor(inverted_data) |
| if isinstance(inverted_data, np.ndarray) and device is not None and torch.device(device).type != "cpu": |
| raise ValueError(f"Inverted data with type of 'numpy.ndarray' support device='cpu', got {device}.") |
| if isinstance(inverted_data, torch.Tensor): |
| inverted_data = inverted_data.to(device=device) |
| d[key] = post_func(inverted_data) if callable(post_func) else inverted_data |
| |
| if InvertibleTransform.trace_key(orig_key) in d: |
| d[InvertibleTransform.trace_key(orig_key)] = inverted_data.applied_operations |
| |
| if orig_meta_key in d: |
| meta_key = meta_key or f"{key}_{meta_key_postfix}" |
| d[meta_key] = inverted.get(orig_meta_key) |
| return d |
|
|
|
|
| class SaveClassificationd(MapTransform): |
| """ |
| Save the classification results and metadata into CSV file or other storage. |
| |
| """ |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| meta_keys: KeysCollection | None = None, |
| meta_key_postfix: str = DEFAULT_POST_FIX, |
| saver: CSVSaver | None = None, |
| output_dir: PathLike = "./", |
| filename: str = "predictions.csv", |
| delimiter: str = ",", |
| overwrite: bool = True, |
| flush: bool = True, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to model output, this transform only supports 1 key. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| meta_keys: explicitly indicate the key of the corresponding metadata dictionary. |
| for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
| the metadata is a dictionary object which contains: filename, original_shape, etc. |
| it can be a sequence of string, map to the `keys`. |
| if None, will try to construct meta_keys by `key_{meta_key_postfix}`. |
| will extract the filename of input image to save classification results. |
| meta_key_postfix: `key_{postfix}` was used to store the metadata in `LoadImaged`. |
| so need the key to extract the metadata of input image, like filename, etc. default is `meta_dict`. |
| for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
| the metadata is a dictionary object which contains: filename, original_shape, etc. |
| this arg only works when `meta_keys=None`. if no corresponding metadata, set to `None`. |
| saver: the saver instance to save classification results, if None, create a CSVSaver internally. |
| the saver must provide `save(data, meta_data)` and `finalize()` APIs. |
| output_dir: if `saver=None`, specify the directory to save the CSV file. |
| filename: if `saver=None`, specify the name of the saved CSV file. |
| delimiter: the delimiter character in the saved file, default to "," as the default output type is `csv`. |
| to be consistent with: https://docs.python.org/3/library/csv.html#csv.Dialect.delimiter. |
| overwrite: if `saver=None`, indicate whether to overwriting existing CSV file content, if True, |
| will clear the file before saving. otherwise, will append new content to the CSV file. |
| flush: if `saver=None`, indicate whether to write the cache data to CSV file immediately |
| in this transform and clear the cache. default to True. |
| If False, may need user to call `saver.finalize()` manually or use `ClassificationSaver` handler. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| if len(self.keys) != 1: |
| raise ValueError("only 1 key is allowed when saving the classification result.") |
| self.saver = saver or CSVSaver( |
| output_dir=output_dir, filename=filename, overwrite=overwrite, flush=flush, delimiter=delimiter |
| ) |
| self.flush = flush |
| self.meta_keys = ensure_tuple_rep(meta_keys, len(self.keys)) |
| self.meta_key_postfix = ensure_tuple_rep(meta_key_postfix, len(self.keys)) |
|
|
| def __call__(self, data): |
| d = dict(data) |
| for key, meta_key, meta_key_postfix in self.key_iterator(d, self.meta_keys, self.meta_key_postfix): |
| if meta_key is None and meta_key_postfix is not None: |
| meta_key = f"{key}_{meta_key_postfix}" |
| meta_data = d[meta_key] if meta_key is not None else None |
| self.saver.save(data=d[key], meta_data=meta_data) |
| if self.flush: |
| self.saver.finalize() |
|
|
| return d |
|
|
| def get_saver(self): |
| """ |
| If want to write content into file, may need to call `finalize` of saver when epoch completed. |
| Or users can also get the cache content from `saver` instead of writing into file. |
| |
| """ |
| return self.saver |
|
|
|
|
| class SobelGradientsd(MapTransform): |
| """Calculate Sobel horizontal and vertical gradients of a grayscale image. |
| |
| Args: |
| keys: keys of the corresponding items to model output. |
| kernel_size: the size of the Sobel kernel. Defaults to 3. |
| spatial_axes: the axes that define the direction of the gradient to be calculated. It calculate the gradient |
| along each of the provide axis. By default it calculate the gradient for all spatial axes. |
| normalize_kernels: if normalize the Sobel kernel to provide proper gradients. Defaults to True. |
| normalize_gradients: if normalize the output gradient to 0 and 1. Defaults to False. |
| padding_mode: the padding mode of the image when convolving with Sobel kernels. Defaults to `"reflect"`. |
| Acceptable values are ``'zeros'``, ``'reflect'``, ``'replicate'`` or ``'circular'``. |
| See ``torch.nn.Conv1d()`` for more information. |
| dtype: kernel data type (torch.dtype). Defaults to `torch.float32`. |
| new_key_prefix: this prefix be prepended to the key to create a new key for the output and keep the value of |
| key intact. By default not prefix is set and the corresponding array to the key will be replaced. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
|
|
| backend = SobelGradients.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| kernel_size: int = 3, |
| spatial_axes: Sequence[int] | int | None = None, |
| normalize_kernels: bool = True, |
| normalize_gradients: bool = False, |
| padding_mode: str = "reflect", |
| dtype: torch.dtype = torch.float32, |
| new_key_prefix: str | None = None, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.transform = SobelGradients( |
| kernel_size=kernel_size, |
| spatial_axes=spatial_axes, |
| normalize_kernels=normalize_kernels, |
| normalize_gradients=normalize_gradients, |
| padding_mode=padding_mode, |
| dtype=dtype, |
| ) |
| self.new_key_prefix = new_key_prefix |
| self.kernel_diff = self.transform.kernel_diff |
| self.kernel_smooth = self.transform.kernel_smooth |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| new_key = key if self.new_key_prefix is None else self.new_key_prefix + key |
| d[new_key] = self.transform(d[key]) |
|
|
| return d |
|
|
|
|
| class DistanceTransformEDTd(MapTransform): |
| """ |
| Applies the Euclidean distance transform on the input. |
| Either GPU based with CuPy / cuCIM or CPU based with scipy. |
| To use the GPU implementation, make sure cuCIM is available and that the data is a `torch.tensor` on a GPU device. |
| |
| Note that the results of the libraries can differ, so stick to one if possible. |
| For details, check out the `SciPy`_ and `cuCIM`_ documentation and / or :func:`monai.transforms.utils.distance_transform_edt`. |
| |
| |
| Note on the input shape: |
| Has to be a channel first array, must have shape: (num_channels, H, W [,D]). |
| Can be of any type but will be converted into binary: 1 wherever image equates to True, 0 elsewhere. |
| Input gets passed channel-wise to the distance-transform, thus results from this function will differ |
| from directly calling ``distance_transform_edt()`` in CuPy or SciPy. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| allow_missing_keys: don't raise exception if key is missing. |
| sampling: Spacing of elements along each dimension. If a sequence, must be of length equal to the input rank -1; |
| if a single number, this is used for all axes. If not specified, a grid spacing of unity is implied. |
| |
| .. _SciPy: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.distance_transform_edt.html |
| .. _cuCIM: https://docs.rapids.ai/api/cucim/nightly/api/#cucim.core.operations.morphology.distance_transform_edt |
| |
| |
| """ |
|
|
| backend = DistanceTransformEDT.backend |
|
|
| def __init__( |
| self, keys: KeysCollection, allow_missing_keys: bool = False, sampling: None | float | list[float] = None |
| ) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.sampling = sampling |
| self.distance_transform = DistanceTransformEDT(sampling=self.sampling) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> Mapping[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.distance_transform(img=d[key]) |
|
|
| return d |
|
|
|
|
| ActivationsD = ActivationsDict = Activationsd |
| AsDiscreteD = AsDiscreteDict = AsDiscreted |
| FillHolesD = FillHolesDict = FillHolesd |
| InvertD = InvertDict = Invertd |
| KeepLargestConnectedComponentD = KeepLargestConnectedComponentDict = KeepLargestConnectedComponentd |
| RemoveSmallObjectsD = RemoveSmallObjectsDict = RemoveSmallObjectsd |
| LabelFilterD = LabelFilterDict = LabelFilterd |
| LabelToContourD = LabelToContourDict = LabelToContourd |
| MeanEnsembleD = MeanEnsembleDict = MeanEnsembled |
| ProbNMSD = ProbNMSDict = ProbNMSd |
| SaveClassificationD = SaveClassificationDict = SaveClassificationd |
| VoteEnsembleD = VoteEnsembleDict = VoteEnsembled |
| EnsembleD = EnsembleDict = Ensembled |
| SobelGradientsD = SobelGradientsDict = SobelGradientsd |
| DistanceTransformEDTD = DistanceTransformEDTDict = DistanceTransformEDTd |
|
|