| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """ |
| A collection of dictionary-based wrappers around the "vanilla" transforms for utility functions |
| defined in :py:class:`monai.transforms.utility.array`. |
| |
| Class names are ended with 'd' to denote dictionary-based transforms. |
| """ |
|
|
| from __future__ import annotations |
|
|
| import re |
| from collections.abc import Callable, Hashable, Mapping |
| from copy import deepcopy |
| from typing import Any, Sequence, cast |
|
|
| import numpy as np |
| import torch |
|
|
| from monai.config import DtypeLike, KeysCollection |
| from monai.config.type_definitions import NdarrayOrTensor |
| from monai.data.meta_tensor import MetaObj, MetaTensor |
| from monai.data.utils import no_collation |
| from monai.transforms.inverse import InvertibleTransform |
| from monai.transforms.traits import MultiSampleTrait, RandomizableTrait |
| from monai.transforms.transform import MapTransform, Randomizable, RandomizableTransform |
| from monai.transforms.utility.array import ( |
| AddCoordinateChannels, |
| AddExtremePointsChannel, |
| AsChannelLast, |
| CastToType, |
| ClassesToIndices, |
| ConvertToMultiChannelBasedOnBratsClasses, |
| CuCIM, |
| DataStats, |
| EnsureChannelFirst, |
| EnsureType, |
| FgBgToIndices, |
| Identity, |
| ImageFilter, |
| IntensityStats, |
| LabelToMask, |
| Lambda, |
| MapLabelValue, |
| RemoveRepeatedChannel, |
| RepeatChannel, |
| SimulateDelay, |
| SplitDim, |
| SqueezeDim, |
| ToCupy, |
| ToDevice, |
| ToNumpy, |
| ToPIL, |
| TorchVision, |
| ToTensor, |
| Transpose, |
| ) |
| from monai.transforms.utils import extreme_points_to_image, get_extreme_points |
| from monai.transforms.utils_pytorch_numpy_unification import concatenate |
| from monai.utils import ensure_tuple, ensure_tuple_rep |
| from monai.utils.enums import PostFix, TraceKeys, TransformBackends |
| from monai.utils.type_conversion import convert_to_dst_type |
|
|
| __all__ = [ |
| "AddCoordinateChannelsD", |
| "AddCoordinateChannelsDict", |
| "AddCoordinateChannelsd", |
| "AddExtremePointsChannelD", |
| "AddExtremePointsChannelDict", |
| "AddExtremePointsChanneld", |
| "AsChannelLastD", |
| "AsChannelLastDict", |
| "AsChannelLastd", |
| "CastToTypeD", |
| "CastToTypeDict", |
| "CastToTyped", |
| "ConcatItemsD", |
| "ConcatItemsDict", |
| "ConcatItemsd", |
| "ConvertToMultiChannelBasedOnBratsClassesD", |
| "ConvertToMultiChannelBasedOnBratsClassesDict", |
| "ConvertToMultiChannelBasedOnBratsClassesd", |
| "CopyItemsD", |
| "CopyItemsDict", |
| "CopyItemsd", |
| "CuCIMd", |
| "CuCIMD", |
| "CuCIMDict", |
| "DataStatsD", |
| "DataStatsDict", |
| "DataStatsd", |
| "DeleteItemsD", |
| "DeleteItemsDict", |
| "DeleteItemsd", |
| "EnsureChannelFirstD", |
| "EnsureChannelFirstDict", |
| "EnsureChannelFirstd", |
| "EnsureTypeD", |
| "EnsureTypeDict", |
| "EnsureTyped", |
| "FgBgToIndicesD", |
| "FgBgToIndicesDict", |
| "FgBgToIndicesd", |
| "IdentityD", |
| "IdentityDict", |
| "Identityd", |
| "IntensityStatsd", |
| "IntensityStatsD", |
| "IntensityStatsDict", |
| "ImageFilterd", |
| "LabelToMaskD", |
| "LabelToMaskDict", |
| "LabelToMaskd", |
| "LambdaD", |
| "LambdaDict", |
| "Lambdad", |
| "MapLabelValueD", |
| "MapLabelValueDict", |
| "MapLabelValued", |
| "FlattenSubKeysd", |
| "FlattenSubKeysD", |
| "FlattenSubKeysDict", |
| "RandCuCIMd", |
| "RandCuCIMD", |
| "RandCuCIMDict", |
| "RandImageFilterd", |
| "RandLambdaD", |
| "RandLambdaDict", |
| "RandLambdad", |
| "RandTorchVisionD", |
| "RandTorchVisionDict", |
| "RandTorchVisiond", |
| "RemoveRepeatedChannelD", |
| "RemoveRepeatedChannelDict", |
| "RemoveRepeatedChanneld", |
| "RepeatChannelD", |
| "RepeatChannelDict", |
| "RepeatChanneld", |
| "SelectItemsD", |
| "SelectItemsDict", |
| "SelectItemsd", |
| "SimulateDelayD", |
| "SimulateDelayDict", |
| "SimulateDelayd", |
| "SplitDimD", |
| "SplitDimDict", |
| "SplitDimd", |
| "SqueezeDimD", |
| "SqueezeDimDict", |
| "SqueezeDimd", |
| "ToCupyD", |
| "ToCupyDict", |
| "ToCupyd", |
| "ToDeviced", |
| "ToDeviceD", |
| "ToDeviceDict", |
| "ToNumpyD", |
| "ToNumpyDict", |
| "ToNumpyd", |
| "ToPILD", |
| "ToPILDict", |
| "ToPILd", |
| "ToTensorD", |
| "ToTensorDict", |
| "ToTensord", |
| "TorchVisionD", |
| "TorchVisionDict", |
| "TorchVisiond", |
| "Transposed", |
| "TransposeDict", |
| "TransposeD", |
| "ClassesToIndicesd", |
| "ClassesToIndicesD", |
| "ClassesToIndicesDict", |
| ] |
|
|
| DEFAULT_POST_FIX = PostFix.meta() |
|
|
|
|
| class Identityd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.Identity`. |
| """ |
|
|
| backend = Identity.backend |
|
|
| def __init__(self, keys: KeysCollection, allow_missing_keys: bool = False) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.identity = Identity() |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.identity(d[key]) |
| return d |
|
|
|
|
| class AsChannelLastd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.AsChannelLast`. |
| """ |
|
|
| backend = AsChannelLast.backend |
|
|
| def __init__(self, keys: KeysCollection, channel_dim: int = 0, allow_missing_keys: bool = False) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| channel_dim: which dimension of input image is the channel, default is the first dimension. |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.converter = AsChannelLast(channel_dim=channel_dim) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class EnsureChannelFirstd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.EnsureChannelFirst`. |
| """ |
|
|
| backend = EnsureChannelFirst.backend |
|
|
| def __init__( |
| self, keys: KeysCollection, strict_check: bool = True, allow_missing_keys: bool = False, channel_dim=None |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| strict_check: whether to raise an error when the meta information is insufficient. |
| allow_missing_keys: don't raise exception if key is missing. |
| channel_dim: This argument can be used to specify the original channel dimension (integer) of the input array. |
| It overrides the `original_channel_dim` from provided MetaTensor input. |
| If the input array doesn't have a channel dim, this value should be ``'no_channel'``. |
| If this is set to `None`, this class relies on `img` or `meta_dict` to provide the channel dimension. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.adjuster = EnsureChannelFirst(strict_check=strict_check, channel_dim=channel_dim) |
|
|
| def __call__(self, data: Mapping[Hashable, torch.Tensor]) -> dict[Hashable, torch.Tensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| meta_dict = d[key].meta if isinstance(d[key], MetaTensor) else None |
| d[key] = self.adjuster(d[key], meta_dict) |
| return d |
|
|
|
|
| class RepeatChanneld(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.RepeatChannel`. |
| """ |
|
|
| backend = RepeatChannel.backend |
|
|
| def __init__(self, keys: KeysCollection, repeats: int, allow_missing_keys: bool = False) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| repeats: the number of repetitions for each element. |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.repeater = RepeatChannel(repeats) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.repeater(d[key]) |
| return d |
|
|
|
|
| class RemoveRepeatedChanneld(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.RemoveRepeatedChannel`. |
| """ |
|
|
| backend = RemoveRepeatedChannel.backend |
|
|
| def __init__(self, keys: KeysCollection, repeats: int, allow_missing_keys: bool = False) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| repeats: the number of repetitions for each element. |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.repeater = RemoveRepeatedChannel(repeats) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.repeater(d[key]) |
| return d |
|
|
|
|
| class SplitDimd(MapTransform, MultiSampleTrait): |
| backend = SplitDim.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| output_postfixes: Sequence[str] | None = None, |
| dim: int = 0, |
| keepdim: bool = True, |
| update_meta: bool = True, |
| list_output: bool = False, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| output_postfixes: the postfixes to construct keys to store split data. |
| for example: if the key of input data is `pred` and split 2 classes, the output |
| data keys will be: pred_(output_postfixes[0]), pred_(output_postfixes[1]) |
| if None, using the index number: `pred_0`, `pred_1`, ... `pred_N`. |
| dim: which dimension of input image is the channel, default to 0. |
| keepdim: if `True`, output will have singleton in the split dimension. If `False`, this |
| dimension will be squeezed. |
| update_meta: if `True`, copy `[key]_meta_dict` for each output and update affine to |
| reflect the cropped image |
| list_output: it `True`, the output will be a list of dictionaries with the same keys as original. |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.output_postfixes = output_postfixes |
| self.splitter = SplitDim(dim, keepdim, update_meta) |
| self.list_output = list_output |
| if self.list_output is None and self.output_postfixes is not None: |
| raise ValueError("`output_postfixes` should not be provided when `list_output` is `True`.") |
|
|
| def __call__( |
| self, data: Mapping[Hashable, torch.Tensor] |
| ) -> dict[Hashable, torch.Tensor] | list[dict[Hashable, torch.Tensor]]: |
| d = dict(data) |
| all_keys = list(set(self.key_iterator(d))) |
|
|
| if self.list_output: |
| output = [] |
| results = [self.splitter(d[key]) for key in all_keys] |
| for row in zip(*results): |
| new_dict = dict(zip(all_keys, row)) |
| |
| for k in set(d.keys()).difference(set(all_keys)): |
| new_dict[k] = deepcopy(d[k]) |
| output.append(new_dict) |
| return output |
|
|
| for key in all_keys: |
| rets = self.splitter(d[key]) |
| postfixes: Sequence = list(range(len(rets))) if self.output_postfixes is None else self.output_postfixes |
| if len(postfixes) != len(rets): |
| raise ValueError(f"count of splits must match output_postfixes, {len(postfixes)} != {len(rets)}.") |
| for i, r in enumerate(rets): |
| split_key = f"{key}_{postfixes[i]}" |
| if split_key in d: |
| raise RuntimeError(f"input data already contains key {split_key}.") |
| d[split_key] = r |
| return d |
|
|
|
|
| class CastToTyped(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.CastToType`. |
| """ |
|
|
| backend = CastToType.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| dtype: Sequence[DtypeLike | torch.dtype] | DtypeLike | torch.dtype = np.float32, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| dtype: convert image to this data type, default is `np.float32`. |
| it also can be a sequence of dtypes or torch.dtype, |
| each element corresponds to a key in ``keys``. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
| MapTransform.__init__(self, keys, allow_missing_keys) |
| self.dtype = ensure_tuple_rep(dtype, len(self.keys)) |
| self.converter = CastToType() |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key, dtype in self.key_iterator(d, self.dtype): |
| d[key] = self.converter(d[key], dtype=dtype) |
|
|
| return d |
|
|
|
|
| class ToTensord(MapTransform, InvertibleTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.ToTensor`. |
| """ |
|
|
| backend = ToTensor.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| dtype: torch.dtype | None = None, |
| device: torch.device | str | None = None, |
| wrap_sequence: bool = True, |
| track_meta: bool | None = None, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| dtype: target data content type to convert, for example: torch.float, etc. |
| device: specify the target device to put the Tensor data. |
| wrap_sequence: if `False`, then lists will recursively call this function, default to `True`. |
| E.g., if `False`, `[1, 2]` -> `[tensor(1), tensor(2)]`, if `True`, then `[1, 2]` -> `tensor([1, 2])`. |
| track_meta: if `True` convert to ``MetaTensor``, otherwise to Pytorch ``Tensor``, |
| if ``None`` behave according to return value of py:func:`monai.data.meta_obj.get_track_meta`. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.converter = ToTensor(dtype=dtype, device=device, wrap_sequence=wrap_sequence, track_meta=track_meta) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| self.push_transform(d, key) |
| return d |
|
|
| def inverse(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| |
| self.pop_transform(d, key) |
| |
| inverse_transform = ToNumpy() |
| |
| d[key] = inverse_transform(d[key]) |
| return d |
|
|
|
|
| class EnsureTyped(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.EnsureType`. |
| |
| Ensure the input data to be a PyTorch Tensor or numpy array, support: `numpy array`, `PyTorch Tensor`, |
| `float`, `int`, `bool`, `string` and `object` keep the original. |
| If passing a dictionary, list or tuple, still return dictionary, list or tuple and recursively convert |
| every item to the expected data type if `wrap_sequence=False`. |
| |
| Note: Currently, we only convert tensor data to numpy array or scalar number in the inverse operation. |
| |
| """ |
|
|
| backend = EnsureType.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| data_type: str = "tensor", |
| dtype: Sequence[DtypeLike | torch.dtype] | DtypeLike | torch.dtype = None, |
| device: torch.device | None = None, |
| wrap_sequence: bool = True, |
| track_meta: bool | None = None, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| data_type: target data type to convert, should be "tensor" or "numpy". |
| dtype: target data content type to convert, for example: np.float32, torch.float, etc. |
| It also can be a sequence of dtype, each element corresponds to a key in ``keys``. |
| device: for Tensor data type, specify the target device. |
| wrap_sequence: if `False`, then lists will recursively call this function, default to `True`. |
| E.g., if `False`, `[1, 2]` -> `[tensor(1), tensor(2)]`, if `True`, then `[1, 2]` -> `tensor([1, 2])`. |
| track_meta: whether to convert to `MetaTensor` when `data_type` is "tensor". |
| If False, the output data type will be `torch.Tensor`. Default to the return value of `get_track_meta`. |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.dtype = ensure_tuple_rep(dtype, len(self.keys)) |
| self.converter = EnsureType( |
| data_type=data_type, device=device, wrap_sequence=wrap_sequence, track_meta=track_meta |
| ) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key, dtype in self.key_iterator(d, self.dtype): |
| d[key] = self.converter(d[key], dtype) |
| return d |
|
|
|
|
| class ToNumpyd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.ToNumpy`. |
| """ |
|
|
| backend = ToNumpy.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| dtype: DtypeLike = None, |
| wrap_sequence: bool = True, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| dtype: target data type when converting to numpy array. |
| wrap_sequence: if `False`, then lists will recursively call this function, default to `True`. |
| E.g., if `False`, `[1, 2]` -> `[array(1), array(2)]`, if `True`, then `[1, 2]` -> `array([1, 2])`. |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.converter = ToNumpy(dtype=dtype, wrap_sequence=wrap_sequence) |
|
|
| def __call__(self, data: Mapping[Hashable, Any]) -> dict[Hashable, Any]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class ToCupyd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.ToCupy`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| dtype: data type specifier. It is inferred from the input by default. |
| if not None, must be an argument of `numpy.dtype`, for more details: |
| https://docs.cupy.dev/en/stable/reference/generated/cupy.array.html. |
| wrap_sequence: if `False`, then lists will recursively call this function, default to `True`. |
| E.g., if `False`, `[1, 2]` -> `[array(1), array(2)]`, if `True`, then `[1, 2]` -> `array([1, 2])`. |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
|
|
| backend = ToCupy.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| dtype: np.dtype | None = None, |
| wrap_sequence: bool = True, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.converter = ToCupy(dtype=dtype, wrap_sequence=wrap_sequence) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class ToPILd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.ToNumpy`. |
| """ |
|
|
| backend = ToPIL.backend |
|
|
| def __init__(self, keys: KeysCollection, allow_missing_keys: bool = False) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.converter = ToPIL() |
|
|
| def __call__(self, data: Mapping[Hashable, Any]) -> dict[Hashable, Any]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class Transposed(MapTransform, InvertibleTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.Transpose`. |
| """ |
|
|
| backend = Transpose.backend |
|
|
| def __init__(self, keys: KeysCollection, indices: Sequence[int] | None, allow_missing_keys: bool = False) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.transform = Transpose(indices) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.transform(d[key]) |
| |
| indices = self.transform.indices or range(d[key].ndim)[::-1] |
| self.push_transform(d, key, extra_info={"indices": indices}) |
| return d |
|
|
| def inverse(self, data: Mapping[Hashable, Any]) -> dict[Hashable, Any]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| transform = self.get_most_recent_transform(d, key) |
| |
| fwd_indices = np.array(transform[TraceKeys.EXTRA_INFO]["indices"]) |
| inv_indices = np.argsort(fwd_indices) |
| inverse_transform = Transpose(inv_indices.tolist()) |
| |
| d[key] = inverse_transform(d[key]) |
| |
| self.pop_transform(d, key) |
| return d |
|
|
|
|
| class DeleteItemsd(MapTransform): |
| """ |
| Delete specified items from data dictionary to release memory. |
| It will remove the key-values and copy the others to construct a new dictionary. |
| """ |
|
|
| backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
| def __init__(self, keys: KeysCollection, sep: str = ".", use_re: Sequence[bool] | bool = False) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to delete, can be "A{sep}B{sep}C" |
| to delete key `C` in nested dictionary, `C` can be regular expression. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| sep: the separator tag to define nested dictionary keys, default to ".". |
| use_re: whether the specified key is a regular expression, it also can be |
| a list of bool values, mapping them to `keys`. |
| """ |
| super().__init__(keys) |
| self.sep = sep |
| self.use_re = ensure_tuple_rep(use_re, len(self.keys)) |
|
|
| def __call__(self, data): |
|
|
| def _delete_item(keys, d, use_re: bool = False): |
| key = keys[0] |
| if len(keys) > 1: |
| d[key] = _delete_item(keys[1:], d[key], use_re) |
| return d |
| return {k: v for k, v in d.items() if (use_re and not re.search(key, f"{k}")) or (not use_re and k != key)} |
|
|
| d = dict(data) |
| for key, use_re in zip(cast(Sequence[str], self.keys), self.use_re): |
| d = _delete_item(key.split(self.sep), d, use_re) |
|
|
| return d |
|
|
|
|
| class SelectItemsd(MapTransform): |
| """ |
| Select only specified items from data dictionary to release memory. |
| It will copy the selected key-values and construct a new dictionary. |
| """ |
|
|
| backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
| def __call__(self, data): |
| return {key: data[key] for key in self.key_iterator(data)} |
|
|
|
|
| class FlattenSubKeysd(MapTransform): |
| """ |
| If an item is dictionary, it flatten the item by moving the sub-items (defined by sub-keys) to the top level. |
| {"pred": {"a": ..., "b", ... }} --> {"a": ..., "b", ... } |
| |
| Args: |
| keys: keys of the corresponding items to be flatten |
| sub_keys: the sub-keys of items to be flatten. If not provided all the sub-keys are flattened. |
| delete_keys: whether to delete the key of the items that their sub-keys are flattened. Default to True. |
| prefix: optional prefix to be added to the sub-keys when moving to the top level. |
| By default no prefix will be added. |
| """ |
|
|
| backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| sub_keys: KeysCollection | None = None, |
| delete_keys: bool = True, |
| prefix: str | None = None, |
| ) -> None: |
| super().__init__(keys) |
| self.sub_keys = sub_keys |
| self.delete_keys = delete_keys |
| self.prefix = prefix |
|
|
| def __call__(self, data): |
| d = dict(data) |
| for key in self.key_iterator(d): |
| |
| sub_keys = d[key].keys() if self.sub_keys is None else self.sub_keys |
|
|
| |
| for sk in sub_keys: |
| |
| sk_top = f"{self.prefix}_{sk}" if self.prefix else sk |
| if sk_top in d: |
| raise ValueError( |
| f"'{sk_top}' already exists in the top-level keys. Please change `prefix` to avoid duplicity." |
| ) |
| d[sk_top] = d[key][sk] |
|
|
| |
| if self.delete_keys: |
| del d[key] |
| return d |
|
|
|
|
| class SqueezeDimd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.SqueezeDim`. |
| """ |
|
|
| backend = SqueezeDim.backend |
|
|
| def __init__( |
| self, keys: KeysCollection, dim: int = 0, update_meta: bool = True, allow_missing_keys: bool = False |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| dim: dimension to be squeezed. Default: 0 (the first dimension) |
| update_meta: whether to update the meta info if the input is a metatensor. Default is ``True``. |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.converter = SqueezeDim(dim=dim, update_meta=update_meta) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class DataStatsd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.DataStats`. |
| """ |
|
|
| backend = DataStats.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| prefix: Sequence[str] | str = "Data", |
| data_type: Sequence[bool] | bool = True, |
| data_shape: Sequence[bool] | bool = True, |
| value_range: Sequence[bool] | bool = True, |
| data_value: Sequence[bool] | bool = False, |
| additional_info: Sequence[Callable] | Callable | None = None, |
| name: str = "DataStats", |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| prefix: will be printed in format: "{prefix} statistics". |
| it also can be a sequence of string, each element corresponds to a key in ``keys``. |
| data_type: whether to show the type of input data. |
| it also can be a sequence of bool, each element corresponds to a key in ``keys``. |
| data_shape: whether to show the shape of input data. |
| it also can be a sequence of bool, each element corresponds to a key in ``keys``. |
| value_range: whether to show the value range of input data. |
| it also can be a sequence of bool, each element corresponds to a key in ``keys``. |
| data_value: whether to show the raw value of input data. |
| it also can be a sequence of bool, each element corresponds to a key in ``keys``. |
| a typical example is to print some properties of Nifti image: affine, pixdim, etc. |
| additional_info: user can define callable function to extract |
| additional info from input data. it also can be a sequence of string, each element |
| corresponds to a key in ``keys``. |
| name: identifier of `logging.logger` to use, defaulting to "DataStats". |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.prefix = ensure_tuple_rep(prefix, len(self.keys)) |
| self.data_type = ensure_tuple_rep(data_type, len(self.keys)) |
| self.data_shape = ensure_tuple_rep(data_shape, len(self.keys)) |
| self.value_range = ensure_tuple_rep(value_range, len(self.keys)) |
| self.data_value = ensure_tuple_rep(data_value, len(self.keys)) |
| self.additional_info = ensure_tuple_rep(additional_info, len(self.keys)) |
| self.printer = DataStats(name=name) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key, prefix, data_type, data_shape, value_range, data_value, additional_info in self.key_iterator( |
| d, self.prefix, self.data_type, self.data_shape, self.value_range, self.data_value, self.additional_info |
| ): |
| d[key] = self.printer(d[key], prefix, data_type, data_shape, value_range, data_value, additional_info) |
| return d |
|
|
|
|
| class SimulateDelayd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.SimulateDelay`. |
| """ |
|
|
| backend = SimulateDelay.backend |
|
|
| def __init__( |
| self, keys: KeysCollection, delay_time: Sequence[float] | float = 0.0, allow_missing_keys: bool = False |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| delay_time: The minimum amount of time, in fractions of seconds, to accomplish this identity task. |
| It also can be a sequence of string, each element corresponds to a key in ``keys``. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.delay_time = ensure_tuple_rep(delay_time, len(self.keys)) |
| self.delayer = SimulateDelay() |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key, delay_time in self.key_iterator(d, self.delay_time): |
| d[key] = self.delayer(d[key], delay_time=delay_time) |
| return d |
|
|
|
|
| class CopyItemsd(MapTransform): |
| """ |
| Copy specified items from data dictionary and save with different key names. |
| It can copy several items together and copy several times. |
| """ |
|
|
| backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| times: int = 1, |
| names: KeysCollection | None = None, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| times: expected copy times, for example, if keys is "img", times is 3, |
| it will add 3 copies of "img" data to the dictionary, default to 1. |
| names: the names corresponding to the newly copied data, |
| the length should match `len(keys) x times`. for example, if keys is ["img", "seg"] |
| and times is 2, names can be: ["img_1", "seg_1", "img_2", "seg_2"]. |
| if None, use "{key}_{index}" as key for copy times `N`, index from `0` to `N-1`. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| Raises: |
| ValueError: When ``times`` is nonpositive. |
| ValueError: When ``len(names)`` is not ``len(keys) * times``. Incompatible values. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| if times < 1: |
| raise ValueError(f"times must be positive, got {times}.") |
| self.times = times |
| names = [f"{k}_{i}" for k in self.keys for i in range(self.times)] if names is None else ensure_tuple(names) |
| if len(names) != (len(self.keys) * times): |
| raise ValueError( |
| "len(names) must match len(keys) * times, " |
| f"got len(names)={len(names)} len(keys) * times={len(self.keys) * times}." |
| ) |
| self.names = names |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| """ |
| Raises: |
| KeyError: When a key in ``self.names`` already exists in ``data``. |
| |
| """ |
| d = dict(data) |
| key_len = len(self.keys) |
| for i in range(self.times): |
| for key, new_key in self.key_iterator(d, self.names[i * key_len : (i + 1) * key_len]): |
| if new_key in d: |
| raise KeyError(f"Key {new_key} already exists in data.") |
| val = d[key] |
| d[new_key] = MetaObj.copy_items(val) if isinstance(val, (torch.Tensor, np.ndarray)) else deepcopy(val) |
| return d |
|
|
|
|
| class ConcatItemsd(MapTransform): |
| """ |
| Concatenate specified items from data dictionary together on the first dim to construct a big array. |
| Expect all the items are numpy array or PyTorch Tensor or MetaTensor. |
| Return the first input's meta information when items are MetaTensor. |
| """ |
|
|
| backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
| def __init__(self, keys: KeysCollection, name: str, dim: int = 0, allow_missing_keys: bool = False) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be concatenated together. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| name: the name corresponding to the key to store the concatenated data. |
| dim: on which dimension to concatenate the items, default is 0. |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.name = name |
| self.dim = dim |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| """ |
| Raises: |
| TypeError: When items in ``data`` differ in type. |
| TypeError: When the item type is not in ``Union[numpy.ndarray, torch.Tensor, MetaTensor]``. |
| |
| """ |
| d = dict(data) |
| output = [] |
| data_type = None |
| for key in self.key_iterator(d): |
| if data_type is None: |
| data_type = type(d[key]) |
| elif not isinstance(d[key], data_type): |
| raise TypeError("All items in data must have the same type.") |
| output.append(d[key]) |
|
|
| if len(output) == 0: |
| return d |
|
|
| if data_type is np.ndarray: |
| d[self.name] = np.concatenate(output, axis=self.dim) |
| elif issubclass(data_type, torch.Tensor): |
| d[self.name] = torch.cat(output, dim=self.dim) |
| else: |
| raise TypeError( |
| f"Unsupported data type: {data_type}, available options are (numpy.ndarray, torch.Tensor, MetaTensor)." |
| ) |
| return d |
|
|
|
|
| class Lambdad(MapTransform, InvertibleTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.Lambda`. |
| |
| For example: |
| |
| .. code-block:: python |
| :emphasize-lines: 2 |
| |
| input_data={'image': np.zeros((10, 2, 2)), 'label': np.ones((10, 2, 2))} |
| lambd = Lambdad(keys='label', func=lambda x: x[:4, :, :]) |
| print(lambd(input_data)['label'].shape) |
| (4, 2, 2) |
| |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| func: Lambda/function to be applied. It also can be a sequence of Callable, |
| each element corresponds to a key in ``keys``. |
| inv_func: Lambda/function of inverse operation if want to invert transforms, default to `lambda x: x`. |
| It also can be a sequence of Callable, each element corresponds to a key in ``keys``. |
| track_meta: If `False`, then standard data objects will be returned (e.g., torch.Tensor` and `np.ndarray`) |
| as opposed to MONAI's enhanced objects. By default, this is `True`. |
| overwrite: whether to overwrite the original data in the input dictionary with lambda function output. it |
| can be bool or str, when setting to str, it will create a new key for the output and keep the value of |
| key intact. default to True. it also can be a sequence of bool or str, each element corresponds to a key |
| in ``keys``. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| Note: The inverse operation doesn't allow to define `extra_info` or access other information, such as the |
| image's original size. If need these complicated information, please write a new InvertibleTransform directly. |
| |
| """ |
|
|
| backend = Lambda.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| func: Sequence[Callable] | Callable, |
| inv_func: Sequence[Callable] | Callable = no_collation, |
| track_meta: bool = True, |
| overwrite: Sequence[bool] | bool | Sequence[str] | str = True, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.func = ensure_tuple_rep(func, len(self.keys)) |
| self.inv_func = ensure_tuple_rep(inv_func, len(self.keys)) |
| self.overwrite = ensure_tuple_rep(overwrite, len(self.keys)) |
| self._lambd = Lambda(track_meta=track_meta) |
|
|
| def __call__(self, data: Mapping[Hashable, torch.Tensor]) -> dict[Hashable, torch.Tensor]: |
| d = dict(data) |
| for key, func, overwrite in self.key_iterator(d, self.func, self.overwrite): |
| ret = self._lambd(img=d[key], func=func) |
| if overwrite and isinstance(overwrite, bool): |
| d[key] = ret |
| elif isinstance(overwrite, str): |
| d[overwrite] = ret |
| return d |
|
|
| def inverse(self, data): |
| d = dict(data) |
| for key, overwrite in self.key_iterator(d, self.overwrite): |
| ret = self._lambd.inverse(data=d[key]) |
| if overwrite: |
| d[key] = ret |
| return d |
|
|
|
|
| class RandLambdad(Lambdad, RandomizableTransform): |
| """ |
| Randomizable version :py:class:`monai.transforms.Lambdad`, the input `func` may contain random logic, |
| or randomly execute the function based on `prob`. so `CacheDataset` will not execute it and cache the results. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| func: Lambda/function to be applied. It also can be a sequence of Callable, |
| each element corresponds to a key in ``keys``. |
| inv_func: Lambda/function of inverse operation if want to invert transforms, default to `lambda x: x`. |
| It also can be a sequence of Callable, each element corresponds to a key in ``keys``. |
| track_meta: If `False`, then standard data objects will be returned (e.g., torch.Tensor` and `np.ndarray`) |
| as opposed to MONAI's enhanced objects. By default, this is `True`. |
| overwrite: whether to overwrite the original data in the input dictionary with lambda function output. |
| default to True. it also can be a sequence of bool, each element corresponds to a key in ``keys``. |
| prob: probability of executing the random function, default to 1.0, with 100% probability to execute. |
| note that all the data specified by `keys` will share the same random probability to execute or not. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| For more details, please check :py:class:`monai.transforms.Lambdad`. |
| |
| Note: The inverse operation doesn't allow to define `extra_info` or access other information, such as the |
| image's original size. If need these complicated information, please write a new InvertibleTransform directly. |
| """ |
|
|
| backend = Lambda.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| func: Sequence[Callable] | Callable, |
| inv_func: Sequence[Callable] | Callable = no_collation, |
| track_meta: bool = True, |
| overwrite: Sequence[bool] | bool = True, |
| prob: float = 1.0, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| Lambdad.__init__( |
| self=self, |
| keys=keys, |
| func=func, |
| inv_func=inv_func, |
| track_meta=track_meta, |
| overwrite=overwrite, |
| allow_missing_keys=allow_missing_keys, |
| ) |
| RandomizableTransform.__init__(self=self, prob=prob, do_transform=True) |
|
|
| def __call__(self, data): |
| self.randomize(data) |
| d = dict(data) |
| for key, func, overwrite in self.key_iterator(d, self.func, self.overwrite): |
| ret = d[key] |
| if not isinstance(ret, MetaTensor): |
| ret = MetaTensor(ret) |
| if self._do_transform: |
| ret = self._lambd(ret, func=func) |
| self.push_transform(ret, extra_info={"lambda_info": self._lambd.pop_transform(ret)}) |
| else: |
| self.push_transform(ret) |
| if overwrite: |
| d[key] = ret |
| return d |
|
|
| def inverse(self, data: Mapping[Hashable, torch.Tensor]) -> dict[Hashable, torch.Tensor]: |
| d = dict(data) |
| for key, overwrite in self.key_iterator(d, self.overwrite): |
| if isinstance(d[key], MetaTensor): |
| tr = self.pop_transform(d[key]) |
| if tr[TraceKeys.DO_TRANSFORM]: |
| d[key].applied_operations.append(tr[TraceKeys.EXTRA_INFO]["lambda_info"]) |
| ret = self._lambd.inverse(d[key]) |
| if overwrite: |
| d[key] = ret |
| return d |
|
|
|
|
| class LabelToMaskd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.LabelToMask`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| select_labels: labels to generate mask from. for 1 channel label, the `select_labels` |
| is the expected label values, like: [1, 2, 3]. for One-Hot format label, the |
| `select_labels` is the expected channel indices. |
| merge_channels: whether to use `np.any()` to merge the result on channel dim. |
| if yes, will return a single channel mask with binary data. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
|
|
| backend = LabelToMask.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| select_labels: Sequence[int] | int, |
| merge_channels: bool = False, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.converter = LabelToMask(select_labels=select_labels, merge_channels=merge_channels) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
|
|
| return d |
|
|
|
|
| class FgBgToIndicesd(MapTransform, MultiSampleTrait): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.FgBgToIndices`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| fg_postfix: postfix to save the computed foreground indices in dict. |
| for example, if computed on `label` and `postfix = "_fg_indices"`, the key will be `label_fg_indices`. |
| bg_postfix: postfix to save the computed background indices in dict. |
| for example, if computed on `label` and `postfix = "_bg_indices"`, the key will be `label_bg_indices`. |
| image_key: if image_key is not None, use ``label == 0 & image > image_threshold`` to determine |
| the negative sample(background). so the output items will not map to all the voxels in the label. |
| image_threshold: if enabled image_key, use ``image > image_threshold`` to determine |
| the valid image content area and select background only in this area. |
| output_shape: expected shape of output indices. if not None, unravel indices to specified shape. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
|
|
| backend = FgBgToIndices.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| fg_postfix: str = "_fg_indices", |
| bg_postfix: str = "_bg_indices", |
| image_key: str | None = None, |
| image_threshold: float = 0.0, |
| output_shape: Sequence[int] | None = None, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.fg_postfix = fg_postfix |
| self.bg_postfix = bg_postfix |
| self.image_key = image_key |
| self.converter = FgBgToIndices(image_threshold, output_shape) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| image = d[self.image_key] if self.image_key else None |
| for key in self.key_iterator(d): |
| d[str(key) + self.fg_postfix], d[str(key) + self.bg_postfix] = self.converter(d[key], image) |
|
|
| return d |
|
|
|
|
| class ClassesToIndicesd(MapTransform, MultiSampleTrait): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.ClassesToIndices`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| indices_postfix: postfix to save the computed indices of all classes in dict. |
| for example, if computed on `label` and `postfix = "_cls_indices"`, the key will be `label_cls_indices`. |
| num_classes: number of classes for argmax label, not necessary for One-Hot label. |
| image_key: if image_key is not None, use ``image > image_threshold`` to define valid region, and only select |
| the indices within the valid region. |
| image_threshold: if enabled image_key, use ``image > image_threshold`` to determine the valid image content |
| area and select only the indices of classes in this area. |
| output_shape: expected shape of output indices. if not None, unravel indices to specified shape. |
| max_samples_per_class: maximum length of indices to sample in each class to reduce memory consumption. |
| Default is None, no subsampling. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
|
|
| backend = ClassesToIndices.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| indices_postfix: str = "_cls_indices", |
| num_classes: int | None = None, |
| image_key: str | None = None, |
| image_threshold: float = 0.0, |
| output_shape: Sequence[int] | None = None, |
| max_samples_per_class: int | None = None, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.indices_postfix = indices_postfix |
| self.image_key = image_key |
| self.converter = ClassesToIndices(num_classes, image_threshold, output_shape, max_samples_per_class) |
|
|
| def __call__(self, data: Mapping[Hashable, Any]): |
| d = dict(data) |
| image = d[self.image_key] if self.image_key else None |
| for key in self.key_iterator(d): |
| d[str(key) + self.indices_postfix] = self.converter(d[key], image) |
|
|
| return d |
|
|
|
|
| class ConvertToMultiChannelBasedOnBratsClassesd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.ConvertToMultiChannelBasedOnBratsClasses`. |
| Convert labels to multi channels based on brats18 classes: |
| label 1 is the necrotic and non-enhancing tumor core |
| label 2 is the peritumoral edema |
| label 4 is the GD-enhancing tumor |
| The possible classes are TC (Tumor core), WT (Whole tumor) |
| and ET (Enhancing tumor). |
| """ |
|
|
| backend = ConvertToMultiChannelBasedOnBratsClasses.backend |
|
|
| def __init__(self, keys: KeysCollection, allow_missing_keys: bool = False): |
| super().__init__(keys, allow_missing_keys) |
| self.converter = ConvertToMultiChannelBasedOnBratsClasses() |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class AddExtremePointsChanneld(Randomizable, MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.AddExtremePointsChannel`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| label_key: key to label source to get the extreme points. |
| background: Class index of background label, defaults to 0. |
| pert: Random perturbation amount to add to the points, defaults to 0.0. |
| sigma: if a list of values, must match the count of spatial dimensions of input data, |
| and apply every value in the list to 1 spatial dimension. if only 1 value provided, |
| use it for all spatial dimensions. |
| rescale_min: minimum value of output data. |
| rescale_max: maximum value of output data. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
|
|
| backend = AddExtremePointsChannel.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| label_key: str, |
| background: int = 0, |
| pert: float = 0.0, |
| sigma: Sequence[float] | float | Sequence[torch.Tensor] | torch.Tensor = 3.0, |
| rescale_min: float = -1.0, |
| rescale_max: float = 1.0, |
| allow_missing_keys: bool = False, |
| ): |
| MapTransform.__init__(self, keys, allow_missing_keys) |
| self.background = background |
| self.pert = pert |
| self.points: list[tuple[int, ...]] = [] |
| self.label_key = label_key |
| self.sigma = sigma |
| self.rescale_min = rescale_min |
| self.rescale_max = rescale_max |
|
|
| def randomize(self, label: NdarrayOrTensor) -> None: |
| self.points = get_extreme_points(label, rand_state=self.R, background=self.background, pert=self.pert) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| label = d[self.label_key] |
| if label.shape[0] != 1: |
| raise ValueError("Only supports single channel labels!") |
|
|
| |
| self.randomize(label[0, :]) |
|
|
| for key in self.key_iterator(d): |
| img = d[key] |
| points_image = extreme_points_to_image( |
| points=self.points, |
| label=label, |
| sigma=self.sigma, |
| rescale_min=self.rescale_min, |
| rescale_max=self.rescale_max, |
| ) |
| points_image, *_ = convert_to_dst_type(points_image, img) |
| d[key] = concatenate([img, points_image], axis=0) |
| return d |
|
|
|
|
| class TorchVisiond(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.TorchVision` for non-randomized transforms. |
| For randomized transforms of TorchVision use :py:class:`monai.transforms.RandTorchVisiond`. |
| |
| Note: |
| As most of the TorchVision transforms only work for PIL image and PyTorch Tensor, this transform expects input |
| data to be dict of PyTorch Tensors, users can easily call `ToTensord` transform to convert Numpy to Tensor. |
| """ |
|
|
| backend = TorchVision.backend |
|
|
| def __init__(self, keys: KeysCollection, name: str, allow_missing_keys: bool = False, *args, **kwargs) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| name: The transform name in TorchVision package. |
| allow_missing_keys: don't raise exception if key is missing. |
| args: parameters for the TorchVision transform. |
| kwargs: parameters for the TorchVision transform. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.name = name |
| self.trans = TorchVision(name, *args, **kwargs) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.trans(d[key]) |
| return d |
|
|
|
|
| class RandTorchVisiond(MapTransform, RandomizableTrait): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.TorchVision` for randomized transforms. |
| For deterministic non-randomized transforms of TorchVision use :py:class:`monai.transforms.TorchVisiond`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| name: The transform name in TorchVision package. |
| allow_missing_keys: don't raise exception if key is missing. |
| args: parameters for the TorchVision transform. |
| kwargs: parameters for the TorchVision transform. |
| |
| Note: |
| |
| - As most of the TorchVision transforms only work for PIL image and PyTorch Tensor, this transform expects input |
| data to be dict of PyTorch Tensors. Users should call `ToTensord` transform first to convert Numpy to Tensor. |
| - This class inherits the ``Randomizable`` purely to prevent any dataset caching to skip the transform |
| computation. If the random factor of the underlying torchvision transform is not derived from `self.R`, |
| the results may not be deterministic. See Also: :py:class:`monai.transforms.Randomizable`. |
| |
| """ |
|
|
| backend = TorchVision.backend |
|
|
| def __init__(self, keys: KeysCollection, name: str, allow_missing_keys: bool = False, *args, **kwargs) -> None: |
| MapTransform.__init__(self, keys, allow_missing_keys) |
| self.name = name |
| self.trans = TorchVision(name, *args, **kwargs) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.trans(d[key]) |
| return d |
|
|
|
|
| class MapLabelValued(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.MapLabelValue`. |
| """ |
|
|
| backend = MapLabelValue.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| orig_labels: Sequence, |
| target_labels: Sequence, |
| dtype: DtypeLike = np.float32, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| orig_labels: original labels that map to others. |
| target_labels: expected label values, 1: 1 map to the `orig_labels`. |
| dtype: convert the output data to dtype, default to float32. |
| if dtype is from PyTorch, the transform will use the pytorch backend, else with numpy backend. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.mapper = MapLabelValue(orig_labels=orig_labels, target_labels=target_labels, dtype=dtype) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.mapper(d[key]) |
| return d |
|
|
|
|
| class IntensityStatsd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.IntensityStats`. |
| Compute statistics for the intensity values of input image and store into the metadata dictionary. |
| For example: if `ops=[lambda x: np.mean(x), "max"]` and `key_prefix="orig"`, may generate below stats: |
| `{"orig_custom_0": 1.5, "orig_max": 3.0}`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| ops: expected operations to compute statistics for the intensity. |
| if a string, will map to the predefined operations, supported: ["mean", "median", "max", "min", "std"] |
| mapping to `np.nanmean`, `np.nanmedian`, `np.nanmax`, `np.nanmin`, `np.nanstd`. |
| if a callable function, will execute the function on input image. |
| key_prefix: the prefix to combine with `ops` name to generate the key to store the results in the |
| metadata dictionary. if some `ops` are callable functions, will use "{key_prefix}_custom_{index}" |
| as the key, where index counts from 0. |
| mask_keys: if not None, specify the mask array for the image to extract only the interested area to compute |
| statistics, mask must have the same shape as the image. |
| it should be a sequence of strings or None, map to the `keys`. |
| channel_wise: whether to compute statistics for every channel of input image separately. |
| if True, return a list of values for every operation, default to False. |
| meta_keys: explicitly indicate the key of the corresponding metadata dictionary. |
| used to store the computed statistics to the meta dict. |
| for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
| the metadata is a dictionary object which contains: filename, original_shape, etc. |
| it can be a sequence of string, map to the `keys`. |
| if None, will try to construct meta_keys by `key_{meta_key_postfix}`. |
| meta_key_postfix: if meta_keys is None, use `key_{postfix}` to fetch the metadata according |
| to the key data, default is `meta_dict`, the metadata is a dictionary object. |
| used to store the computed statistics to the meta dict. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
|
|
| backend = IntensityStats.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| ops: Sequence[str | Callable], |
| key_prefix: str, |
| mask_keys: KeysCollection | None = None, |
| channel_wise: bool = False, |
| meta_keys: KeysCollection | None = None, |
| meta_key_postfix: str = DEFAULT_POST_FIX, |
| allow_missing_keys: bool = False, |
| ) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.stats = IntensityStats(ops=ops, key_prefix=key_prefix, channel_wise=channel_wise) |
| self.mask_keys = ensure_tuple_rep(None, len(self.keys)) if mask_keys is None else ensure_tuple(mask_keys) |
| self.meta_keys = ensure_tuple_rep(None, len(self.keys)) if meta_keys is None else ensure_tuple(meta_keys) |
| if len(self.keys) != len(self.meta_keys): |
| raise ValueError("meta_keys should have the same length as keys.") |
| self.meta_key_postfix = ensure_tuple_rep(meta_key_postfix, len(self.keys)) |
|
|
| def __call__(self, data) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key, mask_key, meta_key, meta_key_postfix in self.key_iterator( |
| d, self.mask_keys, self.meta_keys, self.meta_key_postfix |
| ): |
| meta_key = meta_key or f"{key}_{meta_key_postfix}" |
| d[key], d[meta_key] = self.stats( |
| img=d[key], meta_data=d.get(meta_key), mask=d.get(mask_key) if mask_key is not None else None |
| ) |
| return d |
|
|
|
|
| class ToDeviced(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.ToDevice`. |
| """ |
|
|
| backend = ToDevice.backend |
|
|
| def __init__( |
| self, keys: KeysCollection, device: torch.device | str, allow_missing_keys: bool = False, **kwargs |
| ) -> None: |
| """ |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| device: target device to move the Tensor, for example: "cuda:1". |
| allow_missing_keys: don't raise exception if key is missing. |
| kwargs: other args for the PyTorch `Tensor.to()` API, for more details: |
| https://pytorch.org/docs/stable/generated/torch.Tensor.to.html. |
| """ |
| super().__init__(keys, allow_missing_keys) |
| self.converter = ToDevice(device=device, **kwargs) |
|
|
| def __call__(self, data: Mapping[Hashable, torch.Tensor]) -> dict[Hashable, torch.Tensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.converter(d[key]) |
| return d |
|
|
|
|
| class CuCIMd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.CuCIM` for non-randomized transforms. |
| For randomized transforms of CuCIM use :py:class:`monai.transforms.RandCuCIMd`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| name: The transform name in CuCIM package. |
| allow_missing_keys: don't raise exception if key is missing. |
| args: parameters for the CuCIM transform. |
| kwargs: parameters for the CuCIM transform. |
| |
| Note: |
| CuCIM transforms only work with CuPy arrays, this transform expects input data to be `cupy.ndarray`. |
| Users can call `ToCuPy` transform to convert a numpy array or torch tensor to cupy array. |
| """ |
|
|
| def __init__(self, keys: KeysCollection, name: str, allow_missing_keys: bool = False, *args, **kwargs) -> None: |
| super().__init__(keys=keys, allow_missing_keys=allow_missing_keys) |
| self.name = name |
| self.trans = CuCIM(name, *args, **kwargs) |
|
|
| def __call__(self, data): |
| """ |
| Args: |
| data: Dict[Hashable, `cupy.ndarray`] |
| |
| Returns: |
| Dict[Hashable, `cupy.ndarray`] |
| |
| """ |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.trans(d[key]) |
| return d |
|
|
|
|
| class RandCuCIMd(MapTransform, RandomizableTrait): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.CuCIM` for randomized transforms. |
| For deterministic non-randomized transforms of CuCIM use :py:class:`monai.transforms.CuCIMd`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| name: The transform name in CuCIM package. |
| allow_missing_keys: don't raise exception if key is missing. |
| args: parameters for the CuCIM transform. |
| kwargs: parameters for the CuCIM transform. |
| |
| Note: |
| - CuCIM transform only work with CuPy arrays, so this transform expects input data to be `cupy.ndarray`. |
| Users should call `ToCuPy` transform first to convert a numpy array or torch tensor to cupy array. |
| - This class inherits the ``Randomizable`` purely to prevent any dataset caching to skip the transform |
| computation. If the random factor of the underlying cuCIM transform is not derived from `self.R`, |
| the results may not be deterministic. See Also: :py:class:`monai.transforms.Randomizable`. |
| """ |
|
|
| def __init__(self, keys: KeysCollection, name: str, allow_missing_keys: bool = False, *args, **kwargs) -> None: |
| MapTransform.__init__(self, keys, allow_missing_keys) |
| self.name = name |
| self.trans = CuCIM(name, *args, **kwargs) |
|
|
| def __call__(self, data): |
| """ |
| Args: |
| data: Dict[Hashable, `cupy.ndarray`] |
| |
| Returns: |
| Dict[Hashable, `cupy.ndarray`] |
| |
| """ |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.trans(d[key]) |
| return d |
|
|
|
|
| class AddCoordinateChannelsd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.AddCoordinateChannels`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: :py:class:`monai.transforms.compose.MapTransform` |
| spatial_dims: the spatial dimensions that are to have their coordinates encoded in a channel and |
| appended to the input image. E.g., `(0, 1, 2)` represents `H, W, D` dims and append three channels |
| to the input image, encoding the coordinates of the input's three spatial dimensions. |
| allow_missing_keys: don't raise exception if key is missing. |
| |
| """ |
|
|
| backend = AddCoordinateChannels.backend |
|
|
| def __init__(self, keys: KeysCollection, spatial_dims: Sequence[int], allow_missing_keys: bool = False) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.add_coordinate_channels = AddCoordinateChannels(spatial_dims=spatial_dims) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.add_coordinate_channels(d[key]) |
| return d |
|
|
|
|
| class ImageFilterd(MapTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.ImageFilter`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: monai.transforms.MapTransform |
| kernel: |
| A string specifying the kernel or a custom kernel as `torch.Tenor` or `np.ndarray`. |
| Available options are: `mean`, `laplacian`, `elliptical`, `sobel_{w,h,d}`` |
| kernel_size: |
| A single integer value specifying the size of the quadratic or cubic kernel. |
| Computational complexity increases exponentially with kernel_size, which |
| should be considered when choosing the kernel size. |
| allow_missing_keys: |
| Don't raise exception if key is missing. |
| """ |
|
|
| backend = ImageFilter.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| kernel: str | NdarrayOrTensor, |
| kernel_size: int | None = None, |
| allow_missing_keys: bool = False, |
| **kwargs, |
| ) -> None: |
| super().__init__(keys, allow_missing_keys) |
| self.filter = ImageFilter(kernel, kernel_size, **kwargs) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| for key in self.key_iterator(d): |
| d[key] = self.filter(d[key]) |
| return d |
|
|
|
|
| class RandImageFilterd(MapTransform, RandomizableTransform): |
| """ |
| Dictionary-based wrapper of :py:class:`monai.transforms.RandomFilterKernel`. |
| |
| Args: |
| keys: keys of the corresponding items to be transformed. |
| See also: monai.transforms.MapTransform |
| kernel: |
| A string specifying the kernel or a custom kernel as `torch.Tenor` or `np.ndarray`. |
| Available options are: `mean`, `laplacian`, `elliptical`, `sobel_{w,h,d}`` |
| kernel_size: |
| A single integer value specifying the size of the quadratic or cubic kernel. |
| Computational complexity increases exponentially with kernel_size, which |
| should be considered when choosing the kernel size. |
| prob: |
| Probability the transform is applied to the data |
| allow_missing_keys: |
| Don't raise exception if key is missing. |
| """ |
|
|
| backend = ImageFilter.backend |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| kernel: str | NdarrayOrTensor, |
| kernel_size: int | None = None, |
| prob: float = 0.1, |
| allow_missing_keys: bool = False, |
| **kwargs, |
| ) -> None: |
| MapTransform.__init__(self, keys, allow_missing_keys) |
| RandomizableTransform.__init__(self, prob) |
| self.filter = ImageFilter(kernel, kernel_size, **kwargs) |
|
|
| def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
| d = dict(data) |
| self.randomize(None) |
| if self._do_transform: |
| for key in self.key_iterator(d): |
| d[key] = self.filter(d[key]) |
| return d |
|
|
|
|
| RandImageFilterD = RandImageFilterDict = RandImageFilterd |
| ImageFilterD = ImageFilterDict = ImageFilterd |
| IdentityD = IdentityDict = Identityd |
| AsChannelLastD = AsChannelLastDict = AsChannelLastd |
| EnsureChannelFirstD = EnsureChannelFirstDict = EnsureChannelFirstd |
| RemoveRepeatedChannelD = RemoveRepeatedChannelDict = RemoveRepeatedChanneld |
| RepeatChannelD = RepeatChannelDict = RepeatChanneld |
| SplitDimD = SplitDimDict = SplitDimd |
| CastToTypeD = CastToTypeDict = CastToTyped |
| ToTensorD = ToTensorDict = ToTensord |
| EnsureTypeD = EnsureTypeDict = EnsureTyped |
| ToNumpyD = ToNumpyDict = ToNumpyd |
| ToCupyD = ToCupyDict = ToCupyd |
| ToPILD = ToPILDict = ToPILd |
| TransposeD = TransposeDict = Transposed |
| DeleteItemsD = DeleteItemsDict = DeleteItemsd |
| SelectItemsD = SelectItemsDict = SelectItemsd |
| SqueezeDimD = SqueezeDimDict = SqueezeDimd |
| DataStatsD = DataStatsDict = DataStatsd |
| SimulateDelayD = SimulateDelayDict = SimulateDelayd |
| CopyItemsD = CopyItemsDict = CopyItemsd |
| ConcatItemsD = ConcatItemsDict = ConcatItemsd |
| LambdaD = LambdaDict = Lambdad |
| LabelToMaskD = LabelToMaskDict = LabelToMaskd |
| FgBgToIndicesD = FgBgToIndicesDict = FgBgToIndicesd |
| ClassesToIndicesD = ClassesToIndicesDict = ClassesToIndicesd |
| ConvertToMultiChannelBasedOnBratsClassesD = ConvertToMultiChannelBasedOnBratsClassesDict = ( |
| ConvertToMultiChannelBasedOnBratsClassesd |
| ) |
| AddExtremePointsChannelD = AddExtremePointsChannelDict = AddExtremePointsChanneld |
| TorchVisionD = TorchVisionDict = TorchVisiond |
| RandTorchVisionD = RandTorchVisionDict = RandTorchVisiond |
| RandLambdaD = RandLambdaDict = RandLambdad |
| MapLabelValueD = MapLabelValueDict = MapLabelValued |
| IntensityStatsD = IntensityStatsDict = IntensityStatsd |
| ToDeviceD = ToDeviceDict = ToDeviced |
| CuCIMD = CuCIMDict = CuCIMd |
| RandCuCIMD = RandCuCIMDict = RandCuCIMd |
| AddCoordinateChannelsD = AddCoordinateChannelsDict = AddCoordinateChannelsd |
| FlattenSubKeysD = FlattenSubKeysDict = FlattenSubKeysd |
|
|