| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """ |
| A collection of "vanilla" transforms for spatial operations |
| https://github.com/Project-MONAI/MONAI/wiki/MONAI_Design |
| """ |
|
|
| import warnings |
| from typing import Any, Callable, List, Optional, Sequence, Tuple, Union |
|
|
| import numpy as np |
| import torch |
|
|
| from monai.config import get_torch_version_tuple |
| from monai.data.utils import compute_shape_offset, to_affine_nd, zoom_affine |
| from monai.networks.layers import AffineTransform, GaussianFilter |
| from monai.transforms.compose import Randomizable, Transform |
| from monai.transforms.croppad.array import CenterSpatialCrop |
| from monai.transforms.utils import ( |
| create_control_grid, |
| create_grid, |
| create_rotate, |
| create_scale, |
| create_shear, |
| create_translate, |
| ) |
| from monai.utils import ( |
| GridSampleMode, |
| GridSamplePadMode, |
| InterpolateMode, |
| NumpyPadMode, |
| ensure_tuple, |
| ensure_tuple_rep, |
| ensure_tuple_size, |
| fall_back_tuple, |
| optional_import, |
| ) |
|
|
| nib, _ = optional_import("nibabel") |
|
|
| _torch_interp: Callable[..., torch.Tensor] |
|
|
| if get_torch_version_tuple() >= (1, 5): |
| |
| def _torch_interp(**kwargs): |
| return torch.nn.functional.interpolate(recompute_scale_factor=True, **kwargs) |
|
|
|
|
| else: |
| _torch_interp = torch.nn.functional.interpolate |
|
|
|
|
| class Spacing(Transform): |
| """ |
| Resample input image into the specified `pixdim`. |
| """ |
|
|
| def __init__( |
| self, |
| pixdim: Union[Sequence[float], float], |
| diagonal: bool = False, |
| mode: Union[GridSampleMode, str] = GridSampleMode.BILINEAR, |
| padding_mode: Union[GridSamplePadMode, str] = GridSamplePadMode.BORDER, |
| align_corners: bool = False, |
| dtype: Optional[np.dtype] = np.float64, |
| ) -> None: |
| """ |
| Args: |
| pixdim: output voxel spacing. |
| diagonal: whether to resample the input to have a diagonal affine matrix. |
| If True, the input data is resampled to the following affine:: |
| |
| np.diag((pixdim_0, pixdim_1, ..., pixdim_n, 1)) |
| |
| This effectively resets the volume to the world coordinate system (RAS+ in nibabel). |
| The original orientation, rotation, shearing are not preserved. |
| |
| If False, this transform preserves the axes orientation, orthogonal rotation and |
| translation components from the original affine. This option will not flip/swap axes |
| of the original data. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``"border"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| align_corners: Geometrically, we consider the pixels of the input as squares rather than points. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| dtype: data type for resampling computation. Defaults to ``np.float64`` for best precision. |
| If None, use the data type of input data. To be compatible with other modules, |
| the output data type is always ``np.float32``. |
| """ |
| self.pixdim = np.array(ensure_tuple(pixdim), dtype=np.float64) |
| self.diagonal = diagonal |
| self.mode: GridSampleMode = GridSampleMode(mode) |
| self.padding_mode: GridSamplePadMode = GridSamplePadMode(padding_mode) |
| self.align_corners = align_corners |
| self.dtype = dtype |
|
|
| def __call__( |
| self, |
| data_array: np.ndarray, |
| affine: Optional[np.ndarray] = None, |
| mode: Optional[Union[GridSampleMode, str]] = None, |
| padding_mode: Optional[Union[GridSamplePadMode, str]] = None, |
| align_corners: Optional[bool] = None, |
| dtype: Optional[np.dtype] = None, |
| ) -> Tuple[np.ndarray, np.ndarray, np.ndarray]: |
| """ |
| Args: |
| data_array: in shape (num_channels, H[, W, ...]). |
| affine (matrix): (N+1)x(N+1) original affine matrix for spatially ND `data_array`. Defaults to identity. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| align_corners: Geometrically, we consider the pixels of the input as squares rather than points. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| dtype: data type for resampling computation. Defaults to ``self.dtype``. |
| If None, use the data type of input data. To be compatible with other modules, |
| the output data type is always ``np.float32``. |
| |
| Raises: |
| ValueError: When ``data_array`` has no spatial dimensions. |
| ValueError: When ``pixdim`` is nonpositive. |
| |
| Returns: |
| data_array (resampled into `self.pixdim`), original pixdim, current pixdim. |
| |
| """ |
| _dtype = dtype or self.dtype or data_array.dtype |
| sr = data_array.ndim - 1 |
| if sr <= 0: |
| raise ValueError("data_array must have at least one spatial dimension.") |
| if affine is None: |
| |
| affine = np.eye(sr + 1, dtype=np.float64) |
| affine_ = np.eye(sr + 1, dtype=np.float64) |
| else: |
| affine_ = to_affine_nd(sr, affine) |
| out_d = self.pixdim[:sr] |
| if out_d.size < sr: |
| out_d = np.append(out_d, [1.0] * (out_d.size - sr)) |
| if np.any(out_d <= 0): |
| raise ValueError(f"pixdim must be positive, got {out_d}.") |
| |
| new_affine = zoom_affine(affine_, out_d, diagonal=self.diagonal) |
| output_shape, offset = compute_shape_offset(data_array.shape[1:], affine_, new_affine) |
| new_affine[:sr, -1] = offset[:sr] |
| transform = np.linalg.inv(affine_) @ new_affine |
| |
| transform = to_affine_nd(sr, transform) |
|
|
| |
| if np.allclose(transform, np.diag(np.ones(len(transform))), atol=1e-3): |
| output_data = data_array.copy().astype(np.float32) |
| new_affine = to_affine_nd(affine, new_affine) |
| return output_data, affine, new_affine |
|
|
| |
| affine_xform = AffineTransform( |
| normalized=False, |
| mode=mode or self.mode, |
| padding_mode=padding_mode or self.padding_mode, |
| align_corners=self.align_corners if align_corners is None else align_corners, |
| reverse_indexing=True, |
| ) |
| output_data = affine_xform( |
| |
| torch.as_tensor(np.ascontiguousarray(data_array).astype(_dtype)).unsqueeze(0), |
| torch.as_tensor(np.ascontiguousarray(transform).astype(_dtype)), |
| spatial_size=output_shape, |
| ) |
| output_data = output_data.squeeze(0).detach().cpu().numpy().astype(np.float32) |
| new_affine = to_affine_nd(affine, new_affine) |
| return output_data, affine, new_affine |
|
|
|
|
| class Orientation(Transform): |
| """ |
| Change the input image's orientation into the specified based on `axcodes`. |
| """ |
|
|
| def __init__( |
| self, |
| axcodes: Optional[str] = None, |
| as_closest_canonical: bool = False, |
| labels: Optional[Sequence[Tuple[str, str]]] = tuple(zip("LPI", "RAS")), |
| ) -> None: |
| """ |
| Args: |
| axcodes: N elements sequence for spatial ND input's orientation. |
| e.g. axcodes='RAS' represents 3D orientation: |
| (Left, Right), (Posterior, Anterior), (Inferior, Superior). |
| default orientation labels options are: 'L' and 'R' for the first dimension, |
| 'P' and 'A' for the second, 'I' and 'S' for the third. |
| as_closest_canonical: if True, load the image as closest to canonical axis format. |
| labels: optional, None or sequence of (2,) sequences |
| (2,) sequences are labels for (beginning, end) of output axis. |
| Defaults to ``(('L', 'R'), ('P', 'A'), ('I', 'S'))``. |
| |
| Raises: |
| ValueError: When ``axcodes=None`` and ``as_closest_canonical=True``. Incompatible values. |
| |
| See Also: `nibabel.orientations.ornt2axcodes`. |
| |
| """ |
| if axcodes is None and not as_closest_canonical: |
| raise ValueError("Incompatible values: axcodes=None and as_closest_canonical=True.") |
| if axcodes is not None and as_closest_canonical: |
| warnings.warn("using as_closest_canonical=True, axcodes ignored.") |
| self.axcodes = axcodes |
| self.as_closest_canonical = as_closest_canonical |
| self.labels = labels |
|
|
| def __call__( |
| self, data_array: np.ndarray, affine: Optional[np.ndarray] = None |
| ) -> Tuple[np.ndarray, np.ndarray, np.ndarray]: |
| """ |
| original orientation of `data_array` is defined by `affine`. |
| |
| Args: |
| data_array: in shape (num_channels, H[, W, ...]). |
| affine (matrix): (N+1)x(N+1) original affine matrix for spatially ND `data_array`. Defaults to identity. |
| |
| Raises: |
| ValueError: When ``data_array`` has no spatial dimensions. |
| ValueError: When ``axcodes`` spatiality differs from ``data_array``. |
| |
| Returns: |
| data_array (reoriented in `self.axcodes`), original axcodes, current axcodes. |
| |
| """ |
| sr = data_array.ndim - 1 |
| if sr <= 0: |
| raise ValueError("data_array must have at least one spatial dimension.") |
| if affine is None: |
| affine = np.eye(sr + 1, dtype=np.float64) |
| affine_ = np.eye(sr + 1, dtype=np.float64) |
| else: |
| affine_ = to_affine_nd(sr, affine) |
| src = nib.io_orientation(affine_) |
| if self.as_closest_canonical: |
| spatial_ornt = src |
| else: |
| assert self.axcodes is not None |
| dst = nib.orientations.axcodes2ornt(self.axcodes[:sr], labels=self.labels) |
| if len(dst) < sr: |
| raise ValueError( |
| f"axcodes must match data_array spatially, got axcodes={len(self.axcodes)}D data_array={sr}D" |
| ) |
| spatial_ornt = nib.orientations.ornt_transform(src, dst) |
| ornt = spatial_ornt.copy() |
| ornt[:, 0] += 1 |
| ornt = np.concatenate([np.array([[0, 1]]), ornt]) |
| shape = data_array.shape[1:] |
| data_array = nib.orientations.apply_orientation(data_array, ornt) |
| new_affine = affine_ @ nib.orientations.inv_ornt_aff(spatial_ornt, shape) |
| new_affine = to_affine_nd(affine, new_affine) |
| return data_array, affine, new_affine |
|
|
|
|
| class Flip(Transform): |
| """ |
| Reverses the order of elements along the given spatial axis. Preserves shape. |
| Uses ``np.flip`` in practice. See numpy.flip for additional details. |
| https://docs.scipy.org/doc/numpy/reference/generated/numpy.flip.html |
| |
| Args: |
| spatial_axis: spatial axes along which to flip over. Default is None. |
| """ |
|
|
| def __init__(self, spatial_axis: Optional[Union[Sequence[int], int]]) -> None: |
| self.spatial_axis = spatial_axis |
|
|
| def __call__(self, img: np.ndarray) -> np.ndarray: |
| """ |
| Args: |
| img: channel first array, must have shape: (num_channels, H[, W, ..., ]), |
| """ |
| flipped = list() |
| for channel in img: |
| flipped.append(np.flip(channel, self.spatial_axis)) |
| return np.stack(flipped).astype(img.dtype) |
|
|
|
|
| class Resize(Transform): |
| """ |
| Resize the input image to given spatial size. |
| Implemented using :py:class:`torch.nn.functional.interpolate`. |
| |
| Args: |
| spatial_size: expected shape of spatial dimensions after resize operation. |
| if the components of the `spatial_size` are non-positive values, the transform will use the |
| corresponding components of img size. For example, `spatial_size=(32, -1)` will be adapted |
| to `(32, 64)` if the second spatial dimension size of img is `64`. |
| mode: {``"nearest"``, ``"linear"``, ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| The interpolation mode. Defaults to ``"area"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| align_corners: This only has an effect when mode is |
| 'linear', 'bilinear', 'bicubic' or 'trilinear'. Default: None. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| """ |
|
|
| def __init__( |
| self, |
| spatial_size: Union[Sequence[int], int], |
| mode: Union[InterpolateMode, str] = InterpolateMode.AREA, |
| align_corners: Optional[bool] = None, |
| ) -> None: |
| self.spatial_size = ensure_tuple(spatial_size) |
| self.mode: InterpolateMode = InterpolateMode(mode) |
| self.align_corners = align_corners |
|
|
| def __call__( |
| self, |
| img: np.ndarray, |
| mode: Optional[Union[InterpolateMode, str]] = None, |
| align_corners: Optional[bool] = None, |
| ) -> np.ndarray: |
| """ |
| Args: |
| img: channel first array, must have shape: (num_channels, H[, W, ..., ]). |
| mode: {``"nearest"``, ``"linear"``, ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| The interpolation mode. Defaults to ``self.mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| align_corners: This only has an effect when mode is |
| 'linear', 'bilinear', 'bicubic' or 'trilinear'. Defaults to ``self.align_corners``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| |
| Raises: |
| ValueError: When ``self.spatial_size`` length is less than ``img`` spatial dimensions. |
| |
| """ |
| input_ndim = img.ndim - 1 |
| output_ndim = len(self.spatial_size) |
| if output_ndim > input_ndim: |
| input_shape = ensure_tuple_size(img.shape, output_ndim + 1, 1) |
| img = img.reshape(input_shape) |
| elif output_ndim < input_ndim: |
| raise ValueError( |
| "len(spatial_size) must be greater or equal to img spatial dimensions, " |
| f"got spatial_size={output_ndim} img={input_ndim}." |
| ) |
| spatial_size = fall_back_tuple(self.spatial_size, img.shape[1:]) |
| resized = _torch_interp( |
| input=torch.as_tensor(np.ascontiguousarray(img), dtype=torch.float).unsqueeze(0), |
| size=spatial_size, |
| mode=self.mode.value if mode is None else InterpolateMode(mode).value, |
| align_corners=self.align_corners if align_corners is None else align_corners, |
| ) |
| resized = resized.squeeze(0).detach().cpu().numpy() |
| return resized |
|
|
|
|
| class Rotate(Transform): |
| """ |
| Rotates an input image by given angle using :py:class:`monai.networks.layers.AffineTransform`. |
| |
| Args: |
| angle: Rotation angle(s) in degrees. should a float for 2D, three floats for 3D. |
| keep_size: If it is True, the output shape is kept the same as the input. |
| If it is False, the output shape is adapted so that the |
| input array is contained completely in the output. Default is True. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``"border"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| align_corners: Defaults to False. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| dtype: data type for resampling computation. Defaults to ``np.float64`` for best precision. |
| If None, use the data type of input data. To be compatible with other modules, |
| the output data type is always ``np.float32``. |
| """ |
|
|
| def __init__( |
| self, |
| angle: Union[Sequence[float], float], |
| keep_size: bool = True, |
| mode: Union[GridSampleMode, str] = GridSampleMode.BILINEAR, |
| padding_mode: Union[GridSamplePadMode, str] = GridSamplePadMode.BORDER, |
| align_corners: bool = False, |
| dtype: Optional[np.dtype] = np.float64, |
| ) -> None: |
| self.angle = angle |
| self.keep_size = keep_size |
| self.mode: GridSampleMode = GridSampleMode(mode) |
| self.padding_mode: GridSamplePadMode = GridSamplePadMode(padding_mode) |
| self.align_corners = align_corners |
| self.dtype = dtype |
|
|
| def __call__( |
| self, |
| img: np.ndarray, |
| mode: Optional[Union[GridSampleMode, str]] = None, |
| padding_mode: Optional[Union[GridSamplePadMode, str]] = None, |
| align_corners: Optional[bool] = None, |
| dtype: Optional[np.dtype] = None, |
| ) -> np.ndarray: |
| """ |
| Args: |
| img: channel first array, must have shape: [chns, H, W] or [chns, H, W, D]. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| align_corners: Defaults to ``self.align_corners``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| align_corners: Defaults to ``self.align_corners``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| dtype: data type for resampling computation. Defaults to ``self.dtype``. |
| If None, use the data type of input data. To be compatible with other modules, |
| the output data type is always ``np.float32``. |
| |
| Raises: |
| ValueError: When ``img`` spatially is not one of [2D, 3D]. |
| |
| """ |
| _dtype = dtype or self.dtype or img.dtype |
| im_shape = np.asarray(img.shape[1:]) |
| input_ndim = len(im_shape) |
| if input_ndim not in (2, 3): |
| raise ValueError(f"Unsupported img dimension: {input_ndim}, available options are [2, 3].") |
| _angle = ensure_tuple_rep(self.angle, 1 if input_ndim == 2 else 3) |
| _rad = np.deg2rad(_angle) |
| transform = create_rotate(input_ndim, _rad) |
| shift = create_translate(input_ndim, (im_shape - 1) / 2) |
| if self.keep_size: |
| output_shape = im_shape |
| else: |
| corners = np.asarray(np.meshgrid(*[(0, dim) for dim in im_shape], indexing="ij")).reshape( |
| (len(im_shape), -1) |
| ) |
| corners = transform[:-1, :-1] @ corners |
| output_shape = (corners.ptp(axis=1) + 0.5).astype(int) |
| shift_1 = create_translate(input_ndim, -(output_shape - 1) / 2) |
| transform = shift @ transform @ shift_1 |
|
|
| xform = AffineTransform( |
| normalized=False, |
| mode=mode or self.mode, |
| padding_mode=padding_mode or self.padding_mode, |
| align_corners=self.align_corners if align_corners is None else align_corners, |
| reverse_indexing=True, |
| ) |
| output = xform( |
| torch.as_tensor(np.ascontiguousarray(img).astype(_dtype)).unsqueeze(0), |
| torch.as_tensor(np.ascontiguousarray(transform).astype(_dtype)), |
| spatial_size=output_shape, |
| ) |
| output = output.squeeze(0).detach().cpu().numpy().astype(np.float32) |
| return output |
|
|
|
|
| class Zoom(Transform): |
| """ |
| Zooms an ND image using :py:class:`torch.nn.functional.interpolate`. |
| For details, please see https://pytorch.org/docs/stable/nn.functional.html#interpolate. |
| |
| Different from :py:class:`monai.transforms.resize`, this transform takes scaling factors |
| as input, and provides an option of preserving the input spatial size. |
| |
| Args: |
| zoom: The zoom factor along the spatial axes. |
| If a float, zoom is the same for each spatial axis. |
| If a sequence, zoom should contain one value for each spatial axis. |
| mode: {``"nearest"``, ``"linear"``, ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| The interpolation mode. Defaults to ``"area"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| align_corners: This only has an effect when mode is |
| 'linear', 'bilinear', 'bicubic' or 'trilinear'. Default: None. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| keep_size: Should keep original size (padding/slicing if needed), default is True. |
| """ |
|
|
| def __init__( |
| self, |
| zoom: Union[Sequence[float], float], |
| mode: Union[InterpolateMode, str] = InterpolateMode.AREA, |
| align_corners: Optional[bool] = None, |
| keep_size: bool = True, |
| ) -> None: |
| self.zoom = zoom |
| self.mode: InterpolateMode = InterpolateMode(mode) |
| self.align_corners = align_corners |
| self.keep_size = keep_size |
|
|
| def __call__( |
| self, |
| img: np.ndarray, |
| mode: Optional[Union[InterpolateMode, str]] = None, |
| align_corners: Optional[bool] = None, |
| ) -> np.ndarray: |
| """ |
| Args: |
| img: channel first array, must have shape: (num_channels, H[, W, ..., ]). |
| mode: {``"nearest"``, ``"linear"``, ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| The interpolation mode. Defaults to ``self.mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| align_corners: This only has an effect when mode is |
| 'linear', 'bilinear', 'bicubic' or 'trilinear'. Defaults to ``self.align_corners``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| |
| """ |
| _zoom = ensure_tuple_rep(self.zoom, img.ndim - 1) |
| zoomed = _torch_interp( |
| input=torch.as_tensor(np.ascontiguousarray(img), dtype=torch.float).unsqueeze(0), |
| scale_factor=list(_zoom), |
| mode=self.mode.value if mode is None else InterpolateMode(mode).value, |
| align_corners=self.align_corners if align_corners is None else align_corners, |
| ) |
| zoomed = zoomed.squeeze(0).detach().cpu().numpy() |
| if not self.keep_size or np.allclose(img.shape, zoomed.shape): |
| return zoomed |
|
|
| pad_vec = [[0, 0]] * len(img.shape) |
| slice_vec = [slice(None)] * len(img.shape) |
| for idx, (od, zd) in enumerate(zip(img.shape, zoomed.shape)): |
| diff = od - zd |
| half = abs(diff) // 2 |
| if diff > 0: |
| pad_vec[idx] = [half, diff - half] |
| elif diff < 0: |
| slice_vec[idx] = slice(half, half + od) |
| zoomed = np.pad(zoomed, pad_vec, mode=NumpyPadMode.EDGE.value) |
| return zoomed[tuple(slice_vec)] |
|
|
|
|
| class Rotate90(Transform): |
| """ |
| Rotate an array by 90 degrees in the plane specified by `axes`. |
| """ |
|
|
| def __init__(self, k: int = 1, spatial_axes: Tuple[int, int] = (0, 1)) -> None: |
| """ |
| Args: |
| k: number of times to rotate by 90 degrees. |
| spatial_axes: 2 int numbers, defines the plane to rotate with 2 spatial axes. |
| Default: (0, 1), this is the first two axis in spatial dimensions. |
| """ |
| self.k = k |
| self.spatial_axes = spatial_axes |
|
|
| def __call__(self, img: np.ndarray) -> np.ndarray: |
| """ |
| Args: |
| img: channel first array, must have shape: (num_channels, H[, W, ..., ]), |
| """ |
| rotated = list() |
| for channel in img: |
| rotated.append(np.rot90(channel, self.k, self.spatial_axes)) |
| return np.stack(rotated).astype(img.dtype) |
|
|
|
|
| class RandRotate90(Randomizable, Transform): |
| """ |
| With probability `prob`, input arrays are rotated by 90 degrees |
| in the plane specified by `spatial_axes`. |
| """ |
|
|
| def __init__(self, prob: float = 0.1, max_k: int = 3, spatial_axes: Tuple[int, int] = (0, 1)) -> None: |
| """ |
| Args: |
| prob: probability of rotating. |
| (Default 0.1, with 10% probability it returns a rotated array) |
| max_k: number of rotations will be sampled from `np.random.randint(max_k) + 1`, (Default 3). |
| spatial_axes: 2 int numbers, defines the plane to rotate with 2 spatial axes. |
| Default: (0, 1), this is the first two axis in spatial dimensions. |
| """ |
| self.prob = min(max(prob, 0.0), 1.0) |
| self.max_k = max_k |
| self.spatial_axes = spatial_axes |
|
|
| self._do_transform = False |
| self._rand_k = 0 |
|
|
| def randomize(self, data: Optional[Any] = None) -> None: |
| self._rand_k = self.R.randint(self.max_k) + 1 |
| self._do_transform = self.R.random() < self.prob |
|
|
| def __call__(self, img: np.ndarray) -> np.ndarray: |
| """ |
| Args: |
| img: channel first array, must have shape: (num_channels, H[, W, ..., ]), |
| """ |
| self.randomize() |
| if not self._do_transform: |
| return img |
| rotator = Rotate90(self._rand_k, self.spatial_axes) |
| return rotator(img) |
|
|
|
|
| class RandRotate(Randomizable, Transform): |
| """ |
| Randomly rotate the input arrays. |
| |
| Args: |
| range_x: Range of rotation angle in degrees in the plane defined by the first and second axes. |
| If single number, angle is uniformly sampled from (-range_x, range_x). |
| range_y: Range of rotation angle in degrees in the plane defined by the first and third axes. |
| If single number, angle is uniformly sampled from (-range_y, range_y). |
| range_z: Range of rotation angle in degrees in the plane defined by the second and third axes. |
| If single number, angle is uniformly sampled from (-range_z, range_z). |
| prob: Probability of rotation. |
| keep_size: If it is False, the output shape is adapted so that the |
| input array is contained completely in the output. |
| If it is True, the output shape is the same as the input. Default is True. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``"border"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| align_corners: Defaults to False. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| dtype: data type for resampling computation. Defaults to ``np.float64`` for best precision. |
| If None, use the data type of input data. To be compatible with other modules, |
| the output data type is always ``np.float32``. |
| """ |
|
|
| def __init__( |
| self, |
| range_x: Union[Tuple[float, float], float] = 0.0, |
| range_y: Union[Tuple[float, float], float] = 0.0, |
| range_z: Union[Tuple[float, float], float] = 0.0, |
| prob: float = 0.1, |
| keep_size: bool = True, |
| mode: Union[GridSampleMode, str] = GridSampleMode.BILINEAR, |
| padding_mode: Union[GridSamplePadMode, str] = GridSamplePadMode.BORDER, |
| align_corners: bool = False, |
| dtype: Optional[np.dtype] = np.float64, |
| ) -> None: |
| self.range_x = ensure_tuple(range_x) |
| if len(self.range_x) == 1: |
| self.range_x = tuple(sorted([-self.range_x[0], self.range_x[0]])) |
| self.range_y = ensure_tuple(range_y) |
| if len(self.range_y) == 1: |
| self.range_y = tuple(sorted([-self.range_y[0], self.range_y[0]])) |
| self.range_z = ensure_tuple(range_z) |
| if len(self.range_z) == 1: |
| self.range_z = tuple(sorted([-self.range_z[0], self.range_z[0]])) |
|
|
| self.prob = prob |
| self.keep_size = keep_size |
| self.mode: GridSampleMode = GridSampleMode(mode) |
| self.padding_mode: GridSamplePadMode = GridSamplePadMode(padding_mode) |
| self.align_corners = align_corners |
| self.dtype = dtype |
|
|
| self._do_transform = False |
| self.x = 0.0 |
| self.y = 0.0 |
| self.z = 0.0 |
|
|
| def randomize(self, data: Optional[Any] = None) -> None: |
| self._do_transform = self.R.random_sample() < self.prob |
| self.x = self.R.uniform(low=self.range_x[0], high=self.range_x[1]) |
| self.y = self.R.uniform(low=self.range_y[0], high=self.range_y[1]) |
| self.z = self.R.uniform(low=self.range_z[0], high=self.range_z[1]) |
|
|
| def __call__( |
| self, |
| img: np.ndarray, |
| mode: Optional[Union[GridSampleMode, str]] = None, |
| padding_mode: Optional[Union[GridSamplePadMode, str]] = None, |
| align_corners: Optional[bool] = None, |
| dtype: Optional[np.dtype] = None, |
| ) -> np.ndarray: |
| """ |
| Args: |
| img: channel first array, must have shape 2D: (nchannels, H, W), or 3D: (nchannels, H, W, D). |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| align_corners: Defaults to ``self.align_corners``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| dtype: data type for resampling computation. Defaults to ``self.dtype``. |
| If None, use the data type of input data. To be compatible with other modules, |
| the output data type is always ``np.float32``. |
| """ |
| self.randomize() |
| if not self._do_transform: |
| return img |
| rotator = Rotate( |
| angle=self.x if img.ndim == 3 else (self.x, self.y, self.z), |
| keep_size=self.keep_size, |
| mode=mode or self.mode, |
| padding_mode=padding_mode or self.padding_mode, |
| align_corners=self.align_corners if align_corners is None else align_corners, |
| dtype=dtype or self.dtype or img.dtype, |
| ) |
| return rotator(img) |
|
|
|
|
| class RandFlip(Randomizable, Transform): |
| """ |
| Randomly flips the image along axes. Preserves shape. |
| See numpy.flip for additional details. |
| https://docs.scipy.org/doc/numpy/reference/generated/numpy.flip.html |
| |
| Args: |
| prob: Probability of flipping. |
| spatial_axis: Spatial axes along which to flip over. Default is None. |
| """ |
|
|
| def __init__(self, prob: float = 0.1, spatial_axis: Optional[Union[Sequence[int], int]] = None) -> None: |
| self.prob = prob |
| self.flipper = Flip(spatial_axis=spatial_axis) |
| self._do_transform = False |
|
|
| def randomize(self, data: Optional[Any] = None) -> None: |
| self._do_transform = self.R.random_sample() < self.prob |
|
|
| def __call__(self, img: np.ndarray) -> np.ndarray: |
| """ |
| Args: |
| img: channel first array, must have shape: (num_channels, H[, W, ..., ]), |
| """ |
| self.randomize() |
| if not self._do_transform: |
| return img |
| return self.flipper(img) |
|
|
|
|
| class RandZoom(Randomizable, Transform): |
| """ |
| Randomly zooms input arrays with given probability within given zoom range. |
| |
| Args: |
| prob: Probability of zooming. |
| min_zoom: Min zoom factor. Can be float or sequence same size as image. |
| If a float, select a random factor from `[min_zoom, max_zoom]` then apply to all spatial dims |
| to keep the original spatial shape ratio. |
| If a sequence, min_zoom should contain one value for each spatial axis. |
| max_zoom: Max zoom factor. Can be float or sequence same size as image. |
| If a float, select a random factor from `[min_zoom, max_zoom]` then apply to all spatial dims |
| to keep the original spatial shape ratio. |
| If a sequence, max_zoom should contain one value for each spatial axis. |
| mode: {``"nearest"``, ``"linear"``, ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| The interpolation mode. Defaults to ``"area"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| align_corners: This only has an effect when mode is |
| 'linear', 'bilinear', 'bicubic' or 'trilinear'. Default: None. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| keep_size: Should keep original size (pad if needed), default is True. |
| """ |
|
|
| def __init__( |
| self, |
| prob: float = 0.1, |
| min_zoom: Union[Sequence[float], float] = 0.9, |
| max_zoom: Union[Sequence[float], float] = 1.1, |
| mode: Union[InterpolateMode, str] = InterpolateMode.AREA, |
| align_corners: Optional[bool] = None, |
| keep_size: bool = True, |
| ) -> None: |
| self.min_zoom = ensure_tuple(min_zoom) |
| self.max_zoom = ensure_tuple(max_zoom) |
| assert len(self.min_zoom) == len(self.max_zoom), "min_zoom and max_zoom must have same length." |
| self.prob = prob |
| self.mode: InterpolateMode = InterpolateMode(mode) |
| self.align_corners = align_corners |
| self.keep_size = keep_size |
|
|
| self._do_transform = False |
| self._zoom: Union[List[float], float] = 1.0 |
|
|
| def randomize(self, data: Optional[Any] = None) -> None: |
| self._do_transform = self.R.random_sample() < self.prob |
| self._zoom = [self.R.uniform(l, h) for l, h in zip(self.min_zoom, self.max_zoom)] |
| if len(self._zoom) == 1: |
| |
| self._zoom = self._zoom[0] |
|
|
| def __call__( |
| self, |
| img: np.ndarray, |
| mode: Optional[Union[InterpolateMode, str]] = None, |
| align_corners: Optional[bool] = None, |
| ) -> np.ndarray: |
| """ |
| Args: |
| img: channel first array, must have shape 2D: (nchannels, H, W), or 3D: (nchannels, H, W, D). |
| mode: {``"nearest"``, ``"linear"``, ``"bilinear"``, ``"bicubic"``, ``"trilinear"``, ``"area"``} |
| The interpolation mode. Defaults to ``self.mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| align_corners: This only has an effect when mode is |
| 'linear', 'bilinear', 'bicubic' or 'trilinear'. Defaults to ``self.align_corners``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#interpolate |
| """ |
| |
| self.randomize() |
| _dtype = np.float32 |
| if not self._do_transform: |
| return img.astype(_dtype) |
| zoomer = Zoom(self._zoom, keep_size=self.keep_size) |
| return zoomer( |
| img, |
| mode=mode or self.mode, |
| align_corners=self.align_corners if align_corners is None else align_corners, |
| ).astype(_dtype) |
|
|
|
|
| class AffineGrid(Transform): |
| """ |
| Affine transforms on the coordinates. |
| |
| Args: |
| rotate_range: angle range in radians. rotate_range[0] with be used to generate the 1st rotation |
| parameter from `uniform[-rotate_range[0], rotate_range[0])`. Similarly, `rotate_range[1]` and |
| `rotate_range[2]` are used in 3D affine for the range of 2nd and 3rd axes. |
| shear_range: shear_range[0] with be used to generate the 1st shearing parameter from |
| `uniform[-shear_range[0], shear_range[0])`. Similarly, `shear_range[1]` to |
| `shear_range[N]` controls the range of the uniform distribution used to generate the 2nd to |
| N-th parameter. |
| translate_range : translate_range[0] with be used to generate the 1st shift parameter from |
| `uniform[-translate_range[0], translate_range[0])`. Similarly, `translate_range[1]` |
| to `translate_range[N]` controls the range of the uniform distribution used to generate |
| the 2nd to N-th parameter. |
| scale_range: scaling_range[0] with be used to generate the 1st scaling factor from |
| `uniform[-scale_range[0], scale_range[0]) + 1.0`. Similarly, `scale_range[1]` to |
| `scale_range[N]` controls the range of the uniform distribution used to generate the 2nd to |
| N-th parameter. |
| as_tensor_output: whether to output tensor instead of numpy array. |
| defaults to True. |
| device: device to store the output grid data. |
| |
| """ |
|
|
| def __init__( |
| self, |
| rotate_params: Optional[Union[Sequence[float], float]] = None, |
| shear_params: Optional[Union[Sequence[float], float]] = None, |
| translate_params: Optional[Union[Sequence[float], float]] = None, |
| scale_params: Optional[Union[Sequence[float], float]] = None, |
| as_tensor_output: bool = True, |
| device: Optional[torch.device] = None, |
| ) -> None: |
| self.rotate_params = rotate_params |
| self.shear_params = shear_params |
| self.translate_params = translate_params |
| self.scale_params = scale_params |
|
|
| self.as_tensor_output = as_tensor_output |
| self.device = device |
|
|
| def __call__( |
| self, spatial_size: Optional[Sequence[int]] = None, grid: Optional[Union[np.ndarray, torch.Tensor]] = None |
| ) -> Union[np.ndarray, torch.Tensor]: |
| """ |
| Args: |
| spatial_size: output grid size. |
| grid: grid to be transformed. Shape must be (3, H, W) for 2D or (4, H, W, D) for 3D. |
| |
| Raises: |
| ValueError: When ``grid=None`` and ``spatial_size=None``. Incompatible values. |
| |
| """ |
| if grid is None: |
| if spatial_size is not None: |
| grid = create_grid(spatial_size) |
| else: |
| raise ValueError("Incompatible values: grid=None and spatial_size=None.") |
|
|
| spatial_dims = len(grid.shape) - 1 |
| affine = np.eye(spatial_dims + 1) |
| if self.rotate_params: |
| affine = affine @ create_rotate(spatial_dims, self.rotate_params) |
| if self.shear_params: |
| affine = affine @ create_shear(spatial_dims, self.shear_params) |
| if self.translate_params: |
| affine = affine @ create_translate(spatial_dims, self.translate_params) |
| if self.scale_params: |
| affine = affine @ create_scale(spatial_dims, self.scale_params) |
| affine = torch.as_tensor(np.ascontiguousarray(affine), device=self.device) |
|
|
| grid = torch.tensor(grid) if not torch.is_tensor(grid) else grid.detach().clone() |
| if self.device: |
| grid = grid.to(self.device) |
| grid = (affine.float() @ grid.reshape((grid.shape[0], -1)).float()).reshape([-1] + list(grid.shape[1:])) |
| if self.as_tensor_output: |
| return grid |
| return grid.cpu().numpy() |
|
|
|
|
| class RandAffineGrid(Randomizable, Transform): |
| """ |
| Generate randomised affine grid. |
| """ |
|
|
| def __init__( |
| self, |
| rotate_range: Optional[Union[Sequence[float], float]] = None, |
| shear_range: Optional[Union[Sequence[float], float]] = None, |
| translate_range: Optional[Union[Sequence[float], float]] = None, |
| scale_range: Optional[Union[Sequence[float], float]] = None, |
| as_tensor_output: bool = True, |
| device: Optional[torch.device] = None, |
| ) -> None: |
| """ |
| Args: |
| rotate_range: angle range in radians. rotate_range[0] with be used to generate the 1st rotation |
| parameter from `uniform[-rotate_range[0], rotate_range[0])`. Similarly, `rotate_range[1]` and |
| `rotate_range[2]` are used in 3D affine for the range of 2nd and 3rd axes. |
| shear_range: shear_range[0] with be used to generate the 1st shearing parameter from |
| `uniform[-shear_range[0], shear_range[0])`. Similarly, `shear_range[1]` to |
| `shear_range[N]` controls the range of the uniform distribution used to generate the 2nd to |
| N-th parameter. |
| translate_range : translate_range[0] with be used to generate the 1st shift parameter from |
| `uniform[-translate_range[0], translate_range[0])`. Similarly, `translate_range[1]` |
| to `translate_range[N]` controls the range of the uniform distribution used to generate |
| the 2nd to N-th parameter. |
| scale_range: scaling_range[0] with be used to generate the 1st scaling factor from |
| `uniform[-scale_range[0], scale_range[0]) + 1.0`. Similarly, `scale_range[1]` to |
| `scale_range[N]` controls the range of the uniform distribution used to generate the 2nd to |
| N-th parameter. |
| as_tensor_output: whether to output tensor instead of numpy array. |
| defaults to True. |
| device: device to store the output grid data. |
| |
| See also: |
| - :py:meth:`monai.transforms.utils.create_rotate` |
| - :py:meth:`monai.transforms.utils.create_shear` |
| - :py:meth:`monai.transforms.utils.create_translate` |
| - :py:meth:`monai.transforms.utils.create_scale` |
| """ |
| self.rotate_range = ensure_tuple(rotate_range) |
| self.shear_range = ensure_tuple(shear_range) |
| self.translate_range = ensure_tuple(translate_range) |
| self.scale_range = ensure_tuple(scale_range) |
|
|
| self.rotate_params: Optional[List[float]] = None |
| self.shear_params: Optional[List[float]] = None |
| self.translate_params: Optional[List[float]] = None |
| self.scale_params: Optional[List[float]] = None |
|
|
| self.as_tensor_output = as_tensor_output |
| self.device = device |
|
|
| def randomize(self, data: Optional[Any] = None) -> None: |
| if self.rotate_range: |
| self.rotate_params = [self.R.uniform(-f, f) for f in self.rotate_range if f is not None] |
| if self.shear_range: |
| self.shear_params = [self.R.uniform(-f, f) for f in self.shear_range if f is not None] |
| if self.translate_range: |
| self.translate_params = [self.R.uniform(-f, f) for f in self.translate_range if f is not None] |
| if self.scale_range: |
| self.scale_params = [self.R.uniform(-f, f) + 1.0 for f in self.scale_range if f is not None] |
|
|
| def __call__( |
| self, spatial_size: Optional[Sequence[int]] = None, grid: Optional[Union[np.ndarray, torch.Tensor]] = None |
| ) -> Union[np.ndarray, torch.Tensor]: |
| """ |
| Args: |
| spatial_size: output grid size. |
| grid: grid to be transformed. Shape must be (3, H, W) for 2D or (4, H, W, D) for 3D. |
| |
| Returns: |
| a 2D (3xHxW) or 3D (4xHxWxD) grid. |
| """ |
| self.randomize() |
| affine_grid = AffineGrid( |
| rotate_params=self.rotate_params, |
| shear_params=self.shear_params, |
| translate_params=self.translate_params, |
| scale_params=self.scale_params, |
| as_tensor_output=self.as_tensor_output, |
| device=self.device, |
| ) |
| return affine_grid(spatial_size, grid) |
|
|
|
|
| class RandDeformGrid(Randomizable, Transform): |
| """ |
| Generate random deformation grid. |
| """ |
|
|
| def __init__( |
| self, |
| spacing: Union[Sequence[float], float], |
| magnitude_range: Tuple[float, float], |
| as_tensor_output: bool = True, |
| device: Optional[torch.device] = None, |
| ) -> None: |
| """ |
| Args: |
| spacing: spacing of the grid in 2D or 3D. |
| e.g., spacing=(1, 1) indicates pixel-wise deformation in 2D, |
| spacing=(1, 1, 1) indicates voxel-wise deformation in 3D, |
| spacing=(2, 2) indicates deformation field defined on every other pixel in 2D. |
| magnitude_range: the random offsets will be generated from |
| `uniform[magnitude[0], magnitude[1])`. |
| as_tensor_output: whether to output tensor instead of numpy array. |
| defaults to True. |
| device: device to store the output grid data. |
| """ |
| self.spacing = spacing |
| self.magnitude = magnitude_range |
|
|
| self.rand_mag = 1.0 |
| self.as_tensor_output = as_tensor_output |
| self.random_offset = 0.0 |
| self.device = device |
|
|
| def randomize(self, grid_size: Sequence[int]) -> None: |
| self.random_offset = self.R.normal(size=([len(grid_size)] + list(grid_size))).astype(np.float32) |
| self.rand_mag = self.R.uniform(self.magnitude[0], self.magnitude[1]) |
|
|
| def __call__(self, spatial_size: Sequence[int]) -> Union[np.ndarray, torch.Tensor]: |
| """ |
| Args: |
| spatial_size: spatial size of the grid. |
| """ |
| self.spacing = fall_back_tuple(self.spacing, (1.0,) * len(spatial_size)) |
| control_grid = create_control_grid(spatial_size, self.spacing) |
| self.randomize(control_grid.shape[1:]) |
| control_grid[: len(spatial_size)] += self.rand_mag * self.random_offset |
| if self.as_tensor_output: |
| control_grid = torch.as_tensor(np.ascontiguousarray(control_grid), device=self.device) |
| return control_grid |
|
|
|
|
| class Resample(Transform): |
| def __init__( |
| self, |
| mode: Union[GridSampleMode, str] = GridSampleMode.BILINEAR, |
| padding_mode: Union[GridSamplePadMode, str] = GridSamplePadMode.BORDER, |
| as_tensor_output: bool = False, |
| device: Optional[torch.device] = None, |
| ) -> None: |
| """ |
| computes output image using values from `img`, locations from `grid` using pytorch. |
| supports spatially 2D or 3D (num_channels, H, W[, D]). |
| |
| Args: |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``"border"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| as_tensor_output: whether to return a torch tensor. Defaults to False. |
| device: device on which the tensor will be allocated. |
| """ |
| self.mode: GridSampleMode = GridSampleMode(mode) |
| self.padding_mode: GridSamplePadMode = GridSamplePadMode(padding_mode) |
| self.as_tensor_output = as_tensor_output |
| self.device = device |
|
|
| def __call__( |
| self, |
| img: Union[np.ndarray, torch.Tensor], |
| grid: Optional[Union[np.ndarray, torch.Tensor]] = None, |
| mode: Optional[Union[GridSampleMode, str]] = None, |
| padding_mode: Optional[Union[GridSamplePadMode, str]] = None, |
| ) -> Union[np.ndarray, torch.Tensor]: |
| """ |
| Args: |
| img: shape must be (num_channels, H, W[, D]). |
| grid: shape must be (3, H, W) for 2D or (4, H, W, D) for 3D. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| """ |
|
|
| if not torch.is_tensor(img): |
| img = torch.as_tensor(np.ascontiguousarray(img)) |
| assert grid is not None, "Error, grid argument must be supplied as an ndarray or tensor " |
| grid = torch.tensor(grid) if not torch.is_tensor(grid) else grid.detach().clone() |
| if self.device: |
| img = img.to(self.device) |
| grid = grid.to(self.device) |
|
|
| for i, dim in enumerate(img.shape[1:]): |
| grid[i] = 2.0 * grid[i] / (dim - 1.0) |
| grid = grid[:-1] / grid[-1:] |
| index_ordering: List[int] = list(range(img.ndimension() - 2, -1, -1)) |
| grid = grid[index_ordering] |
| grid = grid.permute(list(range(grid.ndimension()))[1:] + [0]) |
| out = torch.nn.functional.grid_sample( |
| img.unsqueeze(0).float(), |
| grid.unsqueeze(0).float(), |
| mode=self.mode.value if mode is None else GridSampleMode(mode).value, |
| padding_mode=self.padding_mode.value if padding_mode is None else GridSamplePadMode(padding_mode).value, |
| align_corners=True, |
| )[0] |
| if self.as_tensor_output: |
| return out |
| return out.cpu().numpy() |
|
|
|
|
| class Affine(Transform): |
| """ |
| Transform ``img`` given the affine parameters. |
| """ |
|
|
| def __init__( |
| self, |
| rotate_params: Optional[Union[Sequence[float], float]] = None, |
| shear_params: Optional[Union[Sequence[float], float]] = None, |
| translate_params: Optional[Union[Sequence[float], float]] = None, |
| scale_params: Optional[Union[Sequence[float], float]] = None, |
| spatial_size: Optional[Union[Sequence[int], int]] = None, |
| mode: Union[GridSampleMode, str] = GridSampleMode.BILINEAR, |
| padding_mode: Union[GridSamplePadMode, str] = GridSamplePadMode.REFLECTION, |
| as_tensor_output: bool = False, |
| device: Optional[torch.device] = None, |
| ) -> None: |
| """ |
| The affine transformations are applied in rotate, shear, translate, scale order. |
| |
| Args: |
| rotate_params: a rotation angle in radians, a scalar for 2D image, a tuple of 3 floats for 3D. |
| Defaults to no rotation. |
| shear_params: a tuple of 2 floats for 2D, a tuple of 6 floats for 3D. Defaults to no shearing. |
| translate_params: a tuple of 2 floats for 2D, a tuple of 3 floats for 3D. Translation is in |
| pixel/voxel relative to the center of the input image. Defaults to no translation. |
| scale_params: a tuple of 2 floats for 2D, a tuple of 3 floats for 3D. Defaults to no scaling. |
| spatial_size: output image spatial size. |
| if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| the transform will use the spatial size of `img`. |
| if the components of the `spatial_size` are non-positive values, the transform will use the |
| corresponding components of img size. For example, `spatial_size=(32, -1)` will be adapted |
| to `(32, 64)` if the second spatial dimension size of img is `64`. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``"reflection"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| as_tensor_output: the computation is implemented using pytorch tensors, this option specifies |
| whether to convert it back to numpy arrays. |
| device: device on which the tensor will be allocated. |
| """ |
| self.affine_grid = AffineGrid( |
| rotate_params=rotate_params, |
| shear_params=shear_params, |
| translate_params=translate_params, |
| scale_params=scale_params, |
| as_tensor_output=True, |
| device=device, |
| ) |
| self.resampler = Resample(as_tensor_output=as_tensor_output, device=device) |
| self.spatial_size = spatial_size |
| self.mode: GridSampleMode = GridSampleMode(mode) |
| self.padding_mode: GridSamplePadMode = GridSamplePadMode(padding_mode) |
|
|
| def __call__( |
| self, |
| img: Union[np.ndarray, torch.Tensor], |
| spatial_size: Optional[Union[Sequence[int], int]] = None, |
| mode: Optional[Union[GridSampleMode, str]] = None, |
| padding_mode: Optional[Union[GridSamplePadMode, str]] = None, |
| ) -> Union[np.ndarray, torch.Tensor]: |
| """ |
| Args: |
| img: shape must be (num_channels, H, W[, D]), |
| spatial_size: output image spatial size. |
| if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| the transform will use the spatial size of `img`. |
| if `img` has two spatial dimensions, `spatial_size` should have 2 elements [h, w]. |
| if `img` has three spatial dimensions, `spatial_size` should have 3 elements [h, w, d]. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| """ |
| sp_size = fall_back_tuple(spatial_size or self.spatial_size, img.shape[1:]) |
| grid = self.affine_grid(spatial_size=sp_size) |
| return self.resampler( |
| img=img, grid=grid, mode=mode or self.mode, padding_mode=padding_mode or self.padding_mode |
| ) |
|
|
|
|
| class RandAffine(Randomizable, Transform): |
| """ |
| Random affine transform. |
| """ |
|
|
| def __init__( |
| self, |
| prob: float = 0.1, |
| rotate_range: Optional[Union[Sequence[float], float]] = None, |
| shear_range: Optional[Union[Sequence[float], float]] = None, |
| translate_range: Optional[Union[Sequence[float], float]] = None, |
| scale_range: Optional[Union[Sequence[float], float]] = None, |
| spatial_size: Optional[Union[Sequence[float], float]] = None, |
| mode: Union[GridSampleMode, str] = GridSampleMode.BILINEAR, |
| padding_mode: Union[GridSamplePadMode, str] = GridSamplePadMode.REFLECTION, |
| as_tensor_output: bool = True, |
| device: Optional[torch.device] = None, |
| ) -> None: |
| """ |
| Args: |
| prob: probability of returning a randomized affine grid. |
| defaults to 0.1, with 10% chance returns a randomized grid. |
| rotate_range: angle range in radians. rotate_range[0] with be used to generate the 1st rotation |
| parameter from `uniform[-rotate_range[0], rotate_range[0])`. Similarly, `rotate_range[1]` and |
| `rotate_range[2]` are used in 3D affine for the range of 2nd and 3rd axes. |
| shear_range: shear_range[0] with be used to generate the 1st shearing parameter from |
| `uniform[-shear_range[0], shear_range[0])`. Similarly, `shear_range[1]` to |
| `shear_range[N]` controls the range of the uniform distribution used to generate the 2nd to |
| N-th parameter. |
| translate_range : translate_range[0] with be used to generate the 1st shift parameter from |
| `uniform[-translate_range[0], translate_range[0])`. Similarly, `translate_range[1]` |
| to `translate_range[N]` controls the range of the uniform distribution used to generate |
| the 2nd to N-th parameter. |
| scale_range: scaling_range[0] with be used to generate the 1st scaling factor from |
| `uniform[-scale_range[0], scale_range[0]) + 1.0`. Similarly, `scale_range[1]` to |
| `scale_range[N]` controls the range of the uniform distribution used to generate the 2nd to |
| N-th parameter. |
| spatial_size: output image spatial size. |
| if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| the transform will use the spatial size of `img`. |
| if the components of the `spatial_size` are non-positive values, the transform will use the |
| corresponding components of img size. For example, `spatial_size=(32, -1)` will be adapted |
| to `(32, 64)` if the second spatial dimension size of img is `64`. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``"reflection"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| as_tensor_output: the computation is implemented using pytorch tensors, this option specifies |
| whether to convert it back to numpy arrays. |
| device: device on which the tensor will be allocated. |
| |
| See also: |
| - :py:class:`RandAffineGrid` for the random affine parameters configurations. |
| - :py:class:`Affine` for the affine transformation parameters configurations. |
| """ |
|
|
| self.rand_affine_grid = RandAffineGrid( |
| rotate_range=rotate_range, |
| shear_range=shear_range, |
| translate_range=translate_range, |
| scale_range=scale_range, |
| as_tensor_output=True, |
| device=device, |
| ) |
| self.resampler = Resample(as_tensor_output=as_tensor_output, device=device) |
|
|
| self.spatial_size = spatial_size |
| self.mode: GridSampleMode = GridSampleMode(mode) |
| self.padding_mode: GridSamplePadMode = GridSamplePadMode(padding_mode) |
|
|
| self.do_transform = False |
| self.prob = prob |
|
|
| def set_random_state( |
| self, seed: Optional[int] = None, state: Optional[np.random.RandomState] = None |
| ) -> "RandAffine": |
| self.rand_affine_grid.set_random_state(seed, state) |
| super().set_random_state(seed, state) |
| return self |
|
|
| def randomize(self, data: Optional[Any] = None) -> None: |
| self.do_transform = self.R.rand() < self.prob |
| self.rand_affine_grid.randomize() |
|
|
| def __call__( |
| self, |
| img: Union[np.ndarray, torch.Tensor], |
| spatial_size: Optional[Union[Sequence[int], int]] = None, |
| mode: Optional[Union[GridSampleMode, str]] = None, |
| padding_mode: Optional[Union[GridSamplePadMode, str]] = None, |
| ) -> Union[np.ndarray, torch.Tensor]: |
| """ |
| Args: |
| img: shape must be (num_channels, H, W[, D]), |
| spatial_size: output image spatial size. |
| if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| the transform will use the spatial size of `img`. |
| if `img` has two spatial dimensions, `spatial_size` should have 2 elements [h, w]. |
| if `img` has three spatial dimensions, `spatial_size` should have 3 elements [h, w, d]. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| """ |
| self.randomize() |
|
|
| sp_size = fall_back_tuple(spatial_size or self.spatial_size, img.shape[1:]) |
| if self.do_transform: |
| grid = self.rand_affine_grid(spatial_size=sp_size) |
| else: |
| grid = create_grid(spatial_size=sp_size) |
| return self.resampler( |
| img=img, grid=grid, mode=mode or self.mode, padding_mode=padding_mode or self.padding_mode |
| ) |
|
|
|
|
| class Rand2DElastic(Randomizable, Transform): |
| """ |
| Random elastic deformation and affine in 2D |
| """ |
|
|
| def __init__( |
| self, |
| spacing: Union[Tuple[float, float], float], |
| magnitude_range: Tuple[float, float], |
| prob: float = 0.1, |
| rotate_range: Optional[Union[Sequence[float], float]] = None, |
| shear_range: Optional[Union[Sequence[float], float]] = None, |
| translate_range: Optional[Union[Sequence[float], float]] = None, |
| scale_range: Optional[Union[Sequence[float], float]] = None, |
| spatial_size: Optional[Union[Sequence[int], int]] = None, |
| mode: Union[GridSampleMode, str] = GridSampleMode.BILINEAR, |
| padding_mode: Union[GridSamplePadMode, str] = GridSamplePadMode.REFLECTION, |
| as_tensor_output: bool = False, |
| device: Optional[torch.device] = None, |
| ) -> None: |
| """ |
| Args: |
| spacing : distance in between the control points. |
| magnitude_range: the random offsets will be generated from ``uniform[magnitude[0], magnitude[1])``. |
| prob: probability of returning a randomized elastic transform. |
| defaults to 0.1, with 10% chance returns a randomized elastic transform, |
| otherwise returns a ``spatial_size`` centered area extracted from the input image. |
| rotate_range: angle range in radians. rotate_range[0] with be used to generate the 1st rotation |
| parameter from `uniform[-rotate_range[0], rotate_range[0])`. |
| shear_range: shear_range[0] with be used to generate the 1st shearing parameter from |
| `uniform[-shear_range[0], shear_range[0])`. Similarly, `shear_range[1]` controls |
| the range of the uniform distribution used to generate the 2nd parameter. |
| translate_range : translate_range[0] with be used to generate the 1st shift parameter from |
| `uniform[-translate_range[0], translate_range[0])`. Similarly, `translate_range[1]` controls |
| the range of the uniform distribution used to generate the 2nd parameter. |
| scale_range: scaling_range[0] with be used to generate the 1st scaling factor from |
| `uniform[-scale_range[0], scale_range[0]) + 1.0`. Similarly, `scale_range[1]` controls |
| the range of the uniform distribution used to generate the 2nd parameter. |
| spatial_size: specifying output image spatial size [h, w]. |
| if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| the transform will use the spatial size of `img`. |
| if the components of the `spatial_size` are non-positive values, the transform will use the |
| corresponding components of img size. For example, `spatial_size=(32, -1)` will be adapted |
| to `(32, 64)` if the second spatial dimension size of img is `64`. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``"reflection"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| as_tensor_output: the computation is implemented using pytorch tensors, this option specifies |
| whether to convert it back to numpy arrays. |
| device: device on which the tensor will be allocated. |
| |
| See also: |
| - :py:class:`RandAffineGrid` for the random affine parameters configurations. |
| - :py:class:`Affine` for the affine transformation parameters configurations. |
| """ |
| self.deform_grid = RandDeformGrid( |
| spacing=spacing, magnitude_range=magnitude_range, as_tensor_output=True, device=device |
| ) |
| self.rand_affine_grid = RandAffineGrid( |
| rotate_range=rotate_range, |
| shear_range=shear_range, |
| translate_range=translate_range, |
| scale_range=scale_range, |
| as_tensor_output=True, |
| device=device, |
| ) |
| self.resampler = Resample(as_tensor_output=as_tensor_output, device=device) |
|
|
| self.spatial_size = spatial_size |
| self.mode: GridSampleMode = GridSampleMode(mode) |
| self.padding_mode: GridSamplePadMode = GridSamplePadMode(padding_mode) |
| self.prob = prob |
| self.do_transform = False |
|
|
| def set_random_state( |
| self, seed: Optional[int] = None, state: Optional[np.random.RandomState] = None |
| ) -> "Rand2DElastic": |
| self.deform_grid.set_random_state(seed, state) |
| self.rand_affine_grid.set_random_state(seed, state) |
| super().set_random_state(seed, state) |
| return self |
|
|
| def randomize(self, spatial_size: Sequence[int]) -> None: |
| self.do_transform = self.R.rand() < self.prob |
| self.deform_grid.randomize(spatial_size) |
| self.rand_affine_grid.randomize() |
|
|
| def __call__( |
| self, |
| img: Union[np.ndarray, torch.Tensor], |
| spatial_size: Optional[Union[Tuple[int, int], int]] = None, |
| mode: Optional[Union[GridSampleMode, str]] = None, |
| padding_mode: Optional[Union[GridSamplePadMode, str]] = None, |
| ) -> Union[np.ndarray, torch.Tensor]: |
| """ |
| Args: |
| img: shape must be (num_channels, H, W), |
| spatial_size: specifying output image spatial size [h, w]. |
| if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| the transform will use the spatial size of `img`. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| """ |
| sp_size = fall_back_tuple(spatial_size or self.spatial_size, img.shape[1:]) |
| self.randomize(spatial_size=sp_size) |
| if self.do_transform: |
| grid = self.deform_grid(spatial_size=sp_size) |
| grid = self.rand_affine_grid(grid=grid) |
| grid = _torch_interp( |
| input=grid.unsqueeze(0), |
| scale_factor=list(ensure_tuple(self.deform_grid.spacing)), |
| mode=InterpolateMode.BICUBIC.value, |
| align_corners=False, |
| ) |
| grid = CenterSpatialCrop(roi_size=sp_size)(grid[0]) |
| else: |
| grid = create_grid(spatial_size=sp_size) |
| return self.resampler(img, grid, mode=mode or self.mode, padding_mode=padding_mode or self.padding_mode) |
|
|
|
|
| class Rand3DElastic(Randomizable, Transform): |
| """ |
| Random elastic deformation and affine in 3D |
| """ |
|
|
| def __init__( |
| self, |
| sigma_range: Tuple[float, float], |
| magnitude_range: Tuple[float, float], |
| prob: float = 0.1, |
| rotate_range: Optional[Union[Sequence[float], float]] = None, |
| shear_range: Optional[Union[Sequence[float], float]] = None, |
| translate_range: Optional[Union[Sequence[float], float]] = None, |
| scale_range: Optional[Union[Sequence[float], float]] = None, |
| spatial_size: Optional[Union[Sequence[int], int]] = None, |
| mode: Union[GridSampleMode, str] = GridSampleMode.BILINEAR, |
| padding_mode: Union[GridSamplePadMode, str] = GridSamplePadMode.REFLECTION, |
| as_tensor_output: bool = False, |
| device: Optional[torch.device] = None, |
| ) -> None: |
| """ |
| Args: |
| sigma_range: a Gaussian kernel with standard deviation sampled from |
| ``uniform[sigma_range[0], sigma_range[1])`` will be used to smooth the random offset grid. |
| magnitude_range: the random offsets on the grid will be generated from |
| ``uniform[magnitude[0], magnitude[1])``. |
| prob: probability of returning a randomized elastic transform. |
| defaults to 0.1, with 10% chance returns a randomized elastic transform, |
| otherwise returns a ``spatial_size`` centered area extracted from the input image. |
| rotate_range: angle range in radians. rotate_range[0] with be used to generate the 1st rotation |
| parameter from `uniform[-rotate_range[0], rotate_range[0])`. Similarly, `rotate_range[1]` and |
| `rotate_range[2]` are used in 3D affine for the range of 2nd and 3rd axes. |
| shear_range: shear_range[0] with be used to generate the 1st shearing parameter from |
| `uniform[-shear_range[0], shear_range[0])`. Similarly, `shear_range[1]` and `shear_range[2]` |
| controls the range of the uniform distribution used to generate the 2nd and 3rd parameters. |
| translate_range : translate_range[0] with be used to generate the 1st shift parameter from |
| `uniform[-translate_range[0], translate_range[0])`. Similarly, `translate_range[1]` and |
| `translate_range[2]` controls the range of the uniform distribution used to generate |
| the 2nd and 3rd parameters. |
| scale_range: scaling_range[0] with be used to generate the 1st scaling factor from |
| `uniform[-scale_range[0], scale_range[0]) + 1.0`. Similarly, `scale_range[1]` and `scale_range[2]` |
| controls the range of the uniform distribution used to generate the 2nd and 3rd parameters. |
| spatial_size: specifying output image spatial size [h, w, d]. |
| if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| the transform will use the spatial size of `img`. |
| if the components of the `spatial_size` are non-positive values, the transform will use the |
| corresponding components of img size. For example, `spatial_size=(32, 32, -1)` will be adapted |
| to `(32, 32, 64)` if the third spatial dimension size of img is `64`. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``"bilinear"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``"reflection"``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| as_tensor_output: the computation is implemented using pytorch tensors, this option specifies |
| whether to convert it back to numpy arrays. |
| device: device on which the tensor will be allocated. |
| |
| See also: |
| - :py:class:`RandAffineGrid` for the random affine parameters configurations. |
| - :py:class:`Affine` for the affine transformation parameters configurations. |
| """ |
| self.rand_affine_grid = RandAffineGrid(rotate_range, shear_range, translate_range, scale_range, True, device) |
| self.resampler = Resample(as_tensor_output=as_tensor_output, device=device) |
|
|
| self.sigma_range = sigma_range |
| self.magnitude_range = magnitude_range |
| self.spatial_size = spatial_size |
| self.mode: GridSampleMode = GridSampleMode(mode) |
| self.padding_mode: GridSamplePadMode = GridSamplePadMode(padding_mode) |
| self.device = device |
|
|
| self.prob = prob |
| self.do_transform = False |
| self.rand_offset = None |
| self.magnitude = 1.0 |
| self.sigma = 1.0 |
|
|
| def set_random_state( |
| self, seed: Optional[int] = None, state: Optional[np.random.RandomState] = None |
| ) -> "Rand3DElastic": |
| self.rand_affine_grid.set_random_state(seed, state) |
| super().set_random_state(seed, state) |
| return self |
|
|
| def randomize(self, grid_size: Sequence[int]) -> None: |
| self.do_transform = self.R.rand() < self.prob |
| if self.do_transform: |
| self.rand_offset = self.R.uniform(-1.0, 1.0, [3] + list(grid_size)).astype(np.float32) |
| self.magnitude = self.R.uniform(self.magnitude_range[0], self.magnitude_range[1]) |
| self.sigma = self.R.uniform(self.sigma_range[0], self.sigma_range[1]) |
| self.rand_affine_grid.randomize() |
|
|
| def __call__( |
| self, |
| img: Union[np.ndarray, torch.Tensor], |
| spatial_size: Optional[Union[Tuple[int, int, int], int]] = None, |
| mode: Optional[Union[GridSampleMode, str]] = None, |
| padding_mode: Optional[Union[GridSamplePadMode, str]] = None, |
| ) -> Union[np.ndarray, torch.Tensor]: |
| """ |
| Args: |
| img: shape must be (num_channels, H, W, D), |
| spatial_size: specifying spatial 3D output image spatial size [h, w, d]. |
| if `spatial_size` and `self.spatial_size` are not defined, or smaller than 1, |
| the transform will use the spatial size of `img`. |
| mode: {``"bilinear"``, ``"nearest"``} |
| Interpolation mode to calculate output values. Defaults to ``self.mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} |
| Padding mode for outside grid values. Defaults to ``self.padding_mode``. |
| See also: https://pytorch.org/docs/stable/nn.functional.html#grid-sample |
| """ |
| sp_size = fall_back_tuple(spatial_size or self.spatial_size, img.shape[1:]) |
| self.randomize(grid_size=sp_size) |
| grid = create_grid(spatial_size=sp_size) |
| if self.do_transform: |
| assert self.rand_offset is not None |
| grid = torch.as_tensor(np.ascontiguousarray(grid), device=self.device) |
| gaussian = GaussianFilter(3, self.sigma, 3.0).to(device=self.device) |
| offset = torch.as_tensor(self.rand_offset, device=self.device).unsqueeze(0) |
| grid[:3] += gaussian(offset)[0] * self.magnitude |
| grid = self.rand_affine_grid(grid=grid) |
| return self.resampler(img, grid, mode=mode or self.mode, padding_mode=padding_mode or self.padding_mode) |
|
|