| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| from collections.abc import Sequence |
|
|
| import torch |
| import torch.nn as nn |
| import torch.nn.functional as F |
|
|
| from monai.networks.blocks.convolutions import Convolution |
| from monai.networks.blocks.upsample import UpSample |
| from monai.networks.blocks.warp import DVF2DDF, Warp |
| from monai.networks.layers.simplelayers import SkipConnection |
| from monai.utils import alias, export |
|
|
| __all__ = ["VoxelMorphUNet", "voxelmorphunet", "VoxelMorph", "voxelmorph"] |
|
|
|
|
| @export("monai.networks.nets") |
| @alias("voxelmorphunet") |
| class VoxelMorphUNet(nn.Module): |
| """ |
| The backbone network used in VoxelMorph. See :py:class:`monai.networks.nets.VoxelMorph` for more details. |
| |
| A concatenated pair of images (moving and fixed) is first passed through a UNet. The output of the UNet is then |
| passed through a series of convolution blocks to produce the final prediction of the displacement field (DDF) or the |
| stationary velocity field (DVF). |
| |
| In the original implementation, downsample is achieved through maxpooling, here one has the option to use either |
| maxpooling or strided convolution for downsampling. The default is to use maxpooling as it is consistent with the |
| original implementation. Note that for upsampling, the authors of VoxelMorph used nearest neighbor interpolation |
| instead of transposed convolution. In this implementation, only nearest neighbor interpolation is supported in order |
| to be consistent with the original implementation. |
| |
| An instance of this class can be used as a backbone network for constructing a VoxelMorph network. See the |
| documentation of :py:class:`monai.networks.nets.VoxelMorph` for more details and an example on how to construct a |
| VoxelMorph network. |
| |
| Args: |
| spatial_dims: number of spatial dimensions. |
| in_channels: number of channels in the input volume after concatenation of moving and fixed images. |
| unet_out_channels: number of channels in the output of the UNet. |
| channels: number of channels in each layer of the UNet. See the following example for more details. |
| final_conv_channels: number of channels in each layer of the final convolution block. |
| final_conv_act: activation type for the final convolution block. Defaults to LeakyReLU. |
| Since VoxelMorph was originally implemented in tensorflow where the default negative slope for |
| LeakyReLU was 0.2, we use the same default value here. |
| kernel_size: kernel size for all convolution layers in the UNet. Defaults to 3. |
| up_kernel_size: kernel size for all convolution layers in the upsampling path of the UNet. Defaults to 3. |
| act: activation type for all convolution layers in the UNet. Defaults to LeakyReLU with negative slope 0.2. |
| norm: feature normalization type and arguments for all convolution layers in the UNet. Defaults to None. |
| dropout: dropout ratio for all convolution layers in the UNet. Defaults to 0.0 (no dropout). |
| bias: whether to use bias in all convolution layers in the UNet. Defaults to True. |
| use_maxpool: whether to use maxpooling in the downsampling path of the UNet. Defaults to True. |
| Using maxpooling is the consistent with the original implementation of VoxelMorph. |
| But one can optionally use strided convolution instead (i.e. set `use_maxpool` to False). |
| adn_ordering: ordering of activation, dropout, and normalization. Defaults to "NDA". |
| """ |
|
|
| def __init__( |
| self, |
| spatial_dims: int, |
| in_channels: int, |
| unet_out_channels: int, |
| channels: Sequence[int], |
| final_conv_channels: Sequence[int], |
| final_conv_act: tuple | str | None = "LEAKYRELU", |
| kernel_size: Sequence[int] | int = 3, |
| up_kernel_size: Sequence[int] | int = 3, |
| act: tuple | str = "LEAKYRELU", |
| norm: tuple | str | None = None, |
| dropout: float = 0.0, |
| bias: bool = True, |
| use_maxpool: bool = True, |
| adn_ordering: str = "NDA", |
| ) -> None: |
| super().__init__() |
|
|
| if spatial_dims not in (2, 3): |
| raise ValueError("spatial_dims must be either 2 or 3.") |
| if in_channels % 2 != 0: |
| raise ValueError("in_channels must be divisible by 2.") |
| if len(channels) < 2: |
| raise ValueError("the length of `channels` should be no less than 2.") |
| if len(channels) % 2 != 0: |
| raise ValueError("the elements of `channels` should be specified in pairs.") |
| if isinstance(kernel_size, Sequence) and len(kernel_size) != spatial_dims: |
| raise ValueError("the length of `kernel_size` should equal to `dimensions`.") |
| if isinstance(up_kernel_size, Sequence) and len(up_kernel_size) != spatial_dims: |
| raise ValueError("the length of `up_kernel_size` should equal to `dimensions`.") |
|
|
| |
| self.dimensions = spatial_dims |
| self.in_channels = in_channels |
| self.unet_out_channels = unet_out_channels |
| self.channels = channels |
| self.kernel_size = kernel_size |
| self.up_kernel_size = up_kernel_size |
| self.act = ( |
| ("leakyrelu", {"negative_slope": 0.2, "inplace": True}) |
| if isinstance(act, str) and act.upper() == "LEAKYRELU" |
| else act |
| ) |
| self.norm = norm |
| self.dropout = dropout |
| self.bias = bias |
| self.use_maxpool = use_maxpool |
| self.adn_ordering = adn_ordering |
|
|
| |
| self.final_conv_channels = final_conv_channels |
| self.final_conv_act = ( |
| ("leakyrelu", {"negative_slope": 0.2, "inplace": True}) |
| if isinstance(final_conv_act, str) and final_conv_act.upper() == "LEAKYRELU" |
| else final_conv_act |
| ) |
|
|
| def _create_block(inc: int, outc: int, channels: Sequence[int], is_top: bool) -> nn.Module: |
| """ |
| Builds the UNet structure recursively. |
| |
| Args: |
| inc: number of input channels. |
| outc: number of output channels. |
| channels: sequence of channels for each pair of down and up layers. |
| is_top: True if this is the top block. |
| """ |
|
|
| next_c_in, next_c_out = channels[0:2] |
| upc = next_c_in + next_c_out |
|
|
| subblock: nn.Module |
|
|
| if len(channels) > 2: |
| subblock = _create_block(next_c_in, next_c_out, channels[2:], is_top=False) |
| else: |
| |
| subblock = self._get_bottom_layer(next_c_in, next_c_out) |
|
|
| down = self._get_down_layer(inc, next_c_in, is_top) |
| up = self._get_up_layer(upc, outc, is_top) |
|
|
| return self._get_connection_block(down, up, subblock) |
|
|
| def _create_final_conv(inc: int, outc: int, channels: Sequence[int]) -> nn.Module: |
| """ |
| Builds the final convolution blocks. |
| |
| Args: |
| inc: number of input channels, should be the same as `unet_out_channels`. |
| outc: number of output channels, should be the same as `spatial_dims`. |
| channels: sequence of channels for each convolution layer. |
| |
| Note: there is no activation after the last convolution layer as per the original implementation. |
| """ |
|
|
| mod: nn.Module = nn.Sequential() |
|
|
| for i, c in enumerate(channels): |
| mod.add_module( |
| f"final_conv_{i}", |
| Convolution( |
| self.dimensions, |
| inc, |
| c, |
| kernel_size=self.kernel_size, |
| act=self.final_conv_act, |
| norm=self.norm, |
| dropout=self.dropout, |
| bias=self.bias, |
| adn_ordering=self.adn_ordering, |
| ), |
| ) |
| inc = c |
|
|
| mod.add_module( |
| "final_conv_out", |
| Convolution( |
| self.dimensions, |
| inc, |
| outc, |
| kernel_size=self.kernel_size, |
| act=None, |
| norm=self.norm, |
| dropout=self.dropout, |
| bias=self.bias, |
| adn_ordering=self.adn_ordering, |
| ), |
| ) |
|
|
| return mod |
|
|
| self.net = nn.Sequential( |
| _create_block(in_channels, unet_out_channels, self.channels, is_top=True), |
| _create_final_conv(unet_out_channels, self.dimensions, self.final_conv_channels), |
| ) |
|
|
| def _get_connection_block(self, down_path: nn.Module, up_path: nn.Module, subblock: nn.Module) -> nn.Module: |
| """ |
| Returns the block object defining a layer of the UNet structure including the implementation of the skip |
| between encoding (down) and decoding (up) sides of the network. |
| |
| Args: |
| down_path: encoding half of the layer |
| up_path: decoding half of the layer |
| subblock: block defining the next layer in the network. |
| |
| Returns: block for this layer: `nn.Sequential(down_path, SkipConnection(subblock), up_path)` |
| """ |
|
|
| return nn.Sequential(down_path, SkipConnection(subblock), up_path) |
|
|
| def _get_down_layer(self, in_channels: int, out_channels: int, is_top: bool) -> nn.Module: |
| """ |
| In each down layer, the input is first downsampled using maxpooling, |
| then passed through a convolution block, unless this is the top layer |
| in which case the input is passed through a convolution block only |
| without maxpooling first. |
| |
| Args: |
| in_channels: number of input channels. |
| out_channels: number of output channels. |
| is_top: True if this is the top block. |
| """ |
|
|
| mod: Convolution | nn.Sequential |
|
|
| strides = 1 if self.use_maxpool or is_top else 2 |
|
|
| mod = Convolution( |
| self.dimensions, |
| in_channels, |
| out_channels, |
| strides=strides, |
| kernel_size=self.kernel_size, |
| act=self.act, |
| norm=self.norm, |
| dropout=self.dropout, |
| bias=self.bias, |
| adn_ordering=self.adn_ordering, |
| ) |
|
|
| if self.use_maxpool and not is_top: |
| mod = ( |
| nn.Sequential(nn.MaxPool3d(kernel_size=2, stride=2), mod) |
| if self.dimensions == 3 |
| else nn.Sequential(nn.MaxPool2d(kernel_size=2, stride=2), mod) |
| ) |
|
|
| return mod |
|
|
| def _get_bottom_layer(self, in_channels: int, out_channels: int) -> nn.Module: |
| """ |
| Bottom layer (bottleneck) in voxelmorph consists of a typical down layer followed by an upsample layer. |
| |
| Args: |
| in_channels: number of input channels. |
| out_channels: number of output channels. |
| """ |
|
|
| mod: nn.Module |
| upsample: nn.Module |
|
|
| mod = self._get_down_layer(in_channels, out_channels, is_top=False) |
|
|
| upsample = UpSample( |
| self.dimensions, |
| out_channels, |
| out_channels, |
| scale_factor=2, |
| mode="nontrainable", |
| interp_mode="nearest", |
| align_corners=None, |
| ) |
|
|
| return nn.Sequential(mod, upsample) |
|
|
| def _get_up_layer(self, in_channels: int, out_channels: int, is_top: bool) -> nn.Module: |
| """ |
| In each up layer, the input is passed through a convolution block before upsampled, |
| unless this is the top layer in which case the input is passed through a convolution block only |
| without upsampling. |
| |
| Args: |
| in_channels: number of input channels. |
| out_channels: number of output channels. |
| is_top: True if this is the top block. |
| """ |
|
|
| mod: Convolution | nn.Sequential |
|
|
| strides = 1 |
|
|
| mod = Convolution( |
| self.dimensions, |
| in_channels, |
| out_channels, |
| strides=strides, |
| kernel_size=self.up_kernel_size, |
| act=self.act, |
| norm=self.norm, |
| dropout=self.dropout, |
| bias=self.bias, |
| |
| is_transposed=False, |
| adn_ordering=self.adn_ordering, |
| ) |
|
|
| if not is_top: |
| mod = nn.Sequential( |
| mod, |
| UpSample( |
| self.dimensions, |
| out_channels, |
| out_channels, |
| scale_factor=2, |
| mode="nontrainable", |
| interp_mode="nearest", |
| align_corners=None, |
| ), |
| ) |
|
|
| return mod |
|
|
| def forward(self, concatenated_pairs: torch.Tensor) -> torch.Tensor: |
| x = self.net(concatenated_pairs) |
| return x |
|
|
|
|
| voxelmorphunet = VoxelMorphUNet |
|
|
|
|
| @export("monai.networks.nets") |
| @alias("voxelmorph") |
| class VoxelMorph(nn.Module): |
| """ |
| A re-implementation of VoxelMorph framework for medical image registration as described in |
| https://arxiv.org/pdf/1809.05231.pdf. For more details, please refer to VoxelMorph: A Learning Framework for |
| Deformable Medical Image Registration, Guha Balakrishnan, Amy Zhao, Mert R. Sabuncu, John Guttag, Adrian V. Dalca |
| IEEE TMI: Transactions on Medical Imaging. 2019. eprint arXiv:1809.05231. |
| |
| This class is intended to be a general framework, based on which a deformable image registration |
| network can be built. Given a user-specified backbone network (e.g., UNet in the original VoxelMorph paper), this |
| class serves as a wrapper that concatenates the input pair of moving and fixed images, passes through the backbone |
| network, integrate the predicted stationary velocity field (DVF) from the backbone network to obtain the |
| displacement field (DDF), and, finally, warp the moving image using the DDF. |
| |
| To construct a VoxelMorph network, one need to first construct a backbone network |
| (e.g., a :py:class:`monai.networks.nets.VoxelMorphUNet`) and pass it to the constructor of |
| :py:class:`monai.networks.nets.VoxelMorph`. The backbone network should be able to take a pair of moving and fixed |
| images as input and produce a DVF (or DDF, details to be discussed later) as output. |
| |
| When `forward` is called, the input moving and fixed images are first concatenated along the channel dimension and |
| passed through the specified backbone network to produce the prediction of the displacement field (DDF) in the |
| non-diffeomorphic variant (i.e. when `integration_steps` is set to 0) or the stationary velocity field (DVF) in the |
| diffeomorphic variant (i.e. when `integration_steps` is set to a positive integer). The DVF is then integrated using |
| a scaling-and-squaring approach via a :py:class:`monai.networks.blocks.warp.DVF2DDF` module to produce the DDF. |
| Finally, the DDF is used to warp the moving image to the fixed image using a |
| :py:class:`monai.networks.blocks.warp.Warp` module. Optionally, the integration from DVF to DDF can be |
| performed on reduced resolution by specifying `half_res` to be True, in which case the output DVF from the backbone |
| network is first linearly interpolated to half resolution before integration. The output DDF is then linearly |
| interpolated again back to full resolution before being used to warp the moving image. |
| |
| Args: |
| backbone: a backbone network. |
| integration_steps: number of integration steps used for obtaining DDF from DVF via scaling-and-squaring. |
| Defaults to 7. If set to 0, the network will be non-diffeomorphic. |
| half_res: whether to perform integration on half resolution. Defaults to False. |
| spatial_dims: number of spatial dimensions, defaults to 3. |
| |
| Example:: |
| |
| from monai.networks.nets import VoxelMorphUNet, VoxelMorph |
| |
| # The following example construct an instance of VoxelMorph that matches the original VoxelMorph paper |
| # https://arxiv.org/pdf/1809.05231.pdf |
| |
| # First, a backbone network is constructed. In this case, we use a VoxelMorphUNet as the backbone network. |
| backbone = VoxelMorphUNet( |
| spatial_dims=3, |
| in_channels=2, |
| unet_out_channels=32, |
| channels=(16, 32, 32, 32, 32, 32), # this indicates the down block at the top takes 16 channels as |
| # input, the corresponding up block at the top produces 32 |
| # channels as output, the second down block takes 32 channels as |
| # input, and the corresponding up block at the same level |
| # produces 32 channels as output, etc. |
| final_conv_channels=(16, 16) |
| ) |
| |
| # Then, a full VoxelMorph network is constructed using the specified backbone network. |
| net = VoxelMorph( |
| backbone=backbone, |
| integration_steps=7, |
| half_res=False |
| ) |
| |
| # A forward pass through the network would look something like this |
| moving = torch.randn(1, 1, 160, 192, 224) |
| fixed = torch.randn(1, 1, 160, 192, 224) |
| warped, ddf = net(moving, fixed) |
| |
| """ |
|
|
| def __init__( |
| self, |
| backbone: VoxelMorphUNet | nn.Module | None = None, |
| integration_steps: int = 7, |
| half_res: bool = False, |
| spatial_dims: int = 3, |
| ) -> None: |
| super().__init__() |
|
|
| |
| self.backbone = ( |
| backbone |
| if backbone is not None |
| else VoxelMorphUNet( |
| spatial_dims=spatial_dims, |
| in_channels=2, |
| unet_out_channels=32, |
| channels=(16, 32, 32, 32, 32, 32), |
| final_conv_channels=(16, 16), |
| ) |
| ) |
|
|
| |
| self.spatial_dims = spatial_dims |
| self.half_res = half_res |
| self.integration_steps = integration_steps |
| self.diffeomorphic = True if self.integration_steps > 0 else False |
|
|
| |
| if self.diffeomorphic: |
| self.dvf2ddf = DVF2DDF(num_steps=self.integration_steps, mode="bilinear", padding_mode="zeros") |
| self.warp = Warp(mode="bilinear", padding_mode="zeros") |
|
|
| def forward(self, moving: torch.Tensor, fixed: torch.Tensor) -> tuple[torch.Tensor, torch.Tensor]: |
| if moving.shape != fixed.shape: |
| raise ValueError( |
| "The spatial shape of the moving image should be the same as the spatial shape of the fixed image." |
| f" Got {moving.shape} and {fixed.shape} instead." |
| ) |
|
|
| x = self.backbone(torch.cat([moving, fixed], dim=1)) |
|
|
| if x.shape[1] != self.spatial_dims: |
| raise ValueError( |
| "The number of channels in the output of the backbone network should be equal to the" |
| f" number of spatial dimensions {self.spatial_dims}. Got {x.shape[1]} channels instead." |
| ) |
|
|
| if x.shape[2:] != moving.shape[2:]: |
| raise ValueError( |
| "The spatial shape of the output of the backbone network should be equal to the" |
| f" spatial shape of the input images. Got {x.shape[2:]} instead of {moving.shape[2:]}." |
| ) |
|
|
| if self.half_res: |
| x = F.interpolate(x, scale_factor=0.5, mode="trilinear", align_corners=True) * 2.0 |
|
|
| if self.diffeomorphic: |
| x = self.dvf2ddf(x) |
|
|
| if self.half_res: |
| x = F.interpolate(x * 0.5, scale_factor=2.0, mode="trilinear", align_corners=True) |
|
|
| return self.warp(moving, x), x |
|
|
|
|
| voxelmorph = VoxelMorph |
|
|