| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| from collections.abc import Sequence |
|
|
| import torch |
| import torch.nn as nn |
|
|
| from monai.networks.layers.factories import Act, Norm, split_args |
| from monai.networks.nets.regressor import Regressor |
|
|
| __all__ = ["Classifier", "Discriminator", "Critic"] |
|
|
|
|
| class Classifier(Regressor): |
| """ |
| Defines a classification network from Regressor by specifying the output shape as a single dimensional tensor |
| with size equal to the number of classes to predict. The final activation function can also be specified, eg. |
| softmax or sigmoid. |
| |
| Args: |
| in_shape: tuple of integers stating the dimension of the input tensor (minus batch dimension) |
| classes: integer stating the dimension of the final output tensor |
| channels: tuple of integers stating the output channels of each convolutional layer |
| strides: tuple of integers stating the stride (downscale factor) of each convolutional layer |
| kernel_size: integer or tuple of integers stating size of convolutional kernels |
| num_res_units: integer stating number of convolutions in residual units, 0 means no residual units |
| act: name or type defining activation layers |
| norm: name or type defining normalization layers |
| dropout: optional float value in range [0, 1] stating dropout probability for layers, None for no dropout |
| bias: boolean stating if convolution layers should have a bias component |
| last_act: name defining the last activation layer |
| """ |
|
|
| def __init__( |
| self, |
| in_shape: Sequence[int], |
| classes: int, |
| channels: Sequence[int], |
| strides: Sequence[int], |
| kernel_size: Sequence[int] | int = 3, |
| num_res_units: int = 2, |
| act=Act.PRELU, |
| norm=Norm.INSTANCE, |
| dropout: float | None = None, |
| bias: bool = True, |
| last_act: str | None = None, |
| ) -> None: |
| super().__init__(in_shape, (classes,), channels, strides, kernel_size, num_res_units, act, norm, dropout, bias) |
|
|
| if last_act is not None: |
| last_act_name, last_act_args = split_args(last_act) |
| last_act_type = Act[last_act_name] |
|
|
| self.final.add_module("lastact", last_act_type(**last_act_args)) |
|
|
|
|
| class Discriminator(Classifier): |
| """ |
| Defines a discriminator network from Classifier with a single output value and sigmoid activation by default. This |
| is meant for use with GANs or other applications requiring a generic discriminator network. |
| |
| Args: |
| in_shape: tuple of integers stating the dimension of the input tensor (minus batch dimension) |
| channels: tuple of integers stating the output channels of each convolutional layer |
| strides: tuple of integers stating the stride (downscale factor) of each convolutional layer |
| kernel_size: integer or tuple of integers stating size of convolutional kernels |
| num_res_units: integer stating number of convolutions in residual units, 0 means no residual units |
| act: name or type defining activation layers |
| norm: name or type defining normalization layers |
| dropout: optional float value in range [0, 1] stating dropout probability for layers, None for no dropout |
| bias: boolean stating if convolution layers should have a bias component |
| last_act: name defining the last activation layer |
| """ |
|
|
| def __init__( |
| self, |
| in_shape: Sequence[int], |
| channels: Sequence[int], |
| strides: Sequence[int], |
| kernel_size: Sequence[int] | int = 3, |
| num_res_units: int = 2, |
| act=Act.PRELU, |
| norm=Norm.INSTANCE, |
| dropout: float | None = 0.25, |
| bias: bool = True, |
| last_act=Act.SIGMOID, |
| ) -> None: |
| super().__init__(in_shape, 1, channels, strides, kernel_size, num_res_units, act, norm, dropout, bias, last_act) |
|
|
|
|
| class Critic(Classifier): |
| """ |
| Defines a critic network from Classifier with a single output value and no final activation. The final layer is |
| `nn.Flatten` instead of `nn.Linear`, the final result is computed as the mean over the first dimension. This is |
| meant to be used with Wasserstein GANs. |
| |
| Args: |
| in_shape: tuple of integers stating the dimension of the input tensor (minus batch dimension) |
| channels: tuple of integers stating the output channels of each convolutional layer |
| strides: tuple of integers stating the stride (downscale factor) of each convolutional layer |
| kernel_size: integer or tuple of integers stating size of convolutional kernels |
| num_res_units: integer stating number of convolutions in residual units, 0 means no residual units |
| act: name or type defining activation layers |
| norm: name or type defining normalization layers |
| dropout: optional float value in range [0, 1] stating dropout probability for layers, None for no dropout |
| bias: boolean stating if convolution layers should have a bias component |
| """ |
|
|
| def __init__( |
| self, |
| in_shape: Sequence[int], |
| channels: Sequence[int], |
| strides: Sequence[int], |
| kernel_size: Sequence[int] | int = 3, |
| num_res_units: int = 2, |
| act=Act.PRELU, |
| norm=Norm.INSTANCE, |
| dropout: float | None = 0.25, |
| bias: bool = True, |
| ) -> None: |
| super().__init__(in_shape, 1, channels, strides, kernel_size, num_res_units, act, norm, dropout, bias, None) |
|
|
| def _get_final_layer(self, in_shape: Sequence[int]): |
| return nn.Flatten() |
|
|
| def forward(self, x: torch.Tensor) -> torch.Tensor: |
| x = self.net(x) |
| x = self.final(x) |
| x = x.mean(1) |
| return x.view((x.shape[0], -1)) |
|
|