| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| from abc import ABC, abstractmethod |
| from collections.abc import Callable, Sequence |
| from typing import TYPE_CHECKING, Any, cast |
|
|
| import torch |
|
|
| from monai.config import IgniteInfo |
| from monai.transforms import apply_transform |
| from monai.utils import ensure_tuple, min_version, optional_import |
| from monai.utils.enums import CommonKeys, GanKeys |
|
|
| if TYPE_CHECKING: |
| from ignite.engine import EventEnum |
| else: |
| EventEnum, _ = optional_import( |
| "ignite.engine", IgniteInfo.OPT_IMPORT_VERSION, min_version, "EventEnum", as_type="base" |
| ) |
|
|
| __all__ = [ |
| "IterationEvents", |
| "get_devices_spec", |
| "default_prepare_batch", |
| "PrepareBatch", |
| "PrepareBatchDefault", |
| "PrepareBatchExtraInput", |
| "default_make_latent", |
| "engine_apply_transform", |
| "default_metric_cmp_fn", |
| ] |
|
|
|
|
| class IterationEvents(EventEnum): |
| """ |
| Additional Events engine can register and trigger in the iteration process. |
| Refer to the example in ignite: https://pytorch.org/ignite/generated/ignite.engine.events.EventEnum.html. |
| These Events can be triggered during training iteration: |
| `FORWARD_COMPLETED` is the Event when `network(image, label)` completed. |
| `LOSS_COMPLETED` is the Event when `loss(pred, label)` completed. |
| `BACKWARD_COMPLETED` is the Event when `loss.backward()` completed. |
| `MODEL_COMPLETED` is the Event when all the model related operations completed. |
| `INNER_ITERATION_STARTED` is the Event when the iteration has an inner loop and the loop is started. |
| `INNER_ITERATION_COMPLETED` is the Event when the iteration has an inner loop and the loop is completed. |
| """ |
|
|
| FORWARD_COMPLETED = "forward_completed" |
| LOSS_COMPLETED = "loss_completed" |
| BACKWARD_COMPLETED = "backward_completed" |
| MODEL_COMPLETED = "model_completed" |
| INNER_ITERATION_STARTED = "inner_iteration_started" |
| INNER_ITERATION_COMPLETED = "inner_iteration_completed" |
|
|
|
|
| def get_devices_spec(devices: Sequence[torch.device | str] | None = None) -> list[torch.device]: |
| """ |
| Get a valid specification for one or more devices. If `devices` is None get devices for all CUDA devices available. |
| If `devices` is and zero-length structure a single CPU compute device is returned. In any other cases `devices` is |
| returned unchanged. |
| |
| Args: |
| devices: list of devices to request, None for all GPU devices, [] for CPU. |
| |
| Raises: |
| RuntimeError: When all GPUs are selected (``devices=None``) but no GPUs are available. |
| |
| Returns: |
| list of torch.device: list of devices. |
| |
| """ |
| if devices is None: |
| devices = [torch.device(f"cuda:{d:d}") for d in range(torch.cuda.device_count())] |
|
|
| if not devices: |
| raise RuntimeError("No GPU devices available.") |
|
|
| elif len(devices) == 0: |
| devices = [torch.device("cpu")] |
|
|
| else: |
| devices = list(devices) |
|
|
| devices = [torch.device(d) if isinstance(d, str) else d for d in devices] |
| return devices |
|
|
|
|
| def default_prepare_batch( |
| batchdata: dict[str, torch.Tensor] | torch.Tensor | Sequence[torch.Tensor], |
| device: str | torch.device | None = None, |
| non_blocking: bool = False, |
| **kwargs: Any, |
| ) -> tuple[torch.Tensor, torch.Tensor | None] | torch.Tensor: |
| """ |
| Default function to prepare the data for current iteration. |
| |
| The input `batchdata` is either a single tensor, a pair of tensors, or a dictionary of data. In the first case the |
| return value is the tensor and None, in the second case the return value is the two tensors, and in the dictionary |
| case the return value depends on what keys are present. if `CommonKeys.IMAGE` and `CommonKeys.LABEL` are present |
| then the tensors they key to are returned, if only `CommonKeys.IMAGE` is present that tensor and None is returned. |
| If `CommonKeys.REALS` is present this is returned with None. All returned tensors are moved to the given device |
| using the given non-blocking argument before being returned. |
| |
| This function implements the expected API for a `prepare_batch` callable in Ignite: |
| https://pytorch.org/ignite/v0.4.8/generated/ignite.engine.create_supervised_trainer.html |
| |
| Args: |
| batchdata: input batch data which is either a single tensor, a pair, or a dictionary |
| device: device to move every returned tensor to |
| non_blocking: equivalent argument for `Tensor.to` |
| kwargs: further arguments for `Tensor.to` |
| |
| Returns: |
| image, label(optional). |
| """ |
| if not isinstance(batchdata, dict): |
| if isinstance(batchdata, torch.Tensor): |
| return batchdata.to(device=device, non_blocking=non_blocking, **kwargs), None |
| elif len(batchdata) == 2: |
| image, label = batchdata |
| return ( |
| image.to(device=device, non_blocking=non_blocking, **kwargs), |
| label.to(device=device, non_blocking=non_blocking, **kwargs), |
| ) |
|
|
| raise AssertionError("Default prepare_batch expects a single tensor, a tensor pair, or dictionary input data.") |
|
|
| if isinstance(batchdata.get(CommonKeys.LABEL), torch.Tensor): |
| return ( |
| batchdata[CommonKeys.IMAGE].to(device=device, non_blocking=non_blocking, **kwargs), |
| batchdata[CommonKeys.LABEL].to(device=device, non_blocking=non_blocking, **kwargs), |
| ) |
|
|
| if GanKeys.REALS in batchdata: |
| return batchdata[GanKeys.REALS].to(device=device, non_blocking=non_blocking, **kwargs) |
|
|
| return batchdata[CommonKeys.IMAGE].to(device=device, non_blocking=non_blocking, **kwargs), None |
|
|
|
|
| class PrepareBatch(ABC): |
| """ |
| Interface of customized prepare_batch in the trainer or evaluator workflows. |
| It takes the data of current batch, target device and non_blocking flag as input. |
| Args `batchdata`, `device`, `non_blocking` refer to the ignite API: |
| https://pytorch.org/ignite/v0.4.8/generated/ignite.engine.create_supervised_trainer.html. |
| `kwargs` supports other args for `Tensor.to()` API. |
| """ |
|
|
| @abstractmethod |
| def __call__( |
| self, |
| batchdata: dict[str, torch.Tensor], |
| device: str | torch.device | None = None, |
| non_blocking: bool = False, |
| **kwargs: Any, |
| ) -> Any: |
| raise NotImplementedError(f"Subclass {self.__class__.__name__} must implement this method.") |
|
|
|
|
| class PrepareBatchDefault(PrepareBatch): |
| """ |
| This wraps `default_prepare_batch` to return `image` and `label` only, so is consistent with its API. |
| """ |
|
|
| def __call__( |
| self, |
| batchdata: dict[str, torch.Tensor] | torch.Tensor | Sequence[torch.Tensor], |
| device: str | torch.device | None = None, |
| non_blocking: bool = False, |
| **kwargs: Any, |
| ) -> tuple[torch.Tensor, torch.Tensor | None] | torch.Tensor: |
| """ |
| Args `batchdata`, `device`, `non_blocking` refer to the ignite API: |
| https://pytorch.org/ignite/v0.4.8/generated/ignite.engine.create_supervised_trainer.html. |
| `kwargs` supports other args for `Tensor.to()` API. |
| |
| """ |
| return default_prepare_batch(batchdata, device, non_blocking, **kwargs) |
|
|
|
|
| class PrepareBatchExtraInput(PrepareBatch): |
| """ |
| Customized prepare batch callable for trainers or evaluators which support extra input data for the network. |
| Extra items are specified by the `extra_keys` parameter and are extracted from the input dictionary (ie. the batch). |
| This uses `default_prepare_batch` but requires dictionary inputs. |
| |
| Args: |
| extra_keys: If a string or sequence of strings is provided, values from the input dictionary are extracted from |
| those keys and passed to the network as extra positional arguments. If a dictionary is provided, every pair |
| `(k, v)` in that dictionary will become a new keyword argument assigning to `k` the value in the input |
| dictionary keyed to `v`. |
| """ |
|
|
| def __init__(self, extra_keys: str | Sequence[str] | dict[str, str]) -> None: |
| self.extra_keys = extra_keys |
|
|
| def __call__( |
| self, |
| batchdata: dict[str, torch.Tensor], |
| device: str | torch.device | None = None, |
| non_blocking: bool = False, |
| **kwargs: Any, |
| ) -> tuple[torch.Tensor, torch.Tensor, tuple, dict]: |
| """ |
| Args `batchdata`, `device`, `non_blocking` refer to the ignite API: |
| https://pytorch.org/ignite/v0.4.8/generated/ignite.engine.create_supervised_trainer.html. |
| `kwargs` supports other args for `Tensor.to()` API. |
| """ |
| image, label = default_prepare_batch(batchdata, device, non_blocking, **kwargs) |
| args_ = list() |
| kwargs_ = dict() |
|
|
| def _get_data(key: str) -> torch.Tensor: |
| data = batchdata[key] |
|
|
| if isinstance(data, torch.Tensor): |
| data = data.to(device=device, non_blocking=non_blocking, **kwargs) |
|
|
| return data |
|
|
| if isinstance(self.extra_keys, (str, list, tuple)): |
| for k in ensure_tuple(self.extra_keys): |
| args_.append(_get_data(k)) |
| elif isinstance(self.extra_keys, dict): |
| for k, v in self.extra_keys.items(): |
| kwargs_.update({k: _get_data(v)}) |
|
|
| return cast(torch.Tensor, image), cast(torch.Tensor, label), tuple(args_), kwargs_ |
|
|
|
|
| def default_make_latent( |
| num_latents: int, |
| latent_size: int, |
| device: str | torch.device | None = None, |
| non_blocking: bool = False, |
| **kwargs: Any, |
| ) -> torch.Tensor: |
| return torch.randn(num_latents, latent_size).to(device=device, non_blocking=non_blocking, **kwargs) |
|
|
|
|
| def engine_apply_transform(batch: Any, output: Any, transform: Callable[..., dict]) -> tuple[Any, Any]: |
| """ |
| Apply transform on `batch` and `output`. |
| If `batch` and `output` are dictionaries, temporarily combine them for the transform, |
| otherwise, apply the transform for `output` data only. |
| |
| """ |
| if isinstance(batch, dict) and isinstance(output, dict): |
| data = dict(batch) |
| data.update(output) |
| transformed_data = apply_transform(transform, data) |
|
|
| if not isinstance(transformed_data, dict): |
| raise AssertionError("With a dict supplied to apply_transform a single dict return is expected.") |
|
|
| for k, v in transformed_data.items(): |
| |
| |
| if k in output or k not in batch: |
| output[k] = v |
| else: |
| batch[k] = v |
| else: |
| output = apply_transform(transform, output) |
|
|
| return batch, output |
|
|
|
|
| def default_metric_cmp_fn(current_metric: float, prev_best: float) -> bool: |
| """ |
| The default function to compare metric values between current metric and previous best metric. |
| |
| Args: |
| current_metric: metric value of current round computation. |
| prev_best: the best metric value of previous rounds to compare with. |
| |
| """ |
| return current_metric > prev_best |
|
|