| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from typing import Callable, Optional |
|
|
| from torch.utils.data import DataLoader as _TorchDataLoader |
| from torch.utils.data import Dataset, Sampler |
|
|
| from monai.data.utils import list_data_collate, worker_init_fn |
|
|
| __all__ = ["DataLoader"] |
|
|
|
|
| class DataLoader(_TorchDataLoader): |
| """Generates images/labels for train/validation/testing from dataset. |
| It inherits from PyTorch DataLoader and adds callbacks for `collate` and `worker_fn`. |
| |
| Args: |
| dataset: dataset from which to load the data. |
| batch_size: how many samples per batch to load |
| (default: ``1``). |
| shuffle: set to ``True`` to have the data reshuffled |
| at every epoch (default: ``False``). |
| sampler: defines the strategy to draw samples from |
| the dataset. If specified, :attr:`shuffle` must be ``False``. |
| batch_sampler: like :attr:`sampler`, but returns a batch of |
| indices at a time. Mutually exclusive with :attr:`batch_size`, |
| :attr:`shuffle`, :attr:`sampler`, and :attr:`drop_last`. |
| num_workers: how many subprocesses to use for data |
| loading. ``0`` means that the data will be loaded in the main process. |
| (default: ``0``) |
| pin_memory: If ``True``, the data loader will copy Tensors |
| into CUDA pinned memory before returning them. If your data elements |
| are a custom type, or your :attr:`collate_fn` returns a batch that is a custom type, |
| see the example below. |
| drop_last: set to ``True`` to drop the last incomplete batch, |
| if the dataset size is not divisible by the batch size. If ``False`` and |
| the size of dataset is not divisible by the batch size, then the last batch |
| will be smaller. (default: ``False``) |
| timeout: if positive, the timeout value for collecting a batch |
| from workers. Should always be non-negative. (default: ``0``) |
| multiprocessing_context: specify a valid start method for multi-processing. |
| |
| """ |
|
|
| def __init__( |
| self, |
| dataset: Dataset, |
| batch_size: int = 1, |
| shuffle: bool = False, |
| sampler: Optional[Sampler] = None, |
| batch_sampler: Optional[Sampler] = None, |
| num_workers: int = 0, |
| pin_memory: bool = False, |
| drop_last: bool = False, |
| timeout: float = 0.0, |
| multiprocessing_context: Optional[Callable] = None, |
| ) -> None: |
| super().__init__( |
| dataset=dataset, |
| batch_size=batch_size, |
| shuffle=shuffle, |
| sampler=sampler, |
| batch_sampler=batch_sampler, |
| num_workers=num_workers, |
| collate_fn=list_data_collate, |
| pin_memory=pin_memory, |
| drop_last=drop_last, |
| timeout=timeout, |
| worker_init_fn=worker_init_fn, |
| multiprocessing_context=multiprocessing_context, |
| ) |
|
|