| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| """This dataset contains interior images created using DALL-E. |
| The dataset contains 512x512 images split into 5 classes: |
| * bathroom: 1000 images |
| * bedroom: 1000 images |
| * dining_room: 1000 images |
| * kitchen: 1000 images |
| * living_room: 1000 images |
| """ |
|
|
| import datasets |
| from datasets.download.download_manager import DownloadManager |
| from datasets.tasks import ImageClassification |
| from pathlib import Path |
| from typing import List, Iterator |
|
|
| _ALLOWED_IMG_EXT = {".png", ".jpg"} |
|
|
| _CITATION = """\ |
| @InProceedings{huggingface:dataset, |
| title = {Computer Generated interior images}, |
| author={Padilla, Rafael}, |
| year={2023} |
| } |
| """ |
|
|
| _DESCRIPTION = """\ |
| This new dataset contains CG interior images representing interior of houses in 5 classes, with \ |
| 1000 images per class. |
| """ |
|
|
| _HOMEPAGE = "https://huggingface.co/datasets/rafaelpadilla/interior-cgi/" |
|
|
| _LICENSE = "" |
|
|
| _URLS = { |
| "test": "https://huggingface.co/datasets/rafaelpadilla/interior-cgi/resolve/main/data/test.zip", |
| "train": "https://huggingface.co/datasets/rafaelpadilla/interior-cgi/resolve/main/data/train.zip", |
| } |
|
|
| _NAMES = ["bathroom", "bedroom", "dining_room", "kitchen", "living_room"] |
|
|
|
|
| class CGInteriorDataset(datasets.GeneratorBasedBuilder): |
| """CGInterior: Computer Generated Interior images dataset""" |
|
|
| VERSION = datasets.Version("1.1.0") |
|
|
| def _info(self) -> datasets.DatasetInfo: |
| """ |
| Returns the dataset metadata and features. |
| |
| Returns: |
| DatasetInfo: Metadata and features of the dataset. |
| """ |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features( |
| { |
| "image": datasets.Image(), |
| "label_id": datasets.features.ClassLabel(names=_NAMES), |
| "label_name": datasets.Value("string"), |
| } |
| ), |
| supervised_keys=("image", "label_id"), |
| homepage=_HOMEPAGE, |
| license=_LICENSE, |
| citation=_CITATION, |
| task_templates=[ |
| ImageClassification(image_column="image", label_column="label_id") |
| ], |
| ) |
|
|
| def _split_generators( |
| self, dl_manager: DownloadManager |
| ) -> List[datasets.SplitGenerator]: |
| """ |
| Provides the split information and downloads the data. |
| |
| Args: |
| dl_manager (DownloadManager): The DownloadManager to use for downloading and extracting data. |
| |
| Returns: |
| List[SplitGenerator]: List of SplitGenerator objects representing the data splits. |
| """ |
| data_files = dl_manager.download_and_extract(_URLS) |
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={ |
| "files": dl_manager.iter_files(data_files["train"]), |
| }, |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.TEST, |
| gen_kwargs={ |
| "files": dl_manager.iter_files(data_files["test"]), |
| }, |
| ), |
| ] |
|
|
| def _generate_examples(self, files: List[str]) -> Iterator: |
| """ |
| Generates examples for the dataset. |
| |
| Args: |
| files (List[str]): List of image paths. |
| |
| Yields: |
| Dict[str, Union[str, Image]]: A dictionary containing the generated examples. |
| """ |
| for idx, img_path in enumerate(files): |
| path = Path(img_path) |
| if path.suffix in _ALLOWED_IMG_EXT: |
| yield idx, { |
| "image": img_path, |
| "label_id": path.parent.name, |
| "label_name": path.parent.name, |
| } |
|
|