jaguar-hyperview-multigeometry / scripts /publish_hyperview_demo_dataset.py
andandandand's picture
Initialize jaguar HyperView multigeometry Space
974f31e verified
#!/usr/bin/env python3
"""Create and optionally publish a resized HF dataset for the HyperView Space."""
from __future__ import annotations
import argparse
import json
import os
from datetime import datetime, timezone
from pathlib import Path
import numpy as np
import pandas as pd
from PIL import Image
from datasets import Dataset, Image as HFImage
PROJECT_ROOT = Path(__file__).resolve().parents[2]
DEFAULT_DATASET_ROOT = PROJECT_ROOT / "kaggle_jaguar_dataset_v2"
DEFAULT_CORESET_CSV = PROJECT_ROOT / "data/validation_coreset.csv"
DEFAULT_OUTPUT_DIR = PROJECT_ROOT / "HyperViewDemoHuggingFaceSpace/dataset_build"
DEFAULT_REPO_ID = os.environ.get("HF_DATASET_REPO", "hyper3labs/jaguar-hyperview-demo")
def utc_now() -> str:
return datetime.now(timezone.utc).strftime("%Y-%m-%dT%H:%M:%SZ")
def parse_args() -> argparse.Namespace:
parser = argparse.ArgumentParser(
description="Build resized train+validation demo dataset with split tags for HyperView."
)
parser.add_argument("--dataset_root", type=Path, default=DEFAULT_DATASET_ROOT)
parser.add_argument("--coreset_csv", type=Path, default=DEFAULT_CORESET_CSV)
parser.add_argument("--output_dir", type=Path, default=DEFAULT_OUTPUT_DIR)
parser.add_argument("--repo_id", type=str, default=DEFAULT_REPO_ID)
parser.add_argument("--config_name", type=str, default="default")
parser.add_argument("--image_size", type=int, default=384)
parser.add_argument("--jpeg_quality", type=int, default=90)
parser.add_argument(
"--image_variant",
type=str,
default="foreground_only",
choices=["foreground_only", "full_rgb"],
)
parser.add_argument("--max_samples", type=int, default=None)
parser.add_argument("--private", action="store_true")
parser.add_argument("--hf_token_env", type=str, default="HF_TOKEN")
parser.add_argument("--no_push", action="store_true")
return parser.parse_args()
def load_rows(dataset_root: Path, coreset_csv: Path, max_samples: int | None) -> pd.DataFrame:
train_csv = dataset_root / "train.csv"
if not train_csv.exists():
raise FileNotFoundError(f"Missing train.csv at {train_csv}")
train_df = pd.read_csv(train_csv)
coreset_df = pd.read_csv(coreset_csv)
coreset_filenames = set(coreset_df["filename"].astype(str).tolist())
train_df = train_df.copy()
train_df["filename"] = train_df["filename"].astype(str)
train_df["label"] = train_df["ground_truth"].astype(str)
train_df["split_tag"] = np.where(train_df["filename"].isin(coreset_filenames), "validation", "train")
train_df["sample_id"] = train_df["filename"]
if max_samples is not None:
train_df = train_df.iloc[: int(max_samples)].copy()
return train_df[["filename", "label", "split_tag", "sample_id"]]
def load_variant_image(image_path: Path, image_variant: str) -> Image.Image:
if image_variant == "foreground_only":
rgba = Image.open(image_path).convert("RGBA")
rgba_np = np.array(rgba, dtype=np.uint8)
rgb = rgba_np[:, :, :3]
alpha = rgba_np[:, :, 3]
mask = (alpha > 0).astype(np.uint8)
cutout_rgb = (rgb * mask[:, :, np.newaxis]).astype(np.uint8)
return Image.fromarray(cutout_rgb, mode="RGB")
return Image.open(image_path).convert("RGB")
def build_resized_images(
rows_df: pd.DataFrame,
dataset_root: Path,
output_images_dir: Path,
image_size: int,
jpeg_quality: int,
image_variant: str,
) -> pd.DataFrame:
source_images_dir = dataset_root / "train"
if not source_images_dir.exists():
raise FileNotFoundError(f"Missing image directory: {source_images_dir}")
output_images_dir.mkdir(parents=True, exist_ok=True)
records: list[dict[str, str]] = []
for _, row in rows_df.iterrows():
filename = str(row["filename"])
src = source_images_dir / filename
if not src.exists():
raise FileNotFoundError(f"Missing source image: {src}")
image = load_variant_image(src, image_variant=image_variant)
image = image.resize((int(image_size), int(image_size)), Image.Resampling.BICUBIC)
dst_name = f"{Path(filename).stem}.jpg"
dst = output_images_dir / dst_name
image.save(dst, format="JPEG", quality=int(jpeg_quality), optimize=True)
records.append(
{
"image": str(dst),
"label": str(row["label"]),
"filename": filename,
"split_tag": str(row["split_tag"]),
"sample_id": str(row["sample_id"]),
}
)
return pd.DataFrame(records)
def build_hf_dataset(records_df: pd.DataFrame) -> Dataset:
payload = {
"image": records_df["image"].tolist(),
"label": records_df["label"].tolist(),
"filename": records_df["filename"].tolist(),
"split_tag": records_df["split_tag"].tolist(),
"sample_id": records_df["sample_id"].tolist(),
}
dataset = Dataset.from_dict(payload)
dataset = dataset.cast_column("image", HFImage())
return dataset
def maybe_push_to_hub(
dataset: Dataset,
repo_id: str,
config_name: str,
private: bool,
hf_token_env: str,
no_push: bool,
) -> str:
if no_push:
return "skipped (--no_push)"
token = os.environ.get(hf_token_env)
if not token:
raise RuntimeError(
f"Missing Hugging Face token in environment variable {hf_token_env}."
)
dataset.push_to_hub(
repo_id=repo_id,
config_name=config_name,
token=token,
private=bool(private),
)
return f"pushed:{repo_id}:{config_name}"
def main() -> int:
args = parse_args()
output_dir = args.output_dir.resolve()
images_out = output_dir / "images"
dataset_out = output_dir / "hf_dataset"
output_dir.mkdir(parents=True, exist_ok=True)
rows_df = load_rows(
dataset_root=args.dataset_root.resolve(),
coreset_csv=args.coreset_csv.resolve(),
max_samples=args.max_samples,
)
if rows_df.empty:
raise RuntimeError("No dataset rows found for publish pipeline.")
records_df = build_resized_images(
rows_df=rows_df,
dataset_root=args.dataset_root.resolve(),
output_images_dir=images_out,
image_size=int(args.image_size),
jpeg_quality=int(args.jpeg_quality),
image_variant=args.image_variant,
)
dataset = build_hf_dataset(records_df)
dataset.save_to_disk(str(dataset_out))
publish_status = maybe_push_to_hub(
dataset=dataset,
repo_id=args.repo_id,
config_name=args.config_name,
private=args.private,
hf_token_env=args.hf_token_env,
no_push=args.no_push,
)
metadata = {
"generated_at_utc": utc_now(),
"dataset_root": str(args.dataset_root.resolve()),
"coreset_csv": str(args.coreset_csv.resolve()),
"output_dir": str(output_dir),
"repo_id": args.repo_id,
"config_name": args.config_name,
"image_size": int(args.image_size),
"jpeg_quality": int(args.jpeg_quality),
"image_variant": args.image_variant,
"num_rows": int(len(records_df)),
"split_counts": records_df["split_tag"].value_counts().to_dict(),
"push_status": publish_status,
}
metadata_path = output_dir / "publish_metadata.json"
metadata_path.write_text(json.dumps(metadata, indent=2), encoding="utf-8")
print("=== HyperView demo dataset pipeline complete ===")
print(f"Rows: {len(records_df)}")
print(f"HF dataset saved to: {dataset_out}")
print(f"Push status: {publish_status}")
print(f"Metadata: {metadata_path}")
return 0
if __name__ == "__main__":
raise SystemExit(main())