ecom-rlve-code / scripts /make_mini_catalog.py
thebajajra's picture
Upload 12 files
1f3039b verified
#!/usr/bin/env python3
"""Download the first 2M rows of Amazebay-catalog and save locally as a HF dataset.
Usage:
uv run python scripts/make_mini_catalog.py [--max-rows 2000000] [--out data/amazebay-2M]
This streams the dataset so it never loads the full 1.97M+ dataset into memory.
The output is saved in Arrow/Parquet format and can be loaded with:
ds = load_dataset("data/amazebay-2M")
"""
from __future__ import annotations
import argparse
import time
from pathlib import Path
from datasets import Dataset, load_dataset
def main() -> None:
parser = argparse.ArgumentParser(description="Create a mini Amazebay catalog")
parser.add_argument(
"--max-rows", type=int, default=2_000_000,
help="Number of rows to take (default: 2,000,000)",
)
parser.add_argument(
"--out", type=str, default="data/amazebay-2M",
help="Output directory for the saved dataset",
)
parser.add_argument(
"--source", type=str, default="thebajajra/Amazebay-catalog",
help="HuggingFace dataset name",
)
parser.add_argument(
"--batch-size", type=int, default=10_000,
help="Rows to accumulate before printing progress",
)
args = parser.parse_args()
out_path = Path(args.out)
print(f"Streaming {args.source} → first {args.max_rows:,} rows → {out_path}")
t0 = time.time()
ds_stream = load_dataset(args.source, split="train", streaming=True)
# Collect rows in batches for progress reporting
rows: list[dict] = []
for i, row in enumerate(ds_stream):
if i >= args.max_rows:
break
rows.append(row)
if (i + 1) % args.batch_size == 0:
elapsed = time.time() - t0
rate = (i + 1) / elapsed
eta = (args.max_rows - i - 1) / rate if rate > 0 else 0
print(
f" {i + 1:>10,} / {args.max_rows:,} rows "
f"({100 * (i + 1) / args.max_rows:5.1f}%) "
f"{rate:,.0f} rows/s ETA {eta / 60:.1f}min"
)
elapsed = time.time() - t0
print(f"\nCollected {len(rows):,} rows in {elapsed:.1f}s")
# Convert to HF Dataset and save
print("Converting to Arrow dataset...")
ds = Dataset.from_list(rows)
print(f" columns: {ds.column_names}")
print(f" num_rows: {ds.num_rows:,}")
print(f"Saving to {out_path} ...")
out_path.mkdir(parents=True, exist_ok=True)
ds.save_to_disk(str(out_path))
# Also save as parquet for easy inspection
parquet_path = out_path / "catalog.parquet"
ds.to_parquet(str(parquet_path))
print(f"Also saved as parquet: {parquet_path}")
# Quick stats
print("\n--- Quick Stats ---")
cats = {}
valid_prices = 0
for row in rows:
cat = row.get("main_category", "unknown")
cats[cat] = cats.get(cat, 0) + 1
p = row.get("price")
if p and str(p) != "None":
valid_prices += 1
print(f"Total rows: {len(rows):,}")
print(f"Valid prices: {valid_prices:,} ({100 * valid_prices / len(rows):.1f}%)")
print(f"Categories ({len(cats)}):")
for cat, cnt in sorted(cats.items(), key=lambda x: -x[1]):
cat_name = str(cat) if cat is not None else "(None)"
print(f" {cat_name:45s} {cnt:>10,} ({100 * cnt / len(rows):5.1f}%)")
total_sec = time.time() - t0
print(f"\nDone in {total_sec / 60:.1f} min total.")
if __name__ == "__main__":
main()