| |
| """ |
| Repository model inspector. |
| |
| This script is designed to work in `--config-only` mode without importing |
| PyTorch/Diffusers/Transformers. It reads JSON configs from a local Diffusers |
| repository layout and prints a summary. |
| |
| With `--params`, it can also compute parameter counts by scanning |
| `*.safetensors` headers (without loading tensor data into RAM). |
| """ |
|
|
| import argparse |
| import json |
| import math |
| from pathlib import Path |
| from typing import Any, Dict, Iterable, Optional |
|
|
|
|
| def load_json(path: Path) -> Dict[str, Any]: |
| return json.loads(path.read_text(encoding="utf-8")) |
|
|
|
|
| def human_params(value: Optional[int]) -> str: |
| if value is None: |
| return "n/a" |
| if value >= 1_000_000_000: |
| return f"{value/1e9:.2f}B" |
| return f"{value/1e6:.2f}M" |
|
|
|
|
| def read_model_index(model_dir: Path) -> Dict[str, Any]: |
| idx_path = model_dir / "model_index.json" |
| if not idx_path.exists(): |
| return {} |
| return load_json(idx_path) |
|
|
|
|
| def describe_model_index(model_index: Dict[str, Any]) -> None: |
| if not model_index: |
| return |
| print("Pipeline pieces (model_index.json):") |
| for key, val in model_index.items(): |
| if key.startswith("_"): |
| continue |
| print(f" {key:14s} -> {val}") |
| print() |
|
|
|
|
| def detect_pipeline_kind(model_index: Dict[str, Any]) -> str: |
| cls = str(model_index.get("_class_name", "")).lower() |
| if "zimage" in cls or ("transformer" in model_index and "unet" not in model_index): |
| return "zimage" |
| if "stable" in cls or "unet" in model_index: |
| return "sdxl_like" |
| return "unknown" |
|
|
|
|
| def iter_safetensors_files(directory: Path) -> Iterable[Path]: |
| if not directory.exists(): |
| return [] |
| return sorted(p for p in directory.iterdir() if p.is_file() and p.suffix == ".safetensors") |
|
|
|
|
| def count_params_from_safetensors(files: Iterable[Path]) -> int: |
| from safetensors import safe_open |
|
|
| total = 0 |
| for file in files: |
| with safe_open(str(file), framework="np") as f: |
| for key in f.keys(): |
| shape = f.get_slice(key).get_shape() |
| total += math.prod(shape) |
| return int(total) |
|
|
|
|
| def zimage_config_only_summary(model_dir: Path, include_params: bool) -> Dict[str, Any]: |
| model_index = read_model_index(model_dir) |
|
|
| te_cfg_path = model_dir / "text_encoder" / "config.json" |
| transformer_cfg_path = model_dir / "transformer" / "config.json" |
| vae_cfg_path = model_dir / "vae" / "config.json" |
| scheduler_cfg_path = model_dir / "scheduler" / "scheduler_config.json" |
|
|
| te_cfg = load_json(te_cfg_path) if te_cfg_path.exists() else {} |
| transformer_cfg = load_json(transformer_cfg_path) if transformer_cfg_path.exists() else {} |
| vae_cfg = load_json(vae_cfg_path) if vae_cfg_path.exists() else {} |
| scheduler_cfg = load_json(scheduler_cfg_path) if scheduler_cfg_path.exists() else {} |
|
|
| text_encoder_params = None |
| transformer_params = None |
| vae_params = None |
| if include_params: |
| text_encoder_params = count_params_from_safetensors(iter_safetensors_files(model_dir / "text_encoder")) |
| transformer_params = count_params_from_safetensors(iter_safetensors_files(model_dir / "transformer")) |
| vae_params = count_params_from_safetensors(iter_safetensors_files(model_dir / "vae")) |
|
|
| print("[Text encoder]") |
| if te_cfg: |
| arch = te_cfg.get("architectures", []) |
| arch_name = arch[0] if isinstance(arch, list) and arch else "n/a" |
| print(f" architecture={arch_name}") |
| print( |
| " " |
| f"layers={te_cfg.get('num_hidden_layers', 'n/a')}, " |
| f"hidden={te_cfg.get('hidden_size', 'n/a')}, " |
| f"heads={te_cfg.get('num_attention_heads', 'n/a')}, " |
| f"intermediate={te_cfg.get('intermediate_size', 'n/a')}" |
| ) |
| print(f" vocab={te_cfg.get('vocab_size', 'n/a')}, max_positions={te_cfg.get('max_position_embeddings', 'n/a')}") |
| else: |
| print(" [warn] missing text_encoder/config.json") |
| print(f" params={human_params(text_encoder_params)}") |
| print() |
|
|
| print("[Transformer]") |
| if transformer_cfg: |
| print(f" class={transformer_cfg.get('_class_name', 'n/a')}") |
| print( |
| " " |
| f"dim={transformer_cfg.get('dim', 'n/a')}, " |
| f"layers={transformer_cfg.get('n_layers', 'n/a')}, " |
| f"heads={transformer_cfg.get('n_heads', 'n/a')}" |
| ) |
| print(f" in_channels={transformer_cfg.get('in_channels', 'n/a')}, cap_feat_dim={transformer_cfg.get('cap_feat_dim', 'n/a')}") |
| print(f" patch_size={transformer_cfg.get('all_patch_size', 'n/a')}, f_patch_size={transformer_cfg.get('all_f_patch_size', 'n/a')}") |
| else: |
| print(" [warn] missing transformer/config.json") |
| print(f" params={human_params(transformer_params)}") |
| print() |
|
|
| print("[VAE]") |
| if vae_cfg: |
| print(f" class={vae_cfg.get('_class_name', 'n/a')}") |
| print( |
| " " |
| f"sample_size={vae_cfg.get('sample_size', 'n/a')}, " |
| f"in_channels={vae_cfg.get('in_channels', 'n/a')}, " |
| f"latent_channels={vae_cfg.get('latent_channels', 'n/a')}, " |
| f"out_channels={vae_cfg.get('out_channels', 'n/a')}" |
| ) |
| print(f" block_out_channels={vae_cfg.get('block_out_channels', 'n/a')}, scaling_factor={vae_cfg.get('scaling_factor', 'n/a')}") |
| else: |
| print(" [warn] missing vae/config.json") |
| print(f" params={human_params(vae_params)}") |
| print() |
|
|
| print("[Scheduler]") |
| if scheduler_cfg: |
| print( |
| " " |
| f"class={scheduler_cfg.get('_class_name', 'n/a')}, " |
| f"timesteps={scheduler_cfg.get('num_train_timesteps', 'n/a')}, " |
| f"shift={scheduler_cfg.get('shift', 'n/a')}" |
| ) |
| else: |
| print(" [warn] missing scheduler/scheduler_config.json") |
| print() |
|
|
| return { |
| "kind": "zimage", |
| "pipeline": model_index, |
| "text_encoder": {"config": te_cfg, "params": text_encoder_params}, |
| "transformer": {"config": transformer_cfg, "params": transformer_params}, |
| "vae": {"config": vae_cfg, "params": vae_params}, |
| "scheduler": {"config": scheduler_cfg}, |
| } |
|
|
|
|
| def main() -> None: |
| parser = argparse.ArgumentParser(description="Inspect a local Diffusers-style repository layout.") |
| parser.add_argument("--model-dir", type=Path, default=Path(".."), help="Path to the diffusers pipeline directory.") |
| parser.add_argument("--device", default="cpu", help="Unused (kept for CLI compatibility).") |
| parser.add_argument("--fp16", action="store_true", help="Unused (kept for CLI compatibility).") |
| parser.add_argument("--config-only", action="store_true", help="Read JSON configs and print a summary.") |
| parser.add_argument("--params", action="store_true", help="Count parameters from *.safetensors headers (no tensor loading).") |
| parser.add_argument("--json-out", type=Path, default=None, help="Write a JSON summary to this path.") |
| args = parser.parse_args() |
|
|
| model_index = read_model_index(args.model_dir) |
| if not model_index: |
| raise SystemExit(f"model_index.json not found under {args.model_dir}") |
|
|
| describe_model_index(model_index) |
| kind = detect_pipeline_kind(model_index) |
|
|
| if not args.config_only: |
| raise SystemExit("Only --config-only mode is supported by this inspector.") |
|
|
| if kind != "zimage": |
| raise SystemExit(f"Unsupported pipeline kind: {kind} (expected ZImagePipeline-style layout)") |
|
|
| summary = zimage_config_only_summary(args.model_dir, include_params=args.params) |
|
|
| if args.json_out is not None: |
| args.json_out.parent.mkdir(parents=True, exist_ok=True) |
| args.json_out.write_text(json.dumps(summary, indent=2, ensure_ascii=False) + "\n", encoding="utf-8") |
| print(f"[info] wrote JSON summary to {args.json_out}") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|