| | """ |
| | NOTE: Download the Cosmos-Tokenizer repository and pre-trained model weights before running this script. |
| | For full installation and setup instructions, please refer to: |
| | https://github.com/NVIDIA/Cosmos-Tokenizer#readme |
| | """ |
| |
|
| | import math |
| | from pathlib import Path |
| |
|
| | import av |
| | import numpy as np |
| | import torch |
| |
|
| | from cosmos_tokenizer.utils import tensor2numpy |
| | from cosmos_tokenizer.video_lib import CausalVideoTokenizer |
| |
|
| | input_dir = Path("../worldmodel/val_v2.0") |
| | output_dir = Path("/tmp/reconst_1xgpt/") |
| | model_name = "Cosmos-Tokenizer-DV8x8x8" |
| | decoder_path = Path("pretrained_ckpts") / model_name / "decoder.jit" |
| |
|
| | print(f"Output directory exists: {input_dir.exists()}") |
| | print(f"Decoder path exists: {decoder_path.exists()}") |
| |
|
| | rank = 0 |
| | metadata_path = input_dir / f"metadata_{rank}.json" |
| | if not metadata_path.exists(): |
| | raise FileNotFoundError(f"Metadata file not found at {metadata_path}") |
| |
|
| | with open(metadata_path, "r") as f: |
| | metadata_shard = json.load(f) |
| |
|
| | total_frames = metadata_shard["shard_num_frames"] |
| | print(f"Total frames: {total_frames}") |
| |
|
| | encoded_video_dataset = np.memmap(input_dir / f"video_{rank}.bin", dtype=np.int32, mode="r", shape=(math.ceil(total_frames / 17), 3, 32, 32)) |
| |
|
| | print(f"Encoded video dataset shape: {encoded_video_dataset.shape}") |
| |
|
| | indices = torch.tensor(encoded_video_dataset, device="cuda") if not isinstance(encoded_video_dataset, torch.Tensor) else encoded_video_dataset |
| |
|
| | try: |
| | decoder = CausalVideoTokenizer(checkpoint_dec=str(decoder_path)) |
| | if decoder._dec_model is None: |
| | raise RuntimeError(f"Failed to load decoder model from {decoder_path}") |
| | print("Decoder initialized successfully.") |
| | except Exception as e: |
| | raise RuntimeError(f"Error loading decoder: {str(e)}") from e |
| |
|
| | batch_size = 1 |
| | fps = 30 |
| | output_file = output_dir / "reconstructed_video.mp4" |
| |
|
| | first_batch = torch.from_numpy(encoded_video_dataset[0:1]).cuda() |
| | with torch.no_grad(): |
| | first_output = decoder.decode(first_batch).float() |
| | _, _, height, width = first_output.shape[-4:] |
| |
|
| | print(f"Output video dimensions: {width}x{height}") |
| |
|
| |
|
| | ec = av.open(str(output_file), mode="w") |
| | es = ec.add_stream("hevc_nvenc", rate=30) |
| | es.width = 256 |
| | es.height = 256 |
| |
|
| |
|
| | num_batches = math.ceil(len(encoded_video_dataset) / batch_size) |
| | for i in range(num_batches): |
| | start_idx = i * batch_size |
| | end_idx = min((i + 1) * batch_size, len(encoded_video_dataset)) |
| |
|
| | batch = torch.from_numpy(encoded_video_dataset[start_idx:end_idx]).cuda() |
| | with torch.no_grad(): |
| | |
| | reconstructed_batch = decoder.decode(batch) |
| |
|
| | |
| | reconstructed_batch = tensor2numpy(reconstructed_batch) |
| |
|
| | |
| | for this_batch in reconstructed_batch: |
| | for single_frame in this_batch: |
| | |
| | for ep in es.encode(av.VideoFrame.from_ndarray(single_frame, format="rgb24")): |
| | ec.mux(ep) |
| |
|
| | print(f"Processed batch {i + 1}/{num_batches}", flush=True) |
| | if i == 100: |
| | break |
| |
|
| | ec.close() |
| | print(f"Video saved to: {output_file}") |