| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from typing import Tuple |
|
|
| from PIL import Image |
| from torchvision import transforms |
| from transformers import Siglip2ImageProcessorFast |
|
|
| from .tokenizer_wrapper import ImageInfo, JointImageInfo, ResolutionGroup |
|
|
|
|
| def resize_and_crop(image: Image.Image, target_size: Tuple[int, int]) -> Image.Image: |
| tw, th = target_size |
| w, h = image.size |
|
|
| tr = th / tw |
| r = h / w |
|
|
| |
| if r < tr: |
| resize_height = th |
| resize_width = int(round(th / h * w)) |
| else: |
| resize_width = tw |
| resize_height = int(round(tw / w * h)) |
|
|
| image = image.resize((resize_width, resize_height), resample=Image.Resampling.LANCZOS) |
|
|
| |
| crop_top = int(round((resize_height - th) / 2.0)) |
| crop_left = int(round((resize_width - tw) / 2.0)) |
|
|
| image = image.crop((crop_left, crop_top, crop_left + tw, crop_top + th)) |
| return image |
|
|
|
|
| class HunyuanImage3ImageProcessor(object): |
| def __init__(self, config): |
| self.config = config |
|
|
| min_multiple = getattr(config, "image_min_multiple", 0.5) |
| max_multiple = getattr(config, "image_max_multiple", 2.0) |
| step = getattr(config, "image_resolution_step", None) |
| align = getattr(config, "image_resolution_align", 1) |
| max_entries = getattr(config, "image_resolution_count", 33) |
| presets = getattr(config, "image_resolution_presets", None) |
| self.reso_group = ResolutionGroup( |
| base_size=config.image_base_size, |
| step=step, |
| align=align, |
| min_multiple=min_multiple, |
| max_multiple=max_multiple, |
| max_entries=max_entries, |
| presets=presets, |
| ) |
| self.vae_processor = transforms.Compose([ |
| transforms.ToTensor(), |
| transforms.Normalize([0.5], [0.5]), |
| ]) |
| self.vision_encoder_processor = Siglip2ImageProcessorFast.from_dict(config.vit_processor) |
|
|
| def build_image_info(self, image_size): |
| |
| if isinstance(image_size, str): |
| if image_size.startswith("<img_ratio_"): |
| ratio_index = int(image_size.split("_")[-1].rstrip(">")) |
| reso = self.reso_group[ratio_index] |
| image_size = reso.height, reso.width |
| elif 'x' in image_size: |
| image_size = [int(s) for s in image_size.split('x')] |
| elif ':' in image_size: |
| image_size = [int(s) for s in image_size.split(':')] |
| else: |
| raise ValueError( |
| f"`image_size` should be in the format of 'HxW', 'H:W' or <img_ratio_i>, got {image_size}.") |
| assert len(image_size) == 2, f"`image_size` should be in the format of 'HxW', got {image_size}." |
| elif isinstance(image_size, (list, tuple)): |
| assert len(image_size) == 2 and all(isinstance(s, int) for s in image_size), \ |
| f"`image_size` should be a tuple of two integers or a string in the format of 'HxW', got {image_size}." |
| else: |
| raise ValueError(f"`image_size` should be a tuple of two integers or a string in the format of 'WxH', " |
| f"got {image_size}.") |
| image_width, image_height = self.reso_group.get_target_size(image_size[1], image_size[0]) |
| token_height = image_height // (self.config.vae_downsample_factor[0] * self.config.patch_size) |
| token_width = image_width // (self.config.vae_downsample_factor[1] * self.config.patch_size) |
| base_size, ratio_idx = self.reso_group.get_base_size_and_ratio_index(image_size[1], image_size[0]) |
| image_info = ImageInfo( |
| image_type="gen_image", image_width=image_width, image_height=image_height, |
| token_width=token_width, token_height=token_height, base_size=base_size, ratio_index=ratio_idx, |
| ) |
| return image_info |
|
|
| def preprocess(self, image: Image.Image): |
| |
| image_width, image_height = self.reso_group.get_target_size(image.width, image.height) |
| resized_image = resize_and_crop(image, (image_width, image_height)) |
| image_tensor = self.vae_processor(resized_image) |
| token_height = image_height // (self.config.vae_downsample_factor[0] * self.config.patch_size) |
| token_width = image_width // (self.config.vae_downsample_factor[1] * self.config.patch_size) |
| base_size, ratio_index = self.reso_group.get_base_size_and_ratio_index(width=image_width, height=image_height) |
| vae_image_info = ImageInfo( |
| image_type="vae", |
| image_tensor=image_tensor.unsqueeze(0), |
| image_width=image_width, image_height=image_height, |
| token_width=token_width, token_height=token_height, |
| base_size=base_size, ratio_index=ratio_index, |
| ) |
|
|
| |
| inputs = self.vision_encoder_processor(image) |
| image = inputs["pixel_values"].squeeze(0) |
| pixel_attention_mask = inputs["pixel_attention_mask"].squeeze(0) |
| spatial_shapes = inputs["spatial_shapes"].squeeze(0) |
| vision_encoder_kwargs = dict( |
| pixel_attention_mask=pixel_attention_mask, |
| spatial_shapes=spatial_shapes, |
| ) |
| vision_image_info = ImageInfo( |
| image_type="vit", |
| image_tensor=image.unsqueeze(0), |
| image_width=spatial_shapes[1].item() * self.config.vit_processor["patch_size"], |
| image_height=spatial_shapes[0].item() * self.config.vit_processor["patch_size"], |
| token_width=spatial_shapes[1].item(), |
| token_height=spatial_shapes[0].item(), |
| image_token_length=self.config.vit_processor["max_num_patches"], |
| |
| ) |
| return JointImageInfo(vae_image_info, vision_image_info, vision_encoder_kwargs) |
|
|