| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| import math |
| from typing import Any |
|
|
| import numpy as np |
| import torch |
|
|
| from monai.config import KeysCollection, NdarrayOrTensor |
| from monai.networks.layers import GaussianFilter |
| from monai.transforms import MapTransform, Randomizable, SpatialPad |
| from monai.utils import StrEnum, convert_to_numpy, optional_import |
|
|
| measure, _ = optional_import("skimage.measure") |
| morphology, _ = optional_import("skimage.morphology") |
| distance_transform_cdt, _ = optional_import("scipy.ndimage.morphology", name="distance_transform_cdt") |
|
|
|
|
| class NuclickKeys(StrEnum): |
| """ |
| Keys for nuclick transforms. |
| """ |
|
|
| IMAGE = "image" |
| LABEL = "label" |
| OTHERS = "others" |
| FOREGROUND = "foreground" |
|
|
| CENTROID = "centroid" |
| MASK_VALUE = "mask_value" |
| LOCATION = "location" |
|
|
| NUC_POINTS = "nuc_points" |
| BOUNDING_BOXES = "bounding_boxes" |
| IMG_HEIGHT = "img_height" |
| IMG_WIDTH = "img_width" |
| PRED_CLASSES = "pred_classes" |
|
|
|
|
| class FlattenLabeld(MapTransform): |
| """ |
| FlattenLabeld creates labels per closed object contour (defined by a connectivity). For e.g if there are |
| 12 small regions of 1's it will delineate them into 12 different label classes |
| |
| Args: |
| connectivity: Max no. of orthogonal hops to consider a pixel/voxel as a neighbor. Refer skimage.measure.label |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
|
|
| def __init__(self, keys: KeysCollection, connectivity: int = 1, allow_missing_keys: bool = False): |
| super().__init__(keys, allow_missing_keys) |
| self.connectivity = connectivity |
|
|
| def __call__(self, data): |
| d = dict(data) |
| for key in self.keys: |
| img = convert_to_numpy(d[key]) if isinstance(d[key], torch.Tensor) else d[key] |
| d[key] = measure.label(img, connectivity=self.connectivity).astype(np.uint8) |
| return d |
|
|
|
|
| class ExtractPatchd(MapTransform): |
| """ |
| Extracts a patch from the given image and label, however it is based on the centroid location. |
| The centroid location is a 2D coordinate (H, W). The extracted patch is extracted around the centroid, |
| if the centroid is towards the edge, the centroid will not be the center of the image as the patch will be |
| extracted from the edges onwards |
| |
| Args: |
| keys: image, label |
| centroid_key: key where the centroid values are stored, defaults to ``"centroid"`` |
| patch_size: size of the extracted patch |
| allow_missing_keys: don't raise exception if key is missing. |
| pad_kwargs: other arguments for the SpatialPad transform |
| """ |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| centroid_key: str = NuclickKeys.CENTROID, |
| patch_size: tuple[int, int] | int = 128, |
| allow_missing_keys: bool = False, |
| **kwargs: Any, |
| ): |
| super().__init__(keys, allow_missing_keys) |
| self.centroid_key = centroid_key |
| self.patch_size = patch_size |
| self.kwargs = kwargs |
|
|
| def __call__(self, data): |
| d = dict(data) |
|
|
| centroid = d[self.centroid_key] |
| roi_size = (self.patch_size, self.patch_size) |
|
|
| for key in self.keys: |
| img = d[key] |
| x_start, x_end, y_start, y_end = self.bbox(self.patch_size, centroid, img.shape[-2:]) |
| cropped = img[:, x_start:x_end, y_start:y_end] |
| d[key] = SpatialPad(spatial_size=roi_size, **self.kwargs)(cropped) |
| return d |
|
|
| def bbox(self, patch_size, centroid, size): |
| x, y = centroid |
| m, n = size |
|
|
| x_start = int(max(x - patch_size / 2, 0)) |
| y_start = int(max(y - patch_size / 2, 0)) |
| x_end = x_start + patch_size |
| y_end = y_start + patch_size |
| if x_end > m: |
| x_end = m |
| x_start = m - patch_size |
| if y_end > n: |
| y_end = n |
| y_start = n - patch_size |
| return x_start, x_end, y_start, y_end |
|
|
|
|
| class SplitLabeld(MapTransform): |
| """ |
| Extracts a single label from all the given classes, the single label is defined by mask_value, the remaining |
| labels are kept in others |
| |
| Args: |
| label: key of the label source |
| others: other labels storage key, defaults to ``"others"`` |
| mask_value: the mask_value that will be kept for binarization of the label, defaults to ``"mask_value"`` |
| min_area: The smallest allowable object size. |
| others_value: Value/class for other nuclei; Use this to separate core nuclei vs others. |
| to_binary_mask: Convert mask to binary; Set it false to restore original class values |
| """ |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| others: str = NuclickKeys.OTHERS, |
| mask_value: str | None = NuclickKeys.MASK_VALUE, |
| min_area: int = 5, |
| others_value: int = 0, |
| to_binary_mask: bool = True, |
| ): |
| super().__init__(keys, allow_missing_keys=False) |
| self.others = others |
| self.mask_value = mask_value |
| self.min_area = min_area |
| self.others_value = others_value |
| self.to_binary_mask = to_binary_mask |
|
|
| def __call__(self, data): |
| d = dict(data) |
|
|
| if len(self.keys) > 1: |
| print("Only 'label' key is supported, more than 1 key was found") |
| return None |
|
|
| for key in self.keys: |
| label = d[key] if isinstance(d[key], torch.Tensor) else torch.from_numpy(d[key]) |
|
|
| mask = torch.clone(label) |
| if self.mask_value: |
| mask_value = d[self.mask_value] |
| mask[label != mask_value] = 0 |
| else: |
| mask[label >= self.others_value] = 0 |
| mask_value = int(torch.max(mask)) |
|
|
| if self.to_binary_mask: |
| mask[mask > 0] = 1 |
|
|
| others = torch.clone(label) |
| others[label == mask_value] = 0 |
| others[others > 0] = 1 |
| if torch.count_nonzero(others): |
| others = measure.label(convert_to_numpy(others)[0], connectivity=1) |
| others = torch.from_numpy(others)[None] |
|
|
| label = mask.type(torch.uint8) if isinstance(mask, torch.Tensor) else mask |
| others = others.type(torch.uint8) if isinstance(others, torch.Tensor) else others |
|
|
| d[key] = label if isinstance(d[key], torch.Tensor) else convert_to_numpy(label) |
| d[self.others] = others if isinstance(d[key], torch.Tensor) else convert_to_numpy(others) |
|
|
| return d |
|
|
|
|
| class FilterImaged(MapTransform): |
| """ |
| Filters Green and Gray channel of the image using an allowable object size, this pre-processing transform |
| is specific towards NuClick training process. More details can be referred in this paper Koohbanani, |
| Navid Alemi, et al. "NuClick: a deep learning framework for interactive segmentation of microscopic images." |
| Medical Image Analysis 65 (2020): 101771. |
| |
| Args: |
| min_size: The smallest allowable object size |
| allow_missing_keys: don't raise exception if key is missing. |
| """ |
|
|
| def __init__(self, keys: KeysCollection, min_size: int = 500, allow_missing_keys: bool = False): |
| super().__init__(keys, allow_missing_keys) |
| self.min_size = min_size |
|
|
| def __call__(self, data): |
| d = dict(data) |
| for key in self.keys: |
| img = convert_to_numpy(d[key]) if isinstance(d[key], torch.Tensor) else d[key] |
| d[key] = self.filter(img) |
| return d |
|
|
| def filter(self, rgb): |
| mask_not_green = self.filter_green_channel(rgb) |
| mask_not_gray = self.filter_grays(rgb) |
| mask_gray_green = mask_not_gray & mask_not_green |
| mask = ( |
| self.filter_remove_small_objects(mask_gray_green, min_size=self.min_size) |
| if self.min_size |
| else mask_gray_green |
| ) |
|
|
| return rgb * np.dstack([mask, mask, mask]) |
|
|
| def filter_green_channel( |
| self, img_np, green_thresh=200, avoid_overmask=True, overmask_thresh=90, output_type="bool" |
| ): |
| g = img_np[:, :, 1] |
| gr_ch_mask = (g < green_thresh) & (g > 0) |
| mask_percentage = self.mask_percent(gr_ch_mask) |
| if (mask_percentage >= overmask_thresh) and (green_thresh < 255) and (avoid_overmask is True): |
| new_green_thresh = math.ceil((255 - green_thresh) / 2 + green_thresh) |
| gr_ch_mask = self.filter_green_channel( |
| img_np, new_green_thresh, avoid_overmask, overmask_thresh, output_type |
| ) |
| return gr_ch_mask |
|
|
| def filter_grays(self, rgb, tolerance=15): |
| rg_diff = abs(rgb[:, :, 0] - rgb[:, :, 1]) <= tolerance |
| rb_diff = abs(rgb[:, :, 0] - rgb[:, :, 2]) <= tolerance |
| gb_diff = abs(rgb[:, :, 1] - rgb[:, :, 2]) <= tolerance |
| return ~(rg_diff & rb_diff & gb_diff) |
|
|
| def mask_percent(self, img_np): |
| if (len(img_np.shape) == 3) and (img_np.shape[2] == 3): |
| np_sum = img_np[:, :, 0] + img_np[:, :, 1] + img_np[:, :, 2] |
| mask_percentage = 100 - np.count_nonzero(np_sum) / np_sum.size * 100 |
| else: |
| mask_percentage = 100 - np.count_nonzero(img_np) / img_np.size * 100 |
| return mask_percentage |
|
|
| def filter_remove_small_objects(self, img_np, min_size=3000, avoid_overmask=True, overmask_thresh=95): |
| rem_sm = morphology.remove_small_objects(img_np.astype(bool), min_size=min_size) |
| mask_percentage = self.mask_percent(rem_sm) |
| if (mask_percentage >= overmask_thresh) and (min_size >= 1) and (avoid_overmask is True): |
| new_min_size = round(min_size / 2) |
| rem_sm = self.filter_remove_small_objects(img_np, new_min_size, avoid_overmask, overmask_thresh) |
| return rem_sm |
|
|
|
|
| class AddPointGuidanceSignald(Randomizable, MapTransform): |
| """ |
| Adds Guidance Signal to the input image |
| |
| Args: |
| image: key of source image, defaults to ``"image"`` |
| label: key of source label, defaults to ``"label"`` |
| others: source others (other labels from the binary mask which are not being used for training) |
| defaults to ``"others"`` |
| drop_rate: probability of dropping the signal, defaults to ``0.5`` |
| jitter_range: noise added to the points in the point mask for exclusion mask, defaults to ``3`` |
| gaussian: add gaussian |
| sigma: sigma value for gaussian |
| truncated: spreads how many stds for gaussian |
| add_exclusion_map: add exclusion map/signal |
| """ |
|
|
| def __init__( |
| self, |
| image: str = NuclickKeys.IMAGE, |
| label: str = NuclickKeys.LABEL, |
| others: str = NuclickKeys.OTHERS, |
| drop_rate: float = 0.5, |
| jitter_range: int = 0, |
| gaussian: bool = False, |
| sigma: float = 1.0, |
| truncated: float = 2.0, |
| add_exclusion_map: bool = True, |
| use_distance: bool = False, |
| ): |
| MapTransform.__init__(self, image) |
|
|
| self.image = image |
| self.label = label |
| self.others = others |
| self.drop_rate = drop_rate |
| self.jitter_range = jitter_range |
| self.gaussian = gaussian |
| self.sigma = sigma |
| self.truncated = truncated |
| self.add_exclusion_map = add_exclusion_map |
| self.use_distance = use_distance |
|
|
| def __call__(self, data): |
| d = dict(data) |
|
|
| image = d[self.image] if isinstance(d[self.image], torch.Tensor) else torch.from_numpy(d[self.image]) |
| mask = d[self.label] if isinstance(d[self.label], torch.Tensor) else torch.from_numpy(d[self.label]) |
|
|
| inc_sig = self.inclusion_map(mask[0], dtype=image.dtype) |
| inc_sig = self._apply_gaussian(inc_sig) |
| if self.add_exclusion_map: |
| others = d[self.others] if isinstance(d[self.others], torch.Tensor) else torch.from_numpy(d[self.others]) |
| exc_sig = self.exclusion_map( |
| others[0], dtype=image.dtype, drop_rate=self.drop_rate, jitter_range=self.jitter_range |
| ) |
| exc_sig = self._apply_gaussian(exc_sig) |
| image = torch.cat((image, inc_sig[None], exc_sig[None]), dim=0) |
| else: |
| image = torch.cat((image, inc_sig[None]), dim=0) |
|
|
| d[self.image] = image if isinstance(d[self.image], torch.Tensor) else convert_to_numpy(image) |
| return d |
|
|
| def _apply_gaussian(self, t): |
| if not self.gaussian or torch.count_nonzero(t) == 0: |
| return t |
| x = GaussianFilter(spatial_dims=2, truncated=self.truncated, sigma=self.sigma)(t.unsqueeze(0).unsqueeze(0)) |
| return x.squeeze(0).squeeze(0) |
|
|
| def _seed_point(self, label): |
| if distance_transform_cdt is None or not self.use_distance: |
| indices: NdarrayOrTensor |
| if hasattr(torch, "argwhere"): |
| indices = torch.argwhere(label > 0) |
| else: |
| indices = np.argwhere(convert_to_numpy(label) > 0) |
|
|
| if len(indices) > 0: |
| index = self.R.randint(0, len(indices)) |
| return indices[index, 0], indices[index, 1] |
| return None |
|
|
| distance = distance_transform_cdt(label).flatten() |
| probability = np.exp(distance) - 1.0 |
|
|
| idx = np.where(label.flatten() > 0)[0] |
| seed = self.R.choice(idx, size=1, p=probability[idx] / np.sum(probability[idx])) |
| g = np.asarray(np.unravel_index(seed, label.shape)).transpose().tolist()[0] |
| return g[-2], g[-1] |
|
|
| def inclusion_map(self, mask, dtype): |
| point_mask = torch.zeros_like(mask, dtype=dtype) |
| pt = self._seed_point(mask) |
| if pt is not None: |
| point_mask[pt[0], pt[1]] = 1 |
|
|
| return point_mask |
|
|
| def exclusion_map(self, others, dtype, jitter_range, drop_rate): |
| point_mask = torch.zeros_like(others, dtype=dtype) |
| if np.random.choice([True, False], p=[drop_rate, 1 - drop_rate]): |
| return point_mask |
|
|
| max_x = point_mask.shape[0] - 1 |
| max_y = point_mask.shape[1] - 1 |
| stats = measure.regionprops(convert_to_numpy(others)) |
| for stat in stats: |
| if np.random.choice([True, False], p=[drop_rate, 1 - drop_rate]): |
| continue |
|
|
| |
| x, y = stat.centroid |
| x = int(math.floor(x)) |
| y = int(math.floor(y)) |
| if jitter_range: |
| x = x + self.R.randint(low=-jitter_range, high=jitter_range) |
| y = y + self.R.randint(low=-jitter_range, high=jitter_range) |
| x = min(max(0, x), max_x) |
| y = min(max(0, y), max_y) |
| point_mask[x, y] = 1 |
|
|
| return point_mask |
|
|
|
|
| class AddClickSignalsd(MapTransform): |
| """ |
| Adds Click Signal to the input image |
| |
| Args: |
| image: source image, defaults to ``"image"`` |
| foreground: 2D click indices as list, defaults to ``"foreground"`` |
| bb_size: single integer size, defines a bounding box like (bb_size, bb_size) |
| gaussian: add gaussian |
| sigma: sigma value for gaussian |
| truncated: spreads how many stds for gaussian |
| add_exclusion_map: add exclusion map/signal |
| """ |
|
|
| def __init__( |
| self, |
| image: str = NuclickKeys.IMAGE, |
| foreground: str = NuclickKeys.FOREGROUND, |
| bb_size: int = 128, |
| gaussian: bool = False, |
| sigma: float = 1.0, |
| truncated: float = 2.0, |
| add_exclusion_map: bool = True, |
| ): |
| self.image = image |
| self.foreground = foreground |
| self.bb_size = bb_size |
| self.gaussian = gaussian |
| self.sigma = sigma |
| self.truncated = truncated |
| self.add_exclusion_map = add_exclusion_map |
|
|
| def __call__(self, data): |
| d = dict(data) |
|
|
| img = d[self.image] if isinstance(d[self.image], torch.Tensor) else torch.from_numpy(d[self.image]) |
| x = img.shape[-2] |
| y = img.shape[-1] |
|
|
| location = d.get(NuclickKeys.LOCATION.value, (0, 0)) |
| tx, ty = location[0], location[1] |
| pos = d.get(self.foreground) |
| pos = (np.array(pos) - (tx, ty)).astype(int).tolist() if pos else [] |
|
|
| cx = [xy[0] for xy in pos] |
| cy = [xy[1] for xy in pos] |
|
|
| click_map, bounding_boxes = self.get_clickmap_boundingbox(img, cx=cx, cy=cy, x=x, y=y, bb=self.bb_size) |
| if not bounding_boxes: |
| raise ValueError("Failed to create patches from given click points") |
|
|
| patches = self.get_patches_and_signals( |
| img=img, click_map=click_map, bounding_boxes=bounding_boxes, cx=cx, cy=cy, x=x, y=y |
| ) |
|
|
| d[NuclickKeys.BOUNDING_BOXES.value] = bounding_boxes |
| d[NuclickKeys.IMG_WIDTH.value] = x |
| d[NuclickKeys.IMG_HEIGHT.value] = y |
|
|
| d[self.image] = patches if isinstance(d[self.image], torch.Tensor) else convert_to_numpy(patches) |
| return d |
|
|
| def get_clickmap_boundingbox(self, img, cx, cy, x, y, bb=128): |
| click_map = torch.zeros_like(img[0]) |
|
|
| x_del_indices = {i for i in range(len(cx)) if cx[i] >= x or cx[i] < 0} |
| y_del_indices = {i for i in range(len(cy)) if cy[i] >= y or cy[i] < 0} |
| del_indices = list(x_del_indices.union(y_del_indices)) |
| cx = np.delete(cx, del_indices) |
| cy = np.delete(cy, del_indices) |
|
|
| click_map[cx, cy] = 1 |
| bounding_boxes = [] |
| for i in range(len(cx)): |
| x_start = max(0, cx[i] - bb // 2) |
| y_start = max(0, cy[i] - bb // 2) |
| x_end = min(x_start + bb, x) |
| y_end = min(y_start + bb, y) |
|
|
| if x_end - x_start != bb: |
| x_start = x_end - bb |
| if y_end - y_start != bb: |
| y_start = y_end - bb |
| if x_end - x_start == bb and y_end - y_start == bb: |
| bounding_boxes.append([x_start, y_start, x_end, y_end]) |
| else: |
| print(f"Ignore smaller sized bbox ({x_start}, {y_start}, {x_end}, {y_end}) (Min size: {bb}x{bb})") |
| return click_map, bounding_boxes |
|
|
| def get_patches_and_signals(self, img, click_map, bounding_boxes, cx, cy, x, y): |
| patches = [] |
|
|
| x_del_indices = {i for i in range(len(cx)) if cx[i] >= x or cx[i] < 0} |
| y_del_indices = {i for i in range(len(cy)) if cy[i] >= y or cy[i] < 0} |
| del_indices = list(x_del_indices.union(y_del_indices)) |
| cx = np.delete(cx, del_indices) |
| cy = np.delete(cy, del_indices) |
|
|
| for i, bounding_box in enumerate(bounding_boxes): |
| x_start = bounding_box[0] |
| y_start = bounding_box[1] |
| x_end = bounding_box[2] |
| y_end = bounding_box[3] |
|
|
| patch = img[:, x_start:x_end, y_start:y_end] |
|
|
| this_click_map = torch.zeros_like(img[0]) |
| this_click_map[cx[i], cy[i]] = 1 |
|
|
| nuc_points = this_click_map[x_start:x_end, y_start:y_end] |
| nuc_points = self._apply_gaussian(nuc_points) |
|
|
| if self.add_exclusion_map: |
| others_click_map = ((click_map - this_click_map) > 0).type(img.dtype) |
| other_points = others_click_map[x_start:x_end, y_start:y_end] |
| other_points = self._apply_gaussian(other_points) |
| patches.append(torch.cat([patch, nuc_points[None], other_points[None]])) |
| else: |
| patches.append(torch.cat([patch, nuc_points[None]])) |
|
|
| return torch.stack(patches) |
|
|
| def _apply_gaussian(self, t): |
| if not self.gaussian or torch.count_nonzero(t) == 0: |
| return t |
| x = GaussianFilter(spatial_dims=2, truncated=self.truncated, sigma=self.sigma)(t.unsqueeze(0).unsqueeze(0)) |
| return x.squeeze(0).squeeze(0) |
|
|
|
|
| class PostFilterLabeld(MapTransform): |
| """ |
| Performs Filtering of Labels on the predicted probability map |
| |
| Args: |
| thresh: probability threshold for classifying a pixel as a mask |
| min_size: min_size objects that will be removed from the image, refer skimage remove_small_objects |
| min_hole: min_hole that will be removed from the image, refer skimage remove_small_holes |
| do_reconstruction: Boolean Flag, Perform a morphological reconstruction of an image, refer skimage |
| allow_missing_keys: don't raise exception if key is missing. |
| pred_classes: List of Predicted class for each instance |
| """ |
|
|
| def __init__( |
| self, |
| keys: KeysCollection, |
| nuc_points: str = NuclickKeys.NUC_POINTS, |
| bounding_boxes: str = NuclickKeys.BOUNDING_BOXES, |
| img_height: str = NuclickKeys.IMG_HEIGHT, |
| img_width: str = NuclickKeys.IMG_WIDTH, |
| thresh: float = 0.33, |
| min_size: int = 10, |
| min_hole: int = 30, |
| do_reconstruction: bool = False, |
| allow_missing_keys: bool = False, |
| pred_classes: str = NuclickKeys.PRED_CLASSES, |
| ): |
| super().__init__(keys, allow_missing_keys) |
| self.nuc_points = nuc_points |
| self.bounding_boxes = bounding_boxes |
| self.img_height = img_height |
| self.img_width = img_width |
|
|
| self.thresh = thresh |
| self.min_size = min_size |
| self.min_hole = min_hole |
| self.do_reconstruction = do_reconstruction |
| self.pred_classes = pred_classes |
|
|
| def __call__(self, data): |
| d = dict(data) |
|
|
| pred_classes = d.get(self.pred_classes) |
| bounding_boxes = d[self.bounding_boxes] |
| x = d[self.img_width] |
| y = d[self.img_height] |
|
|
| for key in self.keys: |
| label = d[key].astype(np.uint8) |
| masks = self.post_processing(label, self.thresh, self.min_size, self.min_hole) |
| d[key] = self.gen_instance_map(masks, bounding_boxes, x, y, pred_classes=pred_classes).astype(np.uint8) |
| return d |
|
|
| def post_processing(self, preds, thresh=0.33, min_size=10, min_hole=30): |
| masks = preds > thresh |
| for i in range(preds.shape[0]): |
| masks[i] = morphology.remove_small_objects(masks[i], min_size=min_size) |
| masks[i] = morphology.remove_small_holes(masks[i], area_threshold=min_hole) |
| return masks |
|
|
| def gen_instance_map(self, masks, bounding_boxes, x, y, flatten=True, pred_classes=None): |
| instance_map = np.zeros((x, y), dtype=np.uint16) |
| for i, mask in enumerate(masks): |
| bb = bounding_boxes[i] |
| c = pred_classes[i] if pred_classes and i < len(pred_classes) else 1 |
| c = c if flatten else i + 1 |
|
|
| this_map = instance_map[bb[0] : bb[2], bb[1] : bb[3]] |
| this_map = np.where(mask > 0, c, this_map) |
| instance_map[bb[0] : bb[2], bb[1] : bb[3]] = this_map |
|
|
| return instance_map |
|
|
|
|
| class AddLabelAsGuidanced(MapTransform): |
| """ |
| Add Label as new guidance channel |
| |
| Args: |
| source: label/source key which gets added as additional guidance channel |
| """ |
|
|
| def __init__(self, keys: KeysCollection, source: str = "label") -> None: |
| super().__init__(keys, allow_missing_keys=False) |
| self.source = source |
|
|
| def __call__(self, data): |
| d = dict(data) |
| for key in self.keys: |
| image = d[key] if isinstance(d[key], torch.Tensor) else torch.from_numpy(d[key]) |
| label = d[self.source] if isinstance(d[self.source], torch.Tensor) else torch.from_numpy(d[self.source]) |
|
|
| label = label > 0 |
| if len(label.shape) < len(image.shape): |
| label = label[None] |
| image = torch.cat([image, label.type(image.dtype)], dim=len(label.shape) - 3) |
| d[key] = image if isinstance(d[key], torch.Tensor) else convert_to_numpy(image) |
| return d |
|
|
|
|
| class SetLabelClassd(MapTransform): |
| """ |
| Assign class value from the labelmap. This converts multi-dimension tensor to single scalar tensor. |
| |
| Args: |
| offset: offset value to be added to the mask value to determine the final class |
| """ |
|
|
| def __init__(self, keys: KeysCollection, offset: int = -1) -> None: |
| super().__init__(keys, allow_missing_keys=False) |
| self.offset = offset |
|
|
| def __call__(self, data): |
| d = dict(data) |
| for key in self.keys: |
| label = d[key] if isinstance(d[key], torch.Tensor) else torch.from_numpy(d[key]) |
| mask_value = int(torch.max(label)) |
| d[key] = mask_value + self.offset |
| return d |
|
|