| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| import torch |
| from torch import nn |
| from torch.utils import model_zoo |
|
|
| url_map = { |
| "clip_encoding_universal_model_32": ( |
| "https://github.com/Project-MONAI/MONAI-extra-test-data/" |
| "releases/download/0.8.1/clip_encoding_universal_model.pth" |
| ) |
| } |
|
|
|
|
| class TextEncoder(nn.Module): |
| """ |
| Text to vision encoding by Contrastive Language-Image Pre-training (CLIP) or random embedding. |
| The text to vision encoder loads the pre-trained or random initialized weights with connection to 2D/3D vision models. |
| |
| Contrastive Language-Image Pre-training (CLIP), based on: "Radford et al., |
| Learning Transferable Visual Models From Natural Language Supervision <https://arxiv.org/abs/2103.00020>" |
| |
| Connecting text and medical 3D image, based on: "Liu et al., |
| CLIP-Driven Universal Model for Organ Segmentation and Tumor Detection <https://arxiv.org/pdf/2301.00785.pdf>" |
| """ |
|
|
| def __init__( |
| self, |
| out_channels: int, |
| spatial_dims: int = 3, |
| text_dim: int = 512, |
| hidden_size: int = 256, |
| encoding: str = "clip_encoding_universal_model_32", |
| pretrained: bool = True, |
| ) -> None: |
| """ |
| Args: |
| out_channels: number of output channels, to control text-based embedding for classes. |
| spatial_dims: number of spatial dims. |
| text_dim: dimension of text embeddings. |
| hidden_size: dimension of hidden features, compatible to different vision feature dimensions. |
| encoding: the text embedding type, default to use clip text pretrained weights. |
| pretrained: whether to load pretrained weights from e.g., (CLIP) to initialize text embeddings, default to False. |
| """ |
| super().__init__() |
| self.encoding = encoding |
|
|
| self.spatial_dims = spatial_dims |
| if spatial_dims not in (2, 3): |
| raise ValueError("spatial dimension should be 2 or 3.") |
|
|
| if self.encoding == "rand_embedding": |
| self.text_embedding = nn.Embedding(out_channels, hidden_size) |
| else: |
| self.register_buffer("text_embedding", torch.randn(out_channels, text_dim)) |
|
|
| if pretrained: |
| model_url = url_map[self.encoding] |
| pretrain_state_dict = model_zoo.load_url(model_url, map_location="cpu") |
| self.text_embedding.data = pretrain_state_dict.float() |
| else: |
| print(f"{self.encoding} is not implemented, and can not be downloaded, please load your own") |
|
|
| self.text_to_vision = nn.Linear(text_dim, hidden_size) |
|
|
| def forward(self): |
| if self.encoding == "rand_embedding": |
| |
| text_embedding = self.text_embedding.weight |
| else: |
| print(self.text_embedding) |
| text_embedding = nn.functional.relu(self.text_to_vision(self.text_embedding)) |
|
|
| if self.spatial_dims == 3: |
| text_embedding = text_embedding.unsqueeze(2).unsqueeze(2).unsqueeze(2) |
| elif self.spatial_dims == 2: |
| text_embedding = text_embedding.unsqueeze(2).unsqueeze(2) |
|
|
| return text_embedding |
|
|