| | from math import ceil |
| | from typing import List |
| |
|
| | import torch |
| | import torch.nn as nn |
| | from jaxtyping import Float |
| | from neuralop.models import FNO |
| | from scOT.model import ScOT, ScOTConfig, ScOTOutput |
| | from torchvision.models.resnet import BasicBlock, Bottleneck, ResNet |
| |
|
| | from pdeinvbench.utils.types import ( |
| | PDE, |
| | PDE_NUM_PARAMETERS, |
| | PDE_NUM_SPATIAL, |
| | PDE_PARTIALS, |
| | PDE_SPATIAL_SIZE, |
| | PDE_TRAJ_LEN, |
| | ) |
| |
|
| |
|
| | def resolve_number_input_channels( |
| | n_past: int, data_channels: int, use_partials: bool, pde: PDE |
| | ) -> int: |
| | """ |
| | Returns the number of input channels for a pde given args: |
| | - n_past |
| | - data_channels |
| | - use_partials |
| | """ |
| |
|
| | num_partials = PDE_PARTIALS[pde] |
| |
|
| | if use_partials: |
| | |
| | data_channels += num_partials * data_channels |
| |
|
| | in_channels = n_past * data_channels |
| |
|
| | return in_channels |
| |
|
| |
|
| | class FNOEncoder(FNO): |
| | """ |
| | Wrapper around FNO that figures out the input channels based |
| | on the number of past frames and partial derivatives. |
| | :param n_modes: Number of modes to use in the FNO. |
| | :param n_layers: Number of layers in the FNO. |
| | :param n_past: Number of past frames to use. |
| | :param pde: PDE to use for the partial derivatives. |
| | :param data_channels: Number of channels per timestep in the native input data. |
| | :param hidden_channels: Number of channels in the hidden layers. |
| | :param use_partials: Whether to use partial derivatives as input (only applicable to the inverse problem) |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | n_modes: int, |
| | n_layers: int, |
| | n_past: int, |
| | n_future: int, |
| | pde: PDE, |
| | data_channels: int, |
| | hidden_channels: int, |
| | use_partials: bool, |
| | batch_size: int, |
| | ): |
| |
|
| | if use_partials: |
| | |
| | |
| | |
| | out_channels = hidden_channels |
| |
|
| | else: |
| | out_channels = hidden_channels |
| |
|
| | |
| |
|
| | self.use_partials = use_partials |
| |
|
| | in_channels = resolve_number_input_channels( |
| | n_past=n_past, |
| | data_channels=data_channels, |
| | use_partials=use_partials, |
| | pde=pde, |
| | ) |
| |
|
| | |
| | n_modes = [n_modes] * PDE_NUM_SPATIAL[pde] |
| | self.batch_size = batch_size |
| | super(FNOEncoder, self).__init__( |
| | n_modes=n_modes, |
| | n_layers=n_layers, |
| | in_channels=in_channels, |
| | hidden_channels=hidden_channels, |
| | out_channels=out_channels, |
| | ) |
| |
|
| | def forward(self, x, **kwargs): |
| | return super().forward(x) |
| |
|
| |
|
| | class Resnet(nn.Module): |
| | """ |
| | Wrapper around FNO replacing FNO convolution blocks with Resnet Blocks. |
| | """ |
| |
|
| | def __init__(self, *args, **kwargs): |
| | |
| | super(Resnet, self).__init__() |
| | self.in_channels = kwargs["in_channels"] |
| | self.hidden_channels = kwargs["hidden_channels"] |
| | self.n_layers = kwargs["n_layers"] |
| | self.batch_size = kwargs["batch_size"] |
| | self.in_block = BasicBlock( |
| | inplanes=self.in_channels, |
| | planes=self.hidden_channels, |
| | stride=1, |
| | downsample=None, |
| | groups=1, |
| | base_width=64, |
| | dilation=1, |
| | norm_layer=nn.BatchNorm2d, |
| | ) |
| | self.in_block = nn.Sequential( |
| | nn.Conv2d( |
| | self.in_channels, |
| | self.hidden_channels, |
| | kernel_size=(3, 3), |
| | stride=(1, 1), |
| | padding=(1, 1), |
| | bias=False, |
| | ), |
| | nn.BatchNorm2d( |
| | self.hidden_channels, |
| | eps=1e-05, |
| | momentum=0.1, |
| | affine=True, |
| | track_running_stats=True, |
| | ), |
| | nn.ReLU(inplace=True), |
| | nn.Conv2d( |
| | self.hidden_channels, |
| | self.hidden_channels, |
| | kernel_size=(3, 3), |
| | stride=(1, 1), |
| | padding=(1, 1), |
| | bias=False, |
| | ), |
| | nn.BatchNorm2d( |
| | self.hidden_channels, |
| | eps=1e-05, |
| | momentum=0.1, |
| | affine=True, |
| | track_running_stats=True, |
| | ), |
| | ) |
| | self.resnet_blocks = nn.ModuleList( |
| | [ |
| | BasicBlock( |
| | inplanes=self.hidden_channels, |
| | planes=self.hidden_channels, |
| | stride=1, |
| | downsample=None, |
| | groups=1, |
| | base_width=64, |
| | dilation=1, |
| | norm_layer=nn.BatchNorm2d, |
| | ) |
| | for _ in range(kwargs["n_layers"]) |
| | ] |
| | ) |
| |
|
| | def forward(self, x, output_shape=None, **kwargs): |
| | """CN-Resnet's forward pass |
| | |
| | Parameters |
| | ---------- |
| | x : tensor |
| | input tensor |
| | output_shape : {tuple, tuple list, None}, default is None |
| | Gives the option of specifying the exact output shape for odd shaped inputs. |
| | * If None, don't specify an output shape |
| | * If tuple, specifies the output-shape of the **last** FNO Block |
| | * If tuple list, specifies the exact output-shape of each FNO Block |
| | """ |
| | x = self.in_block(x) |
| | for layer_idx in range(self.n_layers): |
| | x = self.resnet_blocks[layer_idx](x) |
| |
|
| | return x |
| |
|
| |
|
| | class ResnetEncoder(Resnet): |
| | """ |
| | Wrapper around Resnet that figures out the input channels based |
| | on the number of past frames and partial derivatives. |
| | :param n_layers: Number of layers in the Resnet. |
| | :param n_past: Number of past frames to use. |
| | :param pde: PDE to use for the partial derivatives. |
| | :param data_channels: Number of channels per timestep in the native input data. |
| | :param hidden_channels: Number of channels in the hidden layers. |
| | :param use_partials: Whether to use partial derivatives as input (only applicable to the inverse problem) |
| | :param mode: One of "oneshot", "autoregressive", "grid_to_soln" |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | n_layers: int, |
| | n_past: int, |
| | n_future: int, |
| | pde: PDE, |
| | data_channels: int, |
| | hidden_channels: int, |
| | use_partials: bool, |
| | batch_size: int, |
| | ): |
| |
|
| | |
| | if use_partials: |
| | |
| | |
| | |
| | out_channels = hidden_channels |
| |
|
| | else: |
| | out_channels = hidden_channels |
| |
|
| | self.use_partials = use_partials |
| | self.pde = pde |
| | in_channels = resolve_number_input_channels( |
| | n_past=n_past, |
| | data_channels=data_channels, |
| | use_partials=use_partials, |
| | pde=pde, |
| | ) |
| |
|
| | super(ResnetEncoder, self).__init__( |
| | n_layers=n_layers, |
| | in_channels=in_channels, |
| | hidden_channels=hidden_channels, |
| | out_channels=out_channels, |
| | batch_size=batch_size, |
| | pde=pde, |
| | ) |
| |
|
| | def forward(self, x, **kwargs): |
| | if self.pde == PDE.KortewegDeVries1D: |
| | x = x.unsqueeze(2) |
| | x = super().forward(x) |
| | return x[:, :, 0, :] |
| | return super().forward(x) |
| |
|
| |
|
| | class ScOTEncoder(nn.Module): |
| | config: ScOTConfig |
| | backbone: ScOT |
| | n_past: int |
| | in_channels: int |
| | use_partials: bool |
| | patch_size: int |
| | padding_mode: str = "constant" |
| |
|
| | def __init__( |
| | self, |
| | |
| | embed_dim: int, |
| | n_layers: int, |
| | hidden_size: int, |
| | patch_size: int, |
| | num_heads: list[int], |
| | skip_connections: list[int], |
| | depths: list[int], |
| | |
| | use_partials: bool, |
| | data_channels: bool, |
| | n_past: int, |
| | pde: PDE, |
| | **kwargs, |
| | ): |
| | super(ScOTEncoder, self).__init__() |
| |
|
| | self.n_past = n_past |
| | self.use_partials = use_partials |
| | self.patch_size = patch_size |
| | self.in_channels = resolve_number_input_channels( |
| | n_past=self.n_past, |
| | use_partials=self.use_partials, |
| | data_channels=data_channels, |
| | pde=pde, |
| | ) |
| |
|
| | |
| | self.spatial_size = PDE_SPATIAL_SIZE[pde][0] |
| | self.pde = pde |
| | self.config = ScOTConfig( |
| | num_layers=n_layers, |
| | num_channels=self.in_channels, |
| | num_out_channels=hidden_size, |
| | depths=depths, |
| | num_heads=num_heads, |
| | skip_connections=skip_connections, |
| | patch_size=self.patch_size, |
| | embed_dim=embed_dim, |
| | image_size=self.spatial_size, |
| | **kwargs, |
| | ) |
| |
|
| | self.backbone = ScOT(self.config) |
| |
|
| | def _pad_input( |
| | self, x: Float[torch.Tensor, "batch channels nx ny"] |
| | ) -> tuple[ |
| | Float[torch.Tensor, "batch channels nx ny"], tuple[int, int, int, int] | None |
| | ]: |
| | _, _, nx, ny = x.shape |
| | assert nx == ny, f"Non-square solutions not supported nx={nx}, ny={ny}" |
| | total_pad: int = (self.patch_size - (nx % self.patch_size)) % self.patch_size |
| | left_pad, right_pad = total_pad // 2, ceil(total_pad / 2) |
| | assert ( |
| | left_pad + right_pad == total_pad |
| | ), f"Incorrect swin padding {left_pad} + {right_pad} = {total_pad}" |
| | if left_pad or right_pad: |
| | pad_vals = (left_pad, right_pad, left_pad, right_pad) |
| | return ( |
| | torch.nn.functional.pad( |
| | x, |
| | pad_vals, |
| | mode=self.padding_mode, |
| | value=0, |
| | ), |
| | pad_vals, |
| | ) |
| | return x, None |
| |
|
| | def forward( |
| | self, |
| | x: Float[torch.Tensor, "batch channels nx ny"], |
| | t: Float[torch.Tensor, "batch nt"] | None = None, |
| | ) -> Float[torch.Tensor, "batch outdim nx ny"]: |
| | |
| | if self.pde == PDE.KortewegDeVries1D: |
| | x = x.unsqueeze(2) |
| | x = x.repeat(1, 1, x.shape[-1], 1) |
| | x, pad_vals = self._pad_input(x) |
| | output: ScOTOutput = self.backbone(pixel_values=x, time=t).output |
| | if pad_vals: |
| | |
| | l, r, _, _ = pad_vals |
| | output = output[..., l:-r, l:-r] |
| | if self.pde == PDE.KortewegDeVries1D: |
| | output = output[:, :, :1, :] |
| | return output |
| |
|