| | import os |
| | from typing import Callable, Dict, List, Tuple, Union |
| |
|
| | import lightning as L |
| | import torch |
| | import typeguard |
| | from functorch.dim import tree_map |
| | from jaxtyping import Float, jaxtyped |
| | from lightning.pytorch.utilities import grad_norm |
| | from pdeinvbench.utils.pytorch_utils import compute_grad_norm |
| | from lightning.pytorch.loggers import WandbLogger |
| | from pdeinvbench.data.utils import unnormalize_params |
| | from pdeinvbench.utils.logging_utils import get_best_model_weights |
| | from copy import deepcopy |
| | import time |
| |
|
| | from pdeinvbench.losses import ( |
| | get_param_metric, |
| | get_pde_residual_function, |
| | pde_residual_reduction, |
| | ) |
| | from pdeinvbench.losses.metrics import ( |
| | classification_metrics_darcy_flow, |
| | param_relative_loss, |
| | ) |
| | from pdeinvbench.models.inverse_model import InverseModel |
| | from pdeinvbench.utils import pytorch_utils as ptu |
| | from pdeinvbench.utils.types import ( |
| | PDE, |
| | DataMetrics, |
| | ParamMetrics, |
| | TypeBatch, |
| | TypeBatch1D, |
| | TypeBatch2D, |
| | TypeLossDict, |
| | TypeParam, |
| | TypePredict1D, |
| | TypePredict2D, |
| | ) |
| |
|
| |
|
| | class InverseModule(L.LightningModule): |
| | """ |
| | Inverse Problem Module. Takes a set of conditioning frames from a PDE trajectory and predicts the value of the PDE parameter(s). |
| | :param model: Model. |
| | :param optimizer: Optimizer. |
| | :param lr_scheduler: Learning rate scheduler. |
| | :param pde: enum of the PDE to use for the residual calculation. |
| | :param param_loss_metric: Metric to use for the parameter loss. |
| | :param inverse_residual_loss_weight: Weight for the PDE residual loss obtained from the predicted parameters. |
| | :param n_past: Number of past frames to condition on. |
| | :param use_partials: whether to append the partial derivatives to the input. |
| | :param params_to_predict: list of strings of parameters to predict. |
| | :param residual_filter: Whether to use residual filtering. |
| | :param batch_size: Batch size. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | model: InverseModel, |
| | optimizer: torch.optim.Optimizer, |
| | lr_scheduler: torch.optim.lr_scheduler, |
| | pde: PDE, |
| | param_loss_metric: ParamMetrics, |
| | inverse_residual_loss_weight: float, |
| | inverse_param_loss_weight: float, |
| | n_past: int, |
| | use_partials: bool, |
| | params_to_predict: List[str], |
| | residual_filter: bool, |
| | batch_size: int, |
| | ): |
| | super().__init__() |
| | self.save_hyperparameters(ignore=["model"]) |
| | self.model = model |
| | self.old_inverse_model = None |
| | self.new_inverse_model = None |
| | self.pde_residual = get_pde_residual_function(pde) |
| | self.pde = pde |
| | self.param_loss_metric = get_param_metric(param_loss_metric) |
| | self.inverse_residual_loss_weight = inverse_residual_loss_weight |
| | self.inverse_param_loss_weight = inverse_param_loss_weight |
| |
|
| | self.validation_step_outputs = [] |
| | self.validation_step_targets = [] |
| | """ |
| | Loss dicts for validation. Each element is a tuple of (losses, pde_params). |
| | PDE_params comes directly from the dataloader. |
| | Loss: Dict[str, torch.Tensor] with keys (shape) |
| | 'data_loss' (B), 'residual_loss' (B), 'loss' (), 'data_loss_per_batch_element' (B), 'residual_per_batch_element' (B). |
| | """ |
| | self.validation_step_loss_dicts = [] |
| | self.n_past = n_past |
| | self.use_partials = use_partials |
| | self.params_to_predict = params_to_predict |
| | self.optimizer = optimizer |
| | self.lr_scheduler = lr_scheduler |
| | self.residual_filter = residual_filter |
| | self.batch_size = batch_size |
| |
|
| | @jaxtyped(typechecker=typeguard.typechecked) |
| | def predict_with_loss( |
| | self, |
| | batch: Union[TypeBatch1D, TypeBatch2D], |
| | stage: str, |
| | gumbel: bool = False, |
| | ) -> Tuple[TypeParam, TypeParam, Dict[str, Union[torch.tensor, TypeBatch]]]: |
| | """ |
| | Common method which computes a prediction of the PDE parameter from the conditioning frames and returns the loss. |
| | :param batch: Batch of data. |
| | :param stage: Which of "train", "validation", "test","tailoring" is the current stage. |
| | """ |
| | _, _, initial_frames, _, _, _, _ = batch |
| | start_time = time.time() |
| | |
| | |
| | |
| | num_repeat_elements = 0 |
| | if initial_frames.shape[0] < self.batch_size: |
| | num_repeat_elements = self.batch_size - initial_frames.shape[0] |
| | batch = tree_map( |
| | lambda x: torch.cat( |
| | [x] + [x[-1].unsqueeze(0) for _ in range(num_repeat_elements)] |
| | ), |
| | batch, |
| | ) |
| | spatial_grid, t, initial_frames, _, _, _, pde_params = batch |
| |
|
| | forward_pass_start_time = time.time() |
| | predicted_pde_params, residual, true_residual = self.model( |
| | initial_frames, pde_params, spatial_grid, t, gumbel=gumbel |
| | ) |
| | forward_pass_end_time = time.time() |
| | forward_pass_time = forward_pass_end_time - forward_pass_start_time |
| | if self.residual_filter: |
| | ids = [ |
| | true_residual.reshape(true_residual.shape[0], -1).max(dim=1)[0] < 100 |
| | ] |
| | residual = residual[ids] |
| | true_residual = true_residual[ids] |
| |
|
| | for key in pde_params: |
| | predicted_pde_params[key] = predicted_pde_params[key][ids] |
| | pde_params[key] = pde_params[key][ids] |
| |
|
| | residual_loss = pde_residual_reduction( |
| | residual |
| | ) |
| | true_residual_loss = pde_residual_reduction(true_residual) |
| | param_losses = self.param_loss_metric( |
| | predicted_pde_params, |
| | pde_params, |
| | reduction="mean", |
| | ) |
| |
|
| | relative_param_losses = param_relative_loss( |
| | {param: prediction for param, prediction in predicted_pde_params.items()}, |
| | pde_params, |
| | reduction="mean", |
| | ) |
| |
|
| | |
| | |
| | backprop_losses: list[torch.Tensor] |
| | if self.pde == PDE.DarcyFlow2D and stage != "tailoring": |
| | backprop_losses = [v for v in param_losses.values() if v.item() != 0] |
| | else: |
| | backprop_losses = list(param_losses.values()) |
| |
|
| | param_loss = torch.stack(backprop_losses).mean() |
| |
|
| | individual_param_losses = { |
| | f"param_loss_{k}": v for k, v in param_losses.items() |
| | } |
| | individual_relative_param_losses = { |
| | f"relative_param_loss_{k}": v for k, v in relative_param_losses.items() |
| | } |
| |
|
| | weighted_param_loss = self.inverse_param_loss_weight * param_loss |
| |
|
| | weighted_residual_loss = self.inverse_residual_loss_weight * residual_loss |
| |
|
| | loss = weighted_param_loss + weighted_residual_loss |
| | losses = { |
| | "param_loss": param_loss, |
| | "residual_loss": residual_loss, |
| | "true_residual_loss": true_residual_loss, |
| | "loss": loss, |
| | **individual_param_losses, |
| | **individual_relative_param_losses, |
| | } |
| |
|
| | if self.pde == PDE.DarcyFlow2D: |
| | |
| | darcy_losses: dict[str, float] = classification_metrics_darcy_flow( |
| | predicted_coeff=predicted_pde_params["coeff"], |
| | true_coeff=pde_params["coeff"], |
| | ) |
| | losses.update(darcy_losses) |
| |
|
| | |
| | if "validation" in stage or "test" in stage or "tailoring" in stage: |
| | |
| | |
| | param_loss_per_batch_element, residual_per_batch_element = ( |
| | self.stratify_losses( |
| | predicted_pde_params, |
| | pde_params, |
| | residual, |
| | ) |
| | ) |
| | losses["param_loss_per_batch_element"] = param_loss_per_batch_element |
| | losses["residual_per_batch_element"] = residual_per_batch_element |
| |
|
| | |
| | if num_repeat_elements > 0: |
| | predicted_pde_params = tree_map( |
| | lambda x: x[: -1 * num_repeat_elements], predicted_pde_params |
| | ) |
| | pde_params = tree_map(lambda x: x[: -1 * num_repeat_elements], pde_params) |
| | losses = tree_map( |
| | lambda x: x[: -1 * num_repeat_elements] if x.numel() > 1 else x, losses |
| | ) |
| | end_time = time.time() |
| | losses["timing_metrics"] = { |
| | "predict_with_loss_time": end_time - start_time, |
| | "forward_pass_time": forward_pass_time, |
| | } |
| | return predicted_pde_params, pde_params, losses |
| |
|
| | @jaxtyped(typechecker=typeguard.typechecked) |
| | def training_step( |
| | self, |
| | batch: Union[TypeBatch1D, TypeBatch2D], |
| | batch_idx: int, |
| | ) -> TypeLossDict: |
| | stage = "train" |
| | _, _, losses = self.predict_with_loss(batch, stage="train") |
| | self.log_metrics(stage, losses) |
| | return losses |
| |
|
| | @jaxtyped(typechecker=typeguard.typechecked) |
| | def validation_step( |
| | self, |
| | batch: Union[TypeBatch1D, TypeBatch2D], |
| | batch_idx: int, |
| | dataloader_idx: int = 0, |
| | stage: str = None, |
| | ) -> Union[None, TypeLossDict]: |
| | stage = "validation" if stage == None else stage |
| | prediction, target, losses = self.predict_with_loss(batch, stage=stage) |
| |
|
| | |
| |
|
| | |
| | if self.pde == PDE.DarcyFlow2D: |
| | for k in prediction.keys(): |
| | prediction[k] = prediction[k].cpu().numpy() |
| | target[k] = target[k].cpu().numpy() |
| | self.validation_step_outputs.append(prediction) |
| | self.validation_step_targets.append(target) |
| | self.validation_step_loss_dicts.append( |
| | ( |
| | ptu.torch_dict_to_numpy(losses), |
| | ptu.torch_dict_to_numpy(batch[-1]), |
| | ) |
| | ) |
| | else: |
| | self.validation_step_outputs.append(prediction) |
| | self.validation_step_targets.append(target) |
| | self.validation_step_loss_dicts.append((losses, batch[-1])) |
| | losses["predictions"] = prediction |
| | losses["targets"] = target |
| | self.log_metrics(stage, losses) |
| | return losses |
| |
|
| | @jaxtyped(typechecker=typeguard.typechecked) |
| | def test_step( |
| | self, |
| | batch: Union[TypeBatch1D, TypeBatch2D], |
| | batch_idx: int, |
| | dataloader_idx: int = 0, |
| | ) -> TypeLossDict: |
| | stage = "test" |
| | self.model.train() |
| | test_losses = self.validation_step(batch, batch_idx, dataloader_idx, stage) |
| | return test_losses |
| |
|
| | def log_metrics(self, stage, losses, prefix_dir=""): |
| | if "validation" in stage or "test" in stage: |
| | on_step = False |
| | on_epoch = True |
| | else: |
| | on_step = True |
| | on_epoch = False |
| |
|
| | main_loss_type = "param_loss" |
| |
|
| | self.log( |
| | os.path.join(stage, prefix_dir, main_loss_type), |
| | losses["param_loss"], |
| | prog_bar=True, |
| | on_step=on_step, |
| | on_epoch=on_epoch, |
| | ) |
| |
|
| | for param in self.params_to_predict: |
| | self.log( |
| | os.path.join(stage, prefix_dir, f"{main_loss_type}_{param}"), |
| | losses[f"param_loss_{param}"], |
| | prog_bar=True, |
| | on_step=on_step, |
| | on_epoch=on_epoch, |
| | ) |
| |
|
| | self.log( |
| | os.path.join(stage, prefix_dir, f"relative_param_loss_{param}"), |
| | losses[f"relative_param_loss_{param}"], |
| | prog_bar=True, |
| | on_step=on_step, |
| | on_epoch=on_epoch, |
| | ) |
| |
|
| | self.log( |
| | os.path.join(stage, prefix_dir, "residual_loss"), |
| | losses["residual_loss"], |
| | prog_bar=True, |
| | on_step=on_step, |
| | on_epoch=on_epoch, |
| | ) |
| |
|
| | self.log( |
| | os.path.join(stage, prefix_dir, "true_residual_loss"), |
| | losses["true_residual_loss"], |
| | prog_bar=True, |
| | on_step=on_step, |
| | on_epoch=on_epoch, |
| | ) |
| |
|
| | if isinstance(self.model, InverseModel): |
| |
|
| | |
| | if self.pde == PDE.DarcyFlow2D: |
| | self.log( |
| | os.path.join( |
| | stage, prefix_dir, "darcy_flow_classification_accuracy" |
| | ), |
| | losses["darcy_flow_classification_accuracy"], |
| | prog_bar=True, |
| | on_step=on_step, |
| | on_epoch=on_epoch, |
| | ) |
| | self.log( |
| | os.path.join(stage, prefix_dir, "darcy_flow_classification_iou"), |
| | losses["darcy_flow_classification_iou"], |
| | prog_bar=True, |
| | on_step=on_step, |
| | on_epoch=on_epoch, |
| | ) |
| |
|
| | self.log( |
| | os.path.join(stage, prefix_dir, "loss"), |
| | losses["loss"], |
| | on_step=on_step, |
| | on_epoch=on_epoch, |
| | ) |
| |
|
| | if "timing_metrics" in losses: |
| | self.log( |
| | os.path.join(stage, prefix_dir, "predict_with_loss_time"), |
| | losses["timing_metrics"]["predict_with_loss_time"], |
| | on_step=on_step, |
| | on_epoch=on_epoch, |
| | ) |
| |
|
| | self.log( |
| | os.path.join(stage, prefix_dir, "forward_pass_time"), |
| | losses["timing_metrics"]["forward_pass_time"], |
| | on_step=on_step, |
| | on_epoch=on_epoch, |
| | ) |
| |
|
| | def configure_optimizers(self): |
| | optimizer = self.optimizer |
| | lr_scheduler = self.lr_scheduler |
| | return [optimizer], [lr_scheduler] |
| |
|
| | def on_before_optimizer_step(self, optimizer): |
| | |
| |
|
| | self.old_inverse_model = torch.cat( |
| | [p.flatten() for p in self.model.parameters()] |
| | ) |
| | inverse_norms = grad_norm(self.model, 2) |
| | inverse_total_norm = sum(inverse_norms.values()) |
| | self.log( |
| | f"outer_loss_grad_norms/inverse_model", |
| | inverse_total_norm, |
| | on_step=True, |
| | ) |
| |
|
| | def optimizer_step(self, *args, **kwargs): |
| | super().optimizer_step(*args, **kwargs) |
| |
|
| | self.new_inverse_model = torch.cat( |
| | [p.flatten() for p in self.model.parameters()] |
| | ) |
| | param_diff_norm = torch.norm(self.new_inverse_model - self.old_inverse_model) |
| | original_param_norm = torch.norm(self.old_inverse_model) |
| | rel_diff = param_diff_norm / original_param_norm |
| | self.log( |
| | "outer_loss_grad_norms/callback_inverse_param_rel_diff", |
| | rel_diff, |
| | on_step=True, |
| | ) |
| |
|
| | def stratify_losses(self, predicted_pde_params, pde_params, residual): |
| | |
| | param_loss_per_batch_element = self.param_loss_metric( |
| | predicted_pde_params, pde_params, reduction="none" |
| | ) |
| |
|
| | |
| | if self.pde == PDE.DarcyFlow2D: |
| | coeff = param_loss_per_batch_element["coeff"] |
| | |
| | if len(coeff.shape) == 4: |
| | coeff = torch.flatten(coeff, start_dim=2).sum(axis=2) |
| |
|
| | param_loss_per_batch_element["coeff"] = coeff |
| |
|
| | param_loss_per_batch_element = ( |
| | torch.stack(list(param_loss_per_batch_element.values())) |
| | .squeeze(-1) |
| | .mean(dim=0) |
| | ) |
| | residual_per_batch_element = pde_residual_reduction( |
| | residual, dim=tuple(range(1, residual.dim())) |
| | ) |
| |
|
| | return param_loss_per_batch_element, residual_per_batch_element |
| |
|
| | def on_test_epoch_start(self): |
| | super().on_test_epoch_start() |
| |
|