| import glob |
| import logging |
| import math |
|
|
| import h5py |
| import numpy as np |
| import torch |
| from scipy import signal |
| from torch.utils.data import Dataset |
|
|
| from pdeinvbench.data.transforms import collapse_time_and_channels_torch_transform |
| from pdeinvbench.data.utils import extract_params_from_path |
| from pdeinvbench.utils.types import PDE, PDE_NUM_SPATIAL, PDE_TRAJ_LEN |
|
|
|
|
| class PDE_MultiParam(Dataset): |
| """Data Loader that loads the multiple parameter version of PDE Datasets.""" |
|
|
| def __init__( |
| self, |
| data_root: str, |
| pde: PDE, |
| n_past: int, |
| dilation: int, |
| cutoff_first_n_frames: int, |
| train: bool, |
| frac_param_combinations: float = 1, |
| frac_ics_per_param: float = 1, |
| random_sample_param: bool = True, |
| downsample_factor: int = 0, |
| every_nth_window: int = 1, |
| window_start_percent: float = 0.0, |
| window_end_percent: float = 1.0, |
| ): |
| """ |
| Args: |
| data_root: path containing the h5 files for the current data split |
| pde: name of the PDE system - one of the enum values. |
| n_past: number of conditioning frames |
| dilation: frequency at which to subsample the ground truth trajectories in the time dimension |
| cutoff_first_n_frames: number of initial frames to cutoff in each trajectory (may want to do this e.g. if initial PDE residuals are very high) |
| train: if training dataloader, windows are randomly sampled from each trajecory, if non-training dataloader we loop through all non-overlapping windows |
| frac_param_combinations: fraction of parameter combinations to use. 1 takes all parameters. "0.x" takes x percent of total parameters |
| frac_ics_per_param: fraction of initial conditions per parameter combination to keep. |
| random_sample_param: (bool) If frac_param_combinations < 1, true means we randomly sample params and false means we grab the first n_frac params. Defaults to true. |
| downsample_factor: downsample a solution field spatially by the 'downsample_factor'. eg if downsample_factor=4, sol field spatial size=[128,128] --downsample--> final spatial size = [32,32] |
| every_nth_window: take every nth window from the list of non-over-lapping windows |
| window_start_percent: percent of the way through the trajectory to start the window after cutoff_first_n_frames |
| window_end_percent: percent of the way through the trajectory to end the window |
| """ |
|
|
| self.data_root = data_root |
| self.pde = pde |
| self.n_past = n_past |
| self.dilation = dilation |
| self.cutoff_first_n_frames = cutoff_first_n_frames |
| self.frac_param_combinations = frac_param_combinations |
| self.frac_ics_per_param = frac_ics_per_param |
| self.random_sample_param = random_sample_param |
| self.train = train |
| self.every_nth_window = every_nth_window |
| assert ( |
| window_start_percent < window_end_percent |
| ), "window_start_percent must be less than window_end_percent" |
| self.window_start_index = int( |
| (PDE_TRAJ_LEN[self.pde] - self.cutoff_first_n_frames) * window_start_percent |
| + self.cutoff_first_n_frames |
| ) |
| self.window_end_index = int( |
| (PDE_TRAJ_LEN[self.pde] - self.cutoff_first_n_frames) * window_end_percent |
| + self.cutoff_first_n_frames |
| ) |
| self.total_trajectory_length = self.window_end_index - self.window_start_index |
|
|
| if self.train: |
| self.num_windows = self.total_trajectory_length - self.n_past - 1 |
| else: |
| self.num_windows = (self.total_trajectory_length) // ( |
| (self.n_past) * self.every_nth_window |
| ) |
|
|
| if self.num_windows == 0 and self.every_nth_window > 1: |
| self.every_nth_window = 1 |
| self.num_windows = (self.total_trajectory_length) // ((self.n_past)) |
|
|
| |
| if self.pde == PDE.DarcyFlow2D: |
| self.num_windows = 1 |
|
|
| self.downsample_factor = downsample_factor |
|
|
| if PDE_NUM_SPATIAL[pde] == 2: |
| self.transforms = [collapse_time_and_channels_torch_transform] |
| else: |
| self.transforms = None |
|
|
| |
| |
| self.h5_paths = glob.glob(f"{self.data_root}/*.h5") |
| if len(self.h5_paths) == 0: |
| self.h5_paths = glob.glob(f"{self.data_root}/*.hdf5") |
| if self.pde == PDE.DarcyFlow2D: |
| self.h5_files = [file for file in self.h5_paths] |
| else: |
| self.h5_files = [h5py.File(file, "r") for file in self.h5_paths] |
|
|
| |
| if self.pde == PDE.ReactionDiffusion2D or self.pde == PDE.TurbulentFlow2D: |
| self.seqs = [list(h5_file.keys()) for h5_file in self.h5_files] |
| elif self.pde == PDE.NavierStokes2D: |
| |
| self.seqs = [h5_file["solutions"] for h5_file in self.h5_files] |
| elif self.pde == PDE.KortewegDeVries1D: |
| self.seqs = [h5_file["tensor"] for h5_file in self.h5_files] |
| elif self.pde == PDE.DarcyFlow2D: |
| |
| self.seqs = [file for file in self.h5_paths] |
| else: |
| self.seqs = [h5py.File(file, "r") for file in self.h5_paths] |
| if self.frac_param_combinations < 1: |
| total_params = math.ceil(len(self.seqs) * self.frac_ics_per_param) |
|
|
| logging.info( |
| f"trimming dataset from length {len(self.seqs)} to {total_params}" |
| ) |
| if self.random_sample_param: |
| |
| |
| assert len(self.seqs) == len(self.h5_paths) and len( |
| self.h5_paths |
| ) == len( |
| self.h5_files |
| ), f"The dataloader variables are mismatched. seqs = {len(self.seqs)}, h5_paths = {len(self.h5_paths)}, h5_files = {len(self.h5_files)}" |
|
|
| |
| |
| num_sequences: int = len(self.seqs) |
| requested_dataset_size: int = int( |
| num_sequences * self.frac_param_combinations |
| ) |
| indices = np.arange(num_sequences) |
| sample_seed: int = 42 |
| rng_generator = np.random.default_rng(seed=sample_seed) |
| sampled_indices = rng_generator.choice( |
| indices, size=requested_dataset_size, replace=False |
| ) |
| logging.info( |
| f"Using random sampling to trim the dataset down from length {len(self.seqs)} to {requested_dataset_size}" |
| ) |
| assert ( |
| len(set(sampled_indices.tolist())) == sampled_indices.shape[0] |
| ), f"Duplicate items in random sampling of PDE parameters!" |
| assert ( |
| sampled_indices.shape[0] == requested_dataset_size |
| ), f"Mismatch between the requested dataset sample size and the new sampled dataset. frac requested = {self.frac_param_combinations}, requested size = {requested_dataset_size}, new size = {sampled_indices.shape[0]}" |
| self.seqs = [self.seqs[i] for i in sampled_indices] |
| self.h5_paths = [self.h5_paths[i] for i in sampled_indices] |
| self.h5_files = [self.h5_files[i] for i in sampled_indices] |
| else: |
| self.seqs = self.seqs[:total_params] |
| self.h5_paths = self.h5_paths[:total_params] |
| self.h5_files = self.h5_files[:total_params] |
|
|
| self.num_params = len(self.seqs) |
| if self.pde == PDE.KortewegDeVries1D: |
| |
| self.num_ics_per_param = self.seqs[0].shape[0] |
| elif self.pde == PDE.DarcyFlow2D: |
| self.num_ics_per_param = 1 |
| elif self.pde != PDE.NavierStokes2D: |
| self.num_ics_per_param = len( |
| min([self.seqs[i] for i in range(len(self.seqs))]) |
| ) |
| else: |
| self.num_ics_per_param = min( |
| [self.seqs[i].shape[0] for i in range(len(self.seqs))] |
| ) |
|
|
| |
|
|
| self.num_ics_per_param = math.ceil( |
| self.num_ics_per_param * self.frac_ics_per_param |
| ) |
| |
| if pde == PDE.ReactionDiffusion2D or pde == PDE.TurbulentFlow2D: |
| self.x = self.h5_files[0]["0001"]["grid"]["x"][:] |
| self.y = self.h5_files[0]["0001"]["grid"]["y"][:] |
| self.t = torch.Tensor(self.h5_files[0]["0001"]["grid"]["t"][:]) |
| elif pde == PDE.NavierStokes2D: |
| self.x = self.h5_files[0]["x-coordinate"][:] |
| self.y = self.h5_files[0]["y-coordinate"][:] |
| self.t = torch.Tensor(self.h5_files[0]["t-coordinate"][:]) |
| elif pde == PDE.DarcyFlow2D: |
| |
| domain_len = 1 |
| d = h5py.File(self.seqs[0], "r") |
| size, _, _ = d["sol"].shape |
| d.close() |
| x = np.linspace(0, domain_len, size, endpoint=False) |
| self.x = torch.from_numpy(x) |
| self.y = torch.from_numpy(x) |
| self.t = ( |
| torch.ones(10, dtype=float) * -1 |
| ) |
| else: |
| |
| self.y = None |
| self.x = self.h5_files[0]["x-coordinate"][:] |
| self.t = torch.Tensor(self.h5_files[0]["t-coordinate"][:]) |
|
|
| if self.downsample_factor != 0: |
| self.y = ( |
| None |
| if self.y is None |
| else signal.decimate(self.y, q=self.downsample_factor, axis=0).copy() |
| ) |
| self.x = signal.decimate(self.x, q=self.downsample_factor, axis=0).copy() |
| self.x = torch.Tensor(self.x) |
| self.y = torch.Tensor(self.y) if self.y is not None else None |
|
|
| logging.info( |
| f"Initialized dataset with {self.num_params} parameter combinations" |
| ) |
|
|
| def __len__(self): |
| """ |
| Number of parameters * number of ICs = number of full trajectories. |
| """ |
| if self.train: |
| return self.num_params * self.num_ics_per_param |
| else: |
| return self.num_params * self.num_ics_per_param * self.num_windows |
|
|
| def __getitem__(self, index: int): |
| """ |
| Loops over all parameters and ICs, and randomly samples time windows. |
| Returns: |
| x: conditioning frames, shape of [n_past, spatial/channel dims] |
| y: target frame(s), shape of [n_future, spatial/channel dims] |
| param_dict: dictionary containing the true PDE parameter for the trajectory. |
| """ |
| |
| if self.train: |
| param_index = index // self.num_ics_per_param |
| ic_index = index % self.num_ics_per_param |
| else: |
| |
| |
| param_index = index // (self.num_ics_per_param * self.num_windows) |
| ic_index = (index // self.num_windows) % self.num_ics_per_param |
| window_index = index % self.num_windows |
| |
| h5_file = self.h5_files[param_index] |
| h5_path = self.h5_paths[param_index] |
| param_dict = extract_params_from_path(h5_path, self.pde) |
|
|
| if self.pde == PDE.ReactionDiffusion2D or self.pde == PDE.TurbulentFlow2D: |
| |
| seq = self.seqs[param_index][ic_index] |
| traj = torch.Tensor( |
| np.array(h5_file[f"{seq}/data"], dtype="f") |
| ) |
| elif self.pde == PDE.NavierStokes2D: |
| seq = self.seqs[param_index] |
| traj = torch.Tensor(seq[ic_index]) |
| |
|
|
| elif self.pde == PDE.DarcyFlow2D: |
| |
| |
| seq = h5py.File(self.seqs[param_index], "r") |
|
|
| coeff = torch.from_numpy(np.asarray(seq["coeff"])) |
| coeff = torch.squeeze(coeff) |
| coeff = torch.unsqueeze(coeff, dim=0) |
| |
| min_val = coeff.min() |
| max_val = coeff.max() |
| |
| coeff = coeff - min_val |
| binary_mask = coeff > 0 |
|
|
| def wrap_scalar(x): |
| return torch.Tensor([x.item()]) |
|
|
| param_dict["coeff"] = binary_mask.float() |
| param_dict["max_val"] = wrap_scalar(max_val) |
| param_dict["min_val"] = wrap_scalar(min_val) |
| traj = torch.from_numpy(np.asarray(seq["sol"])) |
| seq.close() |
| else: |
| seq = self.seqs[param_index] |
| traj = torch.Tensor(np.array(h5_file["tensor"][ic_index])) |
| traj = traj[:: self.dilation] |
|
|
| |
| if traj.shape[0] - self.n_past == 0: |
| start = 0 |
| |
| if self.n_past == 1: |
| raise ValueError("Problem is ill-posed when n_past == 1. ") |
| else: |
| if self.train: |
| start = np.random.randint( |
| self.window_start_index, |
| self.window_end_index - self.n_past, |
| ) |
| else: |
| |
| start = self.window_start_index + ( |
| window_index * (self.n_past) * self.every_nth_window |
| ) |
|
|
| if self.pde != PDE.DarcyFlow2D: |
| traj = traj[start : start + self.n_past] |
| time_frames = self.t[start : start + self.n_past] |
| else: |
| time_frames = -1 * torch.ones(self.n_past, dtype=float) |
| |
| if len(traj.shape) == 4: |
| |
| traj = traj.permute((0, 3, 1, 2)) |
|
|
| if self.downsample_factor != 0: |
| traj = signal.decimate(traj, q=self.downsample_factor, axis=-1) |
| traj = ( |
| torch.Tensor( |
| signal.decimate(traj, q=self.downsample_factor, axis=-2).copy() |
| ) |
| if len(traj.shape) == 4 |
| else torch.Tensor(traj.copy()) |
| ) |
|
|
| |
| if self.pde == PDE.DarcyFlow2D: |
| |
| |
| |
| traj = torch.squeeze(traj) |
| traj = torch.unsqueeze(traj, dim=0) |
| traj = torch.unsqueeze(traj, dim=0) |
| x, y = ( |
| traj, |
| traj, |
| ) |
| x = x.float() |
| y = y.float() |
| else: |
| x, y = torch.split(traj, [self.n_past, 0], dim=0) |
|
|
| if self.transforms is not None: |
| |
| for T in self.transforms: |
| x, y, param_dict = T((x, y, param_dict)) |
|
|
| |
| spatial_grid = (self.x, self.y) if self.y is not None else (self.x,) |
|
|
| ic_index = torch.tensor([ic_index], dtype=float) |
| return spatial_grid, self.t, x, y, time_frames, ic_index, param_dict |
|
|