|
--- |
|
dataset_info: |
|
- config_name: large_100 |
|
features: |
|
- name: lrs |
|
sequence: |
|
array4_d: |
|
shape: |
|
- 3 |
|
- 16 |
|
- 16 |
|
- 16 |
|
dtype: float32 |
|
- name: hr |
|
dtype: |
|
array4_d: |
|
shape: |
|
- 3 |
|
- 64 |
|
- 64 |
|
- 64 |
|
dtype: float32 |
|
splits: |
|
- name: train |
|
num_bytes: 268237120 |
|
num_examples: 80 |
|
- name: validation |
|
num_bytes: 33529640 |
|
num_examples: 10 |
|
- name: test |
|
num_bytes: 33529640 |
|
num_examples: 10 |
|
download_size: 329464088 |
|
dataset_size: 335296400 |
|
- config_name: large_50 |
|
features: |
|
- name: lrs |
|
sequence: |
|
array4_d: |
|
shape: |
|
- 3 |
|
- 16 |
|
- 16 |
|
- 16 |
|
dtype: float32 |
|
- name: hr |
|
dtype: |
|
array4_d: |
|
shape: |
|
- 3 |
|
- 64 |
|
- 64 |
|
- 64 |
|
dtype: float32 |
|
splits: |
|
- name: train |
|
num_bytes: 134118560 |
|
num_examples: 40 |
|
- name: validation |
|
num_bytes: 16764820 |
|
num_examples: 5 |
|
- name: test |
|
num_bytes: 16764820 |
|
num_examples: 5 |
|
download_size: 164732070 |
|
dataset_size: 167648200 |
|
- config_name: small_50 |
|
features: |
|
- name: lrs |
|
sequence: |
|
array4_d: |
|
shape: |
|
- 3 |
|
- 4 |
|
- 4 |
|
- 4 |
|
dtype: float32 |
|
- name: hr |
|
dtype: |
|
array4_d: |
|
shape: |
|
- 3 |
|
- 16 |
|
- 16 |
|
- 16 |
|
dtype: float32 |
|
splits: |
|
- name: train |
|
num_bytes: 2220320 |
|
num_examples: 40 |
|
- name: validation |
|
num_bytes: 277540 |
|
num_examples: 5 |
|
- name: test |
|
num_bytes: 277540 |
|
num_examples: 5 |
|
download_size: 2645696 |
|
dataset_size: 2775400 |
|
--- |
|
|
|
|
|
# Super-resolution of Velocity Fields in Three-dimensional Fluid Dynamics |
|
|
|
This dataset loader attempts to reproduce the data of Wang et al. (2024)'s experiments on Super-resolution of 3D Turbulence. |
|
|
|
References: |
|
- Wang et al. (2024): "Discovering Symmetry Breaking in Physical Systems with Relaxed Group Convolution" |
|
|
|
## Usage |
|
|
|
For a given configuration (e.g. `large_50`): |
|
|
|
```py |
|
>>> ds = datasets.load_dataset("dl2-g32/jhtdb", name="large_50") |
|
>>> ds |
|
DatasetDict({ |
|
train: Dataset({ |
|
features: ['lrs', 'hr'], |
|
num_rows: 40 |
|
}) |
|
validation: Dataset({ |
|
features: ['lrs', 'hr'], |
|
num_rows: 5 |
|
}) |
|
test: Dataset({ |
|
features: ['lrs', 'hr'], |
|
num_rows: 5 |
|
}) |
|
}) |
|
``` |
|
|
|
Each split contains the input `lrs` which corresponds on a sequence of low resolution samples from time `t - ws/2, ..., t, ... ts + ws/2` (ws = window size) and `hr` corresponds to the high resolution sample at time `t`. All the parameters per data point are specified in the corresponding `metadata_*.csv`. |
|
|
|
Specifically, for the default configuration, for each datapoint we have `3` low resolution samples and `1` high resolution sample. Each of the former have shapes `(3, 16, 16, 16)` and the latter has shape `(3, 64, 64, 64)`. |
|
|
|
## Replication |
|
|
|
This dataset is entirely generated by `scripts/generate.py` and each configuration is fully specified in their corresponding `scripts/*.yaml`. |
|
|
|
### Usage |
|
|
|
```sh |
|
python -m scripts.generate --config scripts/small_100.yaml --token edu.jhu.pha.turbulence.testing-201311 |
|
``` |
|
|
|
This will create two folders on `datasets/jhtdb`: |
|
1. A `tmp` folder that will store all samples accross runs to serve as a cache. |
|
2. The corresponding subset, `small_50` for example. This folder will contain a `metadata_*.csv` and data `*.zip` for each split. |
|
|
|
Note: |
|
- For the small variants, the default token is enough, but for the large variants a token has to be requested. More details [here](https://turbulence.pha.jhu.edu/authtoken.aspx). |
|
- For reference, the `large_100` takes ~15 minutes to generate for a total of ~300MB. |
|
|