windtunnel-20k / README.md
rvalerio's picture
Update README.md
cc63030 verified
|
raw
history blame
3.98 kB
metadata
pretty_name: Wind Tunnel 20K Dataset
size_categories:
  - 10K<n<100K
task_categories:
  - feature-extraction
  - graph-ml
  - image-to-3d
language:
  - en
tags:
  - simulation
  - openfoam
  - physics
  - windtunnel
  - inductiva
  - machine learning
  - synthetic

Wind Tunnel 20K Dataset

The Wind Tunnel Dataset contains 20,000 OpenFOAM simulations of 1,000 unique automobile-like objects placed in a virtual wind tunnel. Each object is simulated under 20 distinct conditions: 4 random wind speeds ranging from 10 to 50 m/s, and 5 rotation angles (0°, 180° and 3 random angles). To ensure stable and reliable results, each simulation runs for 300 iterations. The meshes for these automobile-like objects were generated using the Instant Mesh model and sourced from the Stanford Cars Dataset. The entire dataset of 20,000 simulations is organized into three subsets: 70% for training, 20% for validation, and 10% for testing.

The data generation process itself was orchestrated using the Inductiva API, which allowed us to run hundreds of OpenFOAM simulations in parallel on the cloud.

Dataset Structure

data
├── train
│   ├── <SIMULATION_ID>
│   │   ├── input_mesh.obj
│   │   ├── openfoam_mesh.obj
│   │   ├── pressure_field_mesh.vtk
│   │   ├── simulation_metadata.json
│   │   └── streamlines_mesh.ply
│   └── ...
├── validation
│   └── ...
└── test
    └── ...

Dataset Files

Each simulation in the Wind Tunnel Dataset is accompanied by several key files that provide both input and output data. Here’s a breakdown of the files included in each simulation:

  • input_mesh.obj: OBJ file with the input mesh.
  • openfoam_mesh.obj: OBJ file with the OpenFOAM mesh.
  • pressure_field_mesh.vtk: VTK file with the pressure field data.
  • streamlines_mesh.ply: PLY file with the streamlines.
  • metadata.json: JSON with metadata about the input parameters and about some output results such as the force coefficients (obtained via simulation) and the path of the output files.

Downloading the Dataset:

To download the dataset you have to install the Datasets package by HuggingFace:

pip install datasets

1. Using snapshot_download()

from huggingface_hub import snapshot_download

dataset_name = "inductiva/windtunnel"

# Download the entire dataset
snapshot_download(repo_id=dataset_name, repo_type="dataset")

# Download to a specific local directory
snapshot_download(repo_id=dataset_name, repo_type="dataset", local_dir="local_folder")

# Download only the simulation metadata across all simulations
snapshot_download(
    repo_id=dataset_name,
    repo_type="dataset",
    local_dir="local_folder",
    allow_patterns=["*/*/*/simulation_metadata.json"]
)

2. Using load_dataset()

from datasets import load_dataset

# Load the dataset (streaming is supported)
dataset = load_dataset("inductiva/windtunnel", streaming=False)

# Display dataset information
print(dataset)

# Access a sample from the training set
sample = dataset["train"][0]
print("Sample from training set:", sample)

What's next?

If you have any issues using this dataset, feel free to reach out to us at support@intuctiva.ai

To learn more about how we created this dataset—or how you can generate synthetic datasets for Physics-AI models—visit Inductiva.AI or check out our blog post on transforming complex simulation workflows into easy-to-use Python classes.