Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
Cd
float64
0.24
0.34
Cl
float64
-0.16
0.22
Clf
float64
-0.24
0.05
Clr
float64
-0.06
0.25
Cs
float64
-0.02
0.06
0.303512
0.067728
-0.037286
0.105014
0.047668
0.24024
-0.073912
-0.15419
0.080278
0.007835
0.292211
0.147656
-0.02159
0.169246
0.024768
0.278873
0.048312
-0.023833
0.072145
0.030148
0.248894
-0.041229
-0.168807
0.127578
-0.002965
0.279117
0.143125
-0.02998
0.173105
0.021352
0.290326
-0.015452
-0.163937
0.148485
0.026089
0.261619
-0.023936
-0.073291
0.049355
0.01202
0.259749
-0.011651
-0.152916
0.141265
0.03926
0.297232
0.114824
-0.012443
0.127267
0.034604
0.260315
0.072336
-0.063125
0.135462
0.01412
0.315883
0.119675
-0.024812
0.144487
0.04995
0.296761
0.036551
-0.079722
0.116273
0.054041
0.277838
0.061573
-0.043283
0.104856
0.026641
0.34013
0.137444
-0.045533
0.182976
0.037298
0.261434
-0.035261
-0.092611
0.05735
0.005883
0.276181
-0.022107
-0.083419
0.061312
0.017143
0.280791
0.08221
-0.041294
0.123504
0.03933
0.264904
0.050436
-0.096866
0.147302
0.014086
0.262902
0.045324
-0.071357
0.116682
0.010345
0.275042
-0.098488
-0.138631
0.040143
0.009684
0.26212
0.014556
-0.082416
0.096971
0.017591
0.297282
0.115384
0.004209
0.111175
0.028315
0.299789
0.044083
-0.055367
0.09945
0.03691
0.243657
-0.00672
-0.163568
0.156848
0.02098
0.2711
-0.093687
-0.151173
0.057487
0.00438
0.293788
0.033351
-0.102321
0.135672
0.045973
0.28911
0.12005
-0.072846
0.192896
0.033567
0.259263
0.022341
-0.063306
0.085647
0.003594
0.305833
0.092528
-0.013412
0.10594
0.049858
0.27295
-0.042545
-0.157582
0.115037
0.019205
0.274753
0.061941
-0.053633
0.115573
0.020776
0.245122
-0.110359
-0.137736
0.027377
0.006989
0.300113
0.068393
-0.06885
0.137242
0.049451
0.289781
0.057686
-0.042179
0.099866
0.021351
0.26306
-0.042095
-0.126927
0.084832
0.015393
0.294032
0.169134
0.022295
0.146839
0.019896
0.267345
-0.038379
-0.135259
0.096881
0.02315
0.275355
-0.077
-0.177165
0.100165
0.007602
0.29511
0.121145
-0.000355
0.121499
0.04425
0.302083
0.039691
-0.077358
0.117049
0.051553
0.25789
-0.034829
-0.135598
0.100769
0.011702
0.272094
0.063097
-0.101321
0.164418
0.02189
0.285214
0.01954
-0.036507
0.056047
0.002833
0.267114
-0.036386
-0.073648
0.037262
0.0041
0.294993
-0.038042
-0.125859
0.087817
-0.001025
0.283138
0.111147
0.010847
0.100299
0.026867
0.292171
-0.086284
-0.102297
0.016013
0.011121
0.279124
0.014409
-0.166483
0.180893
0.015169
0.28317
0.006762
-0.140287
0.147049
0.030823
0.26413
0.1013
0.003513
0.097787
0.000355
0.309124
0.07243
-0.094636
0.167067
0.061526
0.287932
0.09505
-0.002588
0.097638
0.023847
0.286041
0.010105
-0.057342
0.067447
0.02461
0.261452
-0.018578
-0.082371
0.063794
0.007665
0.290397
0.045843
-0.10257
0.148413
0.025208
0.270046
0.00648
-0.114867
0.121347
0.013631
0.254671
0.041567
-0.098096
0.139663
-0.005645
0.299834
0.041953
-0.082578
0.124532
0.05879
0.295811
0.087658
-0.017989
0.105647
0.017891
0.244706
-0.085962
-0.169217
0.083255
0.015531
0.269819
0.059562
-0.065034
0.124596
0.037898
0.271635
0.047766
-0.04608
0.093846
-0.018914
0.285726
0.087902
-0.050714
0.138616
0.041878
0.271755
0.044772
-0.142205
0.186977
0.030906
0.284482
-0.038672
-0.023281
-0.015391
0.014151
0.27638
0.077848
-0.029011
0.106859
0.027648
0.286744
-0.034279
-0.083536
0.049256
0.013497
0.277618
0.04124
-0.050756
0.091996
0.027039
0.25055
-0.049561
-0.164196
0.114635
0.007922
0.256297
0.017757
-0.064959
0.082717
0.008372
0.294936
0.004656
-0.050475
0.055131
0.025568
0.280572
0.023636
-0.094038
0.117674
0.020035
0.299465
0.099649
-0.070785
0.170433
0.049479
0.279977
0.027941
-0.077814
0.105755
0.032225
0.24306
-0.102213
-0.154583
0.052369
0.007304
0.295536
0.063143
0.008243
0.0549
0.035924
0.266465
-0.04724
-0.099233
0.051993
0.000988
0.32559
0.215056
-0.006914
0.22197
0.018658
0.26605
-0.108676
-0.173629
0.064953
0.016484
0.303849
0.118599
-0.047188
0.165788
0.056395
0.301021
0.129008
-0.034348
0.163356
0.028997
0.266682
-0.067472
-0.108197
0.040726
0.011866
0.271123
0.085508
-0.050375
0.135883
-0.001231
0.283099
0.046615
-0.141679
0.188295
0.00864
0.286398
0.039266
-0.090471
0.129737
0.036261
0.297417
0.122748
-0.044133
0.166881
0.023279
0.286138
0.070449
-0.032839
0.103289
0.026909
0.281171
0.045453
-0.037578
0.083031
0.003036
0.279173
0.027764
-0.063488
0.091252
0.018421
0.286166
0.02212
-0.070307
0.092426
0.011024
0.270141
-0.012678
-0.162893
0.150214
0.047715
0.293429
0.01417
-0.058136
0.072306
0.035666
0.276061
0.059329
-0.096532
0.155861
0.025405
0.298321
0.055108
-0.124631
0.179739
0.050552
0.273651
0.006465
-0.054807
0.061271
0.028059
0.264466
0.021718
-0.128754
0.150472
0.022583
0.278186
-0.026294
-0.064012
0.037718
0.022213
0.278542
0.069847
-0.008373
0.078219
0.022952
0.294075
0.061783
-0.01266
0.074443
0.028975
End of preview. Expand in Data Studio

DrivAerML Point Clouds

A preprocessed, point-cloud version of the DrivAerML high-fidelity CFD dataset, ready for training point-based deep learning surrogates (PointNet, PCT, DGCNN, Graph Neural Operators, etc.) for automotive external aerodynamics.

The original DrivAerML release contains 500 scale-resolving CFD simulations of parametrically morphed DrivAer notchback geometries and ships as 31 TB of raw STL / VTP / VTU / OpenFOAM data. This release distills the surface boundary of each run down to a single compressed .npz file (10 MB) containing the STL point coordinates and the CFD surface fields interpolated onto those points β€” so the full usable set fits in ~5 GB instead of multiple terabytes.

What's in here

For each run:

  • point_cloud_{i}.npz β€” STL surface points with interpolated CFD fields
  • force_mom_{i}.csv β€” time-averaged force and moment coefficients (Cd, Cl, Clf, Clr, Cs)

Plus at the dataset root:

  • splits.json β€” reproducible train/val/test assignment (seed 42, 80/10/10)
  • processing_log.json β€” per-run processing status and nearest-neighbor distance diagnostics

Directory layout

Drivaerml_point_clouds/
β”œβ”€β”€ train/          # 387 runs
β”‚   β”œβ”€β”€ run_1/
β”‚   β”‚   β”œβ”€β”€ point_cloud_1.npz
β”‚   β”‚   └── force_mom_1.csv
β”‚   └── ...
β”œβ”€β”€ val/            # 46 runs
β”‚   └── ...
β”œβ”€β”€ test/           # 52 runs
β”‚   └── ...
β”œβ”€β”€ splits.json
└── processing_log.json

Total: 485 runs (of the 500 original designs, 15 runs are missing from the source dataset: 167, 211, 218, 221, 248, 282, 291, 295, 316, 325, 329, 364, 370, 376, 403, 473).

Note on the Hugging Face dataset viewer: the viewer only previews the tabular force_mom_*.csv files (the Cd/Cl/Clf/Clr/Cs coefficients). The actual point-cloud payload lives in the .npz files, which the viewer does not render β€” clone the repo or stream it with huggingface_hub to access the geometry and fields.

.npz contents

Each point_cloud_{i}.npz has three arrays:

Key Shape Dtype Description
points (N, 3) float32 XYZ coordinates of STL surface points
fields (N, 5) float32 CFD surface fields interpolated to each STL point
field_names (5,) str Names of the columns in fields

N varies per run (roughly 300k points, matching the STL resolution).

The five field columns are:

  1. CpMeanTrim β€” time-averaged pressure coefficient
  2. wallShearStressMeanTrim_mag β€” magnitude of time-averaged wall shear stress
  3. wallShearStressMeanTrim_x β€” x-component
  4. wallShearStressMeanTrim_y β€” y-component
  5. wallShearStressMeanTrim_z β€” z-component

How the preprocessing works

The original DrivAerML surface data is stored in boundary_{i}.vtp: a high-resolution (8M point) mesh with CFD fields attached as cell data. The STL geometry drivaer_{i}.stl is a lower-resolution (300k point) representation of the same surface but without any fields.

To produce a clean point cloud where every geometry point carries its own CFD values, the pipeline runs the following per run:

  1. Load the .vtp boundary mesh and the .stl geometry.
  2. Convert the VTP's cell-centered fields to point-centered via PyVista's cell_data_to_point_data() (averaging adjacent cells to the shared vertex).
  3. Build a scipy.spatial.cKDTree over the VTP point coordinates.
  4. For each STL point, take the K=1 nearest neighbor in the VTP point cloud and copy its field values.
  5. Save points, fields, and field_names as a compressed .npz.
  6. Delete the raw VTP and STL to keep disk usage bounded while processing.

The K=1 nearest-neighbor mapping was chosen deliberately for benchmark comparability with existing DrivAerML / DrivAerNet++ leaderboard models (TripNet, FIGConvNet, RegDGCNN), all of which operate on the STL vertices directly. The nearest-neighbor distances are logged in processing_log.json for every run so this mapping can be audited.

The 80/10/10 train/val/test split is computed with numpy.random.default_rng(seed=42) over a sorted list of successfully-processed run IDs, making it fully reproducible from the splits.json manifest.

Loading

With datasets (CSV force/moment preview only)

The Hugging Face dataset loader will pick up the force_mom_*.csv files automatically:

from datasets import load_dataset
ds = load_dataset("Jrhoss/Drivaerml_point_clouds")
# ds["train"][0] -> {"Cd": 0.30, "Cl": 0.07, "Clf": -0.04, "Clr": 0.10, "Cs": 0.05}

Point clouds (recommended)

Clone the repo or snapshot-download it, then load .npz files directly:

from huggingface_hub import snapshot_download
import numpy as np
from pathlib import Path

root = Path(snapshot_download("Jrhoss/Drivaerml_point_clouds", repo_type="dataset"))

run = np.load(root / "train" / "run_1" / "point_cloud_1.npz", allow_pickle=True)
points = run["points"]            # (N, 3)
fields = run["fields"]            # (N, 5)
names  = run["field_names"]       # ['CpMeanTrim', 'wallShearStressMeanTrim_mag', ...]

Minimal PyTorch Dataset

import numpy as np
import pandas as pd
import torch
from pathlib import Path
from torch.utils.data import Dataset

class DrivAerMLPointClouds(Dataset):
    def __init__(self, root, split="train", n_points=16384):
        self.run_dirs = sorted((Path(root) / split).glob("run_*"),
                               key=lambda p: int(p.name.split("_")[1]))
        self.n_points = n_points

    def __len__(self):
        return len(self.run_dirs)

    def __getitem__(self, idx):
        run_dir = self.run_dirs[idx]
        run_id = int(run_dir.name.split("_")[1])

        npz = np.load(run_dir / f"point_cloud_{run_id}.npz", allow_pickle=True)
        points, fields = npz["points"], npz["fields"]

        # Random subsample for batching
        sel = np.random.choice(len(points), self.n_points, replace=False)
        points, fields = points[sel], fields[sel]

        # Integrated force/moment coefficients
        fm = pd.read_csv(run_dir / f"force_mom_{run_id}.csv").iloc[0]
        coeffs = torch.tensor([fm["Cd"], fm["Cl"], fm["Clf"], fm["Clr"], fm["Cs"]],
                              dtype=torch.float32)

        return {
            "points":  torch.from_numpy(points),   # (n_points, 3)
            "fields":  torch.from_numpy(fields),   # (n_points, 5)
            "coeffs":  coeffs,                     # (5,)
            "run_id":  run_id,
        }

Suggested tasks

  • Per-point surface field regression β€” predict CpMeanTrim and wall shear stress vectors from geometry alone. Comparable to the TripNet / FIGConvNet / RegDGCNN benchmarks (which report ~20% relative L2 error).
  • Integrated coefficient regression β€” predict Cd, Cl, etc. from the point cloud (global pooling over the surface).
  • Coupled prediction β€” joint learning of per-point fields and integrated coefficients, using the integrated values as a physics-informed auxiliary loss.

Known limitations

  • K=1 interpolation is not exact. It preserves the CFD field values faithfully where VTP and STL points are co-located, but introduces small errors at points where the STL has higher local resolution than the VTP. processing_log.json reports nn_dist_mean, nn_dist_max, and nn_dist_p99 per run so you can filter out any pathological cases.
  • Fields are time-averaged only. Transient information (e.g. unsteady vortex shedding in the wake) is not preserved; the source dataset contains scale-resolving data but only the trim-averaged fields are interpolated here.
  • Surface only. The volumetric flow field (the 50 GB-per-run volume_{i}.vtu) is not included β€” go to the source dataset for volumetric surrogate modeling.

Reproducing this dataset

The full preprocessing script is shipped alongside this repo (process_drivaerml.py). To regenerate from scratch:

pip install pyvista numpy scipy tqdm huggingface_hub
python process_drivaerml.py --output_dir ./drivaerml_processed --start 1 --end 500

The script streams one run at a time β€” downloading the VTP + STL + force/moment CSV from neashton/drivaerml, producing the .npz, then deleting the raw files β€” so it runs comfortably in a few tens of GB of free disk regardless of total dataset size.

License

CC-BY-SA 4.0, inherited from the source dataset. If you use this data you must give appropriate credit, indicate any changes, and distribute any derivative works under the same license.

Citation

Please cite the original DrivAerML paper:

@article{ashton2024drivaer,
  title   = {DrivAerML: High-Fidelity Computational Fluid Dynamics Dataset for Road-Car External Aerodynamics},
  author  = {Ashton, N. and Mockett, C. and Fuchs, M. and Fliessbach, L. and Hetmann, H.
             and Knacke, T. and Schonwald, N. and Skaperdas, V. and Fotiadis, G.
             and Walle, A. and Hupertz, B. and Maddix, D.},
  journal = {arXiv preprint arXiv:2408.11969},
  year    = {2024},
  url     = {https://arxiv.org/abs/2408.11969}
}

Acknowledgments

All credit for the underlying CFD data goes to the DrivAerML team (Neil Ashton et al., AWS / UpstreamCFD / BETA-CAE / Siemens Energy / Ford). This repository only redistributes a preprocessed surface-point-cloud view of that work. For the full multi-terabyte dataset including volumetric fields, OpenFOAM meshes, slice images, and residual plots, see neashton/drivaerml.

Downloads last month
468

Paper for Jrhoss/Drivaerml_point_clouds