RealPDEBench: A Benchmark for Complex Physical Systems with Real-World Data
Peiyan Hu, Haodong Feng, Hongyuan Liu, Tongtong Yan, Wenhao Deng, Tianrun Gao, Rong Zheng, Haoren Zheng, Chenglei Yu, Chuanrui Wang, Kaiwen Li, Zhi-Ming Ma, Dezhi Zhou, Xingcai Lu, Dixia Fan, Tailin Wu
We propose the first benchmark for complex physical systems with paired real-world data and simulated data, and explore how to bridge simulated and real-world data.
Abstract
Predicting the evolution of complex physical systems remains a central problem in science and engineering. Despite rapid progress in scientific Machine Learning (ML) models, a critical bottleneck is the lack of expensive real-world data, resulting in most current models being trained and validated on simulated data. Beyond limiting the development and evaluation of scientific ML, this gap also hinders research into essential tasks such as sim-to-real transfer. We introduce RealPDEBench, the first benchmark for scientific ML that integrates real-world measurements with paired numerical simulations. RealPDEBench consists of five datasets, three tasks, nine metrics, and ten baselines. We first present five real-world measured datasets with paired simulated datasets across different complex physical systems. We further define three tasks, which allow comparisons between real-world and simulated data, and facilitate the development of methods to bridge the two. Moreover, we design nine evaluation metrics, spanning data-oriented and physics-oriented metrics, and finally benchmark ten representative baselines, including state-of-the-art models, pretrained PDE foundation models, and a traditional method. Experiments reveal significant discrepancies between simulated and real-world data, while showing that pretraining with simulated data consistently improves both accuracy and convergence. In this work, we hope to provide insights from real-world data, advancing scientific ML toward bridging the sim-to-real gap and real-world deployment. Our benchmark, datasets, and instructions are available at https://realpdebench.github.io/.
RealPDEBench first benchmark integrating real-world measurements with paired simulations across five physical systems for scientific ML evaluation.
- First real-world scientific ML benchmark combining real-world and simulated data for complex physical systems
- Three categories of tasks, nine evaluation metrics, and ten baselines for comprehensive assessment
- Evidence that pretraining with simulated data improves both accuracy and convergence on real-world data
- physics-informed machine learning
- sim-to-real transfer
- five real-world measured datasets with paired simulations
Benchmark scope limited to specific domains without coverage of electromagnetics, structural mechanics, or aerodynamics
from the paperNo dedicated metrics for Combustion system
from the paperDoes not systematically explore strong out-of-distribution regimes
from the paper
Extend to additional physical domains and introduce domain-specific metrics
from the paperDesign out-of-distribution tasks and maintain long-term development toward more data and models
from the paper
Author keywords
- complex physical system
- PDE
- benchmark
- real-world data
- prediction
Related orals
On the Wasserstein Geodesic Principal Component Analysis of probability measures
Geodesic PCA for probability distributions using Wasserstein geometry with neural network parametrization for continuous distributions.
TabStruct: Measuring Structural Fidelity of Tabular Data
TabStruct benchmark evaluates tabular data generators on structural fidelity and conventional dimensions using global utility metric without ground-truth causal structures.
Monocular Normal Estimation via Shading Sequence Estimation
RoSE estimates surface normals via shading sequence prediction, addressing 3D misalignment in monocular normal estimation.
TTSDS2: Resources and Benchmark for Evaluating Human-Quality Text to Speech Systems
TTSDS2 metric robustly correlates with human judgments for TTS evaluation across diverse speech domains maintaining >0.5 Spearman correlation.
World-In-World: World Models in a Closed-Loop World
Introduces closed-loop benchmark evaluating generative world models on embodied task performance rather than visual quality.