Towards Experiment Execution in Support of Community Benchmark Workflows for HPC

📅 2025-07-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address low reusability of HPC benchmarks, poor cross-platform portability, and inefficient resource validation, this paper proposes the “benchmark carpentry” paradigm—a lightweight, reusable experimental execution framework. Methodologically, it integrates Cloudmesh’s experiment executor with HPE SmartSim, incorporating standardized workflow templates, AI/ML–simulation coupling mechanisms, and a unified experimental management interface. Its key contribution is the first application of craftsmanship principles to benchmarking process design, enabling automated, cross-domain and cross-architecture benchmark deployment and capability assessment. Evaluated on representative scientific computing workloads—including cloud masking analysis, seismic forecasting, and CFD surrogate modeling—the framework achieves ≥92% workflow reproducibility and reduces average deployment time by 68%, significantly improving resource configuration efficiency. It establishes a scalable, community-driven paradigm for HPC capability validation.

Technology Category

Application Category

📝 Abstract
A key hurdle is demonstrating compute resource capability with limited benchmarks. We propose workflow templates as a solution, offering adaptable designs for specific scientific applications. Our paper identifies common usage patterns for these templates, drawn from decades of HPC experience, including recent work with the MLCommons Science working group. We found that focusing on simple experiment management tools within the broader computational workflow improves adaptability, especially in education. This concept, which we term benchmark carpentry, is validated by two independent tools: Cloudmesh's Experiment Executor and Hewlett Packard Enterprise's SmartSim. Both frameworks, with significant functional overlap, have been tested across various scientific applications, including conduction cloudmask, earthquake prediction, simulation-AI/ML interactions, and the development of computational fluid dynamics surrogates.
Problem

Research questions and friction points this paper is trying to address.

Demonstrating HPC compute capability with limited benchmarks
Creating adaptable workflow templates for scientific applications
Improving experiment management tools for broader workflow adaptability
Innovation

Methods, ideas, or system contributions that make the work stand out.

Workflow templates for adaptable scientific applications
Benchmark carpentry improves experiment management adaptability
Validated by Cloudmesh and HPE SmartSim frameworks
🔎 Similar Papers
No similar papers found.
Gregor von Laszewski
Gregor von Laszewski
University of Virginia
W
Wesley Brewer
Oak Ridge National Laboratory, Oak Ridge, TN 37831, USA
Sean R. Wilkinson
Sean R. Wilkinson
Research Scientist, Oak Ridge National Laboratory
BioinformaticsData ScienceHigh Performance ComputingFAIRWorkflows
A
Andrew Shao
Hewlett Packard Enterprise Canada, Victoria, British Columbia, Canada
J
J. P. Fleischer
University of Florida, Gainesville, FL 32611, USA
H
Harshad Pitkar
Cummins, Columbus, IN 47201, USA
C
Christine R. Kirkpatrick
San Diego Supercomputer Center, University of California, San Diego, La Jolla, CA 92093, USA
G
Geoffrey C. Fox
Biocomplexity Institute, University of Virginia, Charlottesville, VA 22911, USA