Extra-P and Score-P: Automated Performance Modeling for HPC

Extra-P's empirical performance modeling and Score-P's measurement infrastructure: from automated scalability bug detection to noise-resilient modeling for exascale systems.

Performance modeling in high-performance computing has traditionally been a manual, expert-driven process. Developers would run small-scale experiments, manually derive scaling formulas, and hope these predictions held at production scale. This approach is time-consuming, error-prone, and doesn’t scale to the complexity of modern HPC applications with thousands of parameters and code paths.

The Extra-P (Exascale Performance Prediction) framework, developed collaboratively by TU Darmstadt and ETH Zürich’s SPCL group, automates this entire process. Combined with Score-P, the unified measurement infrastructure from the VI-HPS community, these tools enable developers to automatically discover performance models, detect scalability bugs before they manifest at scale, and make data-driven optimization decisions.

This post traces the evolution of Extra-P from its foundational work on automated scalability bug detection to recent advances in noise-resilient modeling using deep neural networks, all built on Score-P’s robust measurement infrastructure.

The Extra-P Ecosystem

Extra-P and Score-P form a complete performance analysis pipeline:

The partnership between TU Darmstadt’s Laboratory for Parallel Programming and ETH Zürich’s Scalable Parallel Computing Lab (SPCL) has driven continuous innovation in both tools since 2013. Score-P serves as the measurement backend not just for Extra-P but for the entire VI-HPS tool ecosystem including Vampir, Scalasca, and TAU.

Core Methodology

Automated Scalability Analysis

The foundational insight behind Extra-P is that performance bottlenecks often follow predictable mathematical patterns—but discovering these patterns manually is impractical for complex codes.

Multi-Parameter Modeling

Real applications don’t scale along a single dimension—they have multiple problem sizes, decompositions, and algorithmic parameters.

Noise-Resilient Modeling

Performance measurements in real systems are noisy—network jitter, OS interference, and hardware variability create measurement uncertainty that can corrupt performance models.

Score-P: Measurement Infrastructure

System Architecture

Score-P serves as the unified measurement infrastructure for the entire VI-HPS tool ecosystem, providing a common foundation for diverse performance tools.

Language and Platform Support

Score-P has evolved to support diverse programming models and execution platforms as HPC software has diversified.

Advanced Applications

Isoefficiency and Configuration

Understanding how applications scale requires more than just runtime measurements—it requires principled analysis of efficiency and configuration choices.

Case Studies and Real-World Applications

The Path Forward

The Extra-P and Score-P ecosystem has fundamentally changed how HPC developers approach performance analysis. What once required manual derivation of scaling formulas and expert intuition is now automated, systematic, and data-driven.

Key Achievements:

Current Frontiers:

Open Challenges:

The partnership between measurement infrastructure (Score-P) and automated analysis (Extra-P) has created a powerful platform for performance understanding. As HPC systems continue to grow in complexity, automated empirical modeling becomes not just useful but essential—manual approaches simply cannot scale to the complexity of exascale applications with millions of lines of code and thousands of configuration parameters.

For researchers and practitioners, the Extra-P/Score-P ecosystem provides both a mature toolchain for immediate use and an active research platform for advancing the state of performance analysis. The code is open source, the methodology is well-documented, and the community continues to push the boundaries of what automated performance modeling can achieve.