Your new experience awaits. Try the new design now and help us make it even better

TECHNOLOGY AND CODE article

Front. High Perform. Comput.

Sec. Benchmarking

Extra-P – Empirical Performance Modeling Made Easy

Provisionally accepted
Alexandru  CalotoiuAlexandru Calotoiu1Marcin  CopikMarcin Copik1Fabian  CzappaFabian Czappa2Alexander  GeissAlexander Geiss2Gustavo  MoraisGustavo Morais2Marcus  RitterMarcus Ritter3Sergei  ShudlerSergei Shudler4Torsten  HoeflerTorsten Hoefler1Felix  WolfFelix Wolf2*
  • 1Eidgenossische Technische Hochschule Zurich Departement Informatik, Zürich, Switzerland
  • 2Technische Universitat Darmstadt, Darmstadt, Germany
  • 3ABB AG, Mannheim, Germany
  • 4e.solutions GmbH, Erlangen, Germany

The final, formatted version of the article will be published soon.

High-performance computing (HPC) applications face challenges in achieving scalability, with bottlenecks often discovered only late in the development cycle. Performance modeling offers a means to predict and understand scalability, but analytical approaches require deep expertise and are often impractical for large, complex codes. To address this, the Extra-P project provides a user-friendly tool for empirical performance modeling, enabling automated model generation from a small number of carefully selected experiments. This paper presents an overview of Extra-P, its underlying methodology—the Performance Model Normal Form (PMNF)—and its evolution into a mature tool for detecting and analyzing scalability issues. We discuss strategies to reduce experiment costs through parameter selection, sparse modeling, and Gaussian process regression, as well as techniques for mitigating the impact of noise using iterative refinement and deep learning. Furthermore, we highlight novel use cases, including segmented modeling and validation of user expectations, and of course demonstrate how Extra-P can uncover hidden bottlenecks in real-world applications such as HOMME or MPI libraries. Finally, we outline the software's architecture and future directions, emphasizing the potential for integration with AI-driven methods and adaptation to increasingly heterogeneous hardware.

Keywords: Empirical Performance Modeling, machine learning, parallelism, performance, Scalability

Received: 26 Sep 2025; Accepted: 24 Dec 2025.

Copyright: © 2025 Calotoiu, Copik, Czappa, Geiss, Morais, Ritter, Shudler, Hoefler and Wolf. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Felix Wolf

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.