Code Scaling

INCITE requires applicants to include code performance and scaling results for allocations requests. This document is now mandatory. Below are resources for select CIG codes that can be used in your proposals.


The scaling capabilities of ASPECT for large-scale 3D mantle convection simulations on Stampede are shown for a 3D box model (left) and for a spherical shell model (right). These demonstrate that this next-generation code scales well on problems up to tens of millions of elements and up to over 1000 processors.


Strong scalability of 3D box model (1283 or 2563 elements) on Stampede. [pdf] [tiff]


ASPECT sphere.png

Strong scaling of ASPECT's spherical shell model. The wall clock time for one time step is measured by a mantle convection model including temperature dependent viscosity and composition perturbation in a spherical shell. [pdf]


Same as above for SDSC Comet [pdf]

The weak and strong scalability of ASPECT was also examined on a local cluster in detail demonstrating good scalability up to several thousand processors for large problems.

ASPECT weak.png



ASPECT Strong.png



The largest scaling test above corresponds to roughly a 17 million elements and scales well up to 16384 cores.


Calypso strong.png

Strong scaling on TACC Stampede for different sized simulations.  Ideal scaling is plotted by a dotted line. [pdf]



Calypso MIC.png 

Strong scaling using the MIC processor on TACC Stampede.  Ideal scaling is plotted by a dotted line. [pdf]


Tests were perfromed on TACC Lonestar for a regional model with 6.4 million unknowns using up to 256 cores.

CitcomS scaling.png

Speedup (left) [pdf] and scalability (right) [pdf] of CitcomS on Lonestar.


Rayleigh is based on the solar dynamo code ASH which scales efficiently to 10,000 cores. We expect similar performance for Rayleigh.

ASH Strong.png

Rayleigh's strong scaling on the TACC Stampede system for different sized simulations. Ideal scaling is plotted by a dotted line.  [pdf]

Rayleigh has been benchmarked on ALCF MIRA using up to 131,072 single-threaded cores for the Geodynamo Working Group benchmarking exercise. 


Left: The parallel efficiency for the non-magnetic and magnetic runs at the 20483 problem size. Efficiency is defined as the ratio of the realized speedup to the ideal speed-up. deal performance is taken relative to 16,384 cores, the minimum core count or this problem size. Right: Strong scaling results for the ame set of runs. Ideal scaling for each series is indicated by the dashed line. [pdf] [png]

Sign In