Skip to main content

Aurora (Intel-HPE): Website

Aurora, Argonne’s first exascale computer, features Intel’s Data Center GPU Max Series GPUs and Intel XEON Max Series CPUs with HBM, a CPU-GPU interconnect (PCI-e), the HPE Slingshot fabric system interconnect, and a Cray EX platform.  Aurora will featur several technological innovations and will also include a revolutionary I/O system – the Distributed Asynchronous Object Store (DAOS) – to support new types of machine learning and data science workloads alongside traditional modeling and simulation workloads.

The public information for Aurora is listed below.

System Spec Aurora System Spec
System Performance ≥2EF DP peak Number of Nodes
Compute Node 2 x 4th gen Intel XEON Max Series CPUs with HBM; 6 x Intel Data Center GPU Max Series GPU Architecture
Intel Data Center GPU Max Series; Tile based, chiplets, HBM stack, Foveros 3d integration
CPU-GPU Interconnect PCIe Aggregate System Memory
Aggregate DDRS Memory
10.6PB Aggregate CPU HBM
Aggregate GPU HBM
8.1PB Node Memory Architecture
Unified memory architecture, RAMBO
System Interconnect HPE Slingshot 11; Dragonfly topology with adaptive routing Network Switch
25.6 TB/s per switch, from 64-200 GB ports (25GB/s per direction)
High-Performance Storage
230 PB, 31 TB/s (DAOS) Platform HPE Cray XE with the HPE HPCM software stack + Intel enhancements
Compilers Intel, LLVM, GCC Programming Models
Intel oneAPI, OpenMP, DPC++/SYCL
Programming languages and models
Fortran, C, C++, OpenMP 5.x (Intel, Cray, and possibly LLVM compilers), UPC (Cray), Coarray Fortran (Intel), Data Parallel C++ (Intel and LLVM compilers), Open SHMEM, Python, Numba, MPI, OpenCL Programming Tools
Open|Speedshop, TAU, HPCToolkit, Score-P, Darshan, Intel Trace Analyser and Collector, Intel Vtune, Advisor, and Inspector, PAPI, GNU gprof
Debugging and Correctness Tools Stack Trace Analysis Tool, gdb, Cray Abnormal Termination Processing Math Libraries Intel MKL, Intel MKL-DNN, ScaLAPACK
GUI and Viz APIs, I/O Libraries X11, Motif, QT, NetCDF, Parallel, NetCDF, HDF5 Frameworks TensorFlow, PyTorch, Scikit-learn, Spark Mllib, GraphX, Intel DAAL, Intel MKL-DNN

The Aurora software stack includes the HPE HPCM software stack, the Intel oneAPI software development kit, and data and learning frameworks. Supported programming models include MPI, Intel OneAPI, OpenMP, SYCL/DPC++., Kokkos, RAJA, and others.

For INCITE 2025, Aurora allocations will be requested in native Aurora node-hours.

The most recent public information on Aurora can be found at

Aurora Website

For more information about any of the ALCF resources, please visit

Frontier (Cray-HPE): Website

Frontier is a HPE Cray EX supercomputer located at the Oak Ridge Leadership Computing Facility. With a theoretical peak double-precision performance of approximately 2 exaflops (2 quintillion calculations per second), it is the fastest system in the world for a wide range of traditional computational science applications. The system has 74 Olympus rack HPE cabinets, each with 128 AMD compute nodes, and a total of 9,408 AMD compute nodes.

Each Frontier compute node consists of a single 64-core AMD 3rd Gen EPYC CPU and 512 GB of DDR4 memory. Each node also contains [4x] AMD MI250X, each with 2 Graphics Compute Dies (GCDs) for a total of 8 GCDs per node. Each GCD has 64 GB of high-bandwidth memory (HBM2E) for a total of 512 GB of HBM memory per node.

Frontier System Configuration
Architecture: HPE/AMD
Node: 1 3rd Gen AMD EPYC CPU
4 AMD Instinct 250X GPUs
Number of Nodes 9,408
GPU link
AMD Infinity Fabric
4x HPE Slingshot NICs providing 100 GB/s network bandwidth
File system 679 PB, GPFS center-wide file system
Peak Performance >1.5 EF

Further details may be found

Polaris (HPE): Website

Developed in collaboration with Hewlett Packard Enterprise (HPE), the 44-petaflop system called Polaris, is available to users through INCITE 2025 program. Polaris is a hybrid CPU/GPU leading-edge testbed system that will give scientists and application developers a platform to test and optimize codes for Aurora, Argonne’s new Intel-HPE exascale supercomputer.

The Polaris software environment is equipped with the HPE Cray programming environment, PE Performance Cluster Manager (HPCM) system software, and the ability to test programming models, such as OpenMP and SYCL, that will be available on Aurora and the next generation of DOE’s high performance computing systems. Polaris users will also benefit from NVIDIA’s HPC software development kit, a suite of compilers, libraries, and tools for GPU code development.

Polaris System Configuration
HPE Apollo 6500 Gen10+
Compute Node
1 AMD EPYC “Milan” processor + 4 NVIDIA A100 GPUs
System Size
560 nodes
System Peak 44 PF DP
System Memory
280 TB CPU DDR4 + 87.5 TB GPU HBM2
Peak Power 1.8 MW
Node Performance 78 TF DP
Memory/node 512 GB CPU DDR4 + 160 GB GPU HBM2
Interconnect HPE Slingshot
Node-to-Node Interconnect 200 Gb
Programming Models OpenMP 4.5/5, SYCL, Kokkos, RAJA, HIP
Performance/Debugging GPU tools, PAPI, TAU, HPCToolkit, DDT
Frameworks Python/Numba, TensorFlow, PyTorch