Frontier Supercomputer Achieves New Exascale Milestones

Two-and-a-half years after breaking the exascale barrier, the Frontier supercomputer at the Department of Energy's Oak Ridge National Laboratory continues to set new standards for its computing speed and performance.

The HPE Cray EX supercomputing system reported new highs for problem-solving speeds this week, updated for the TOP500 announcement at the International Conference for High Performance Computing, Networking, Storage and Analysis , or SC24, in Atlanta. The Frontier team achieved a High-Performance Linpack, or HPL, score of 1.35 exaflops, or 1.35 quintillion calculations per second using double-precision arithmetic, the 64-bit international scientific standard for computational accuracy.

"Frontier's new numbers reflect the tremendous contributions of OLCF's computing and computational experts who know how to best optimize bleeding-edge high-performance computing systems to serve the evolving needs of our scientific user community," said Gina Tourassi, associate laboratory director for computing and computational sciences at ORNL. "They never stop pushing boundaries, throughout the life of the system."

The score earned Frontier the No. 2 spot on the November 2024 TOP500 list , which ranks the fastest supercomputers in the world. Frontier debuted atop the list in May 2022 at 1.1. exaflops as the first machine to achieve exascale performance at more than a quintillion calculations per second.

"Our in-house team of experts understands just how to get the most from this system in terms of performance," said Ashley Barker, director of the Oak Ridge Leadership Computing Facility, which houses Frontier. "They're the world's most experienced exascale computing team, and they ran this test on Frontier while the machine ran other science problems for users."

Large group of staff standing in front of computing cabinets in data center that houses the Frontier supercomputer at Oak Ridge National Laboratory
The Frontier supercomputing team at ORNL broke the exascale barrier nearly three years ago and remains the most experienced exascale computing team. Credit: Carlos Jones/ORNL, U.S. Dept. of Energy

Frontier's improved speed marks a jump of roughly 150 petaflops, or 150 quadrillion calculations per second - roughly equivalent to the performance of its predecessor supercomputer Summit , which was decommissioned last week.

"We could already solve amazingly large problems on Frontier - the biggest science problems on the planet," said Al Geist, an ORNL Corporate Fellow and Frontier project director. "This year, we've basically picked up the power of another supercomputer (like Summit)."

Frontier relies on a constellation of thousands of nodes, each a self-contained supercomputer of one CPU and four GPUs, connected by more than 90 miles of cable to enable them to communicate and work together on large problems. Frontier debuted in 2022 with more than 9,400 nodes, and crews have since added another 400 nodes for a current total of more than 9,800. The latest HPL benchmark ran across 9,500 nodes.

Geist estimates about half of Frontier's improved score can be credited to the new nodes, which had been used for application development and testing by the Exascale Computing Project . The ECP oversaw development of software applications for Frontier and other exascale machines, such as the Aurora supercomputer at Argonne National Laboratory and the El Capitan supercomputer at Lawrence Livermore National Laboratory, and completed its work earlier this year.

On top of the new nodes and exascale experience gained by OLCF computing experts, Frontier boasts improvements to its math libraries developed by HPE, which built Frontier, and AMD, maker of the CPUs and GPUs that power Frontier.

Besides the updated HPL number, the Frontier team logged a new High-Performance Linpack-Mixed Precision, or HPL-MxP, score of 11.4 exaflops, or 11.4 quintillion calculations per second, running across all 9,800 nodes. That's a jump of more than an exaflop from the previous HPL-MxP score of 10.2 exaflops.

The detailed simulations used to model such phenomena as cancer cells, supernovas, the coronavirus or the atomic structure of elements require 64-bit precision, a computationally demanding standard of accuracy. Machine-learning algorithms used for artificial intelligence typically require less precision - sometimes as little as 32-, 24- or 16-bit accuracy.

"This shows how capable Frontier is for AI problems," Barker said. "The results we get from mixed-precision runs are equally accurate, but they're reached in a different way, mathematically. For some questions, double precision will remain the standard, but at these potential speeds, we expect more researchers will start to explore what kind of results they can achieve via mixed precision. They'd like to take advantage of this ability to solve problems faster by a factor of 10, and our team knows how to make that happen."

Frontier is an HPE Cray EX system with more than 9,800 nodes, each equipped with a third-generation AMD EPYC™ CPU and four AMD Instinct™ MI250X GPUs. The OLCF is a DOE Office of Science user facility.

UT-Battelle manages ORNL for DOE's Office of Science, the single largest supporter of basic research in the physical sciences in the United States. DOE's Office of Science is working to address some of the most pressing challenges of our time. For more information, visit https://energy.gov/science . - Matt Lakin

/Public Release. This material from the originating organization/author(s) might be of the point-in-time nature, and edited for clarity, style and length. Mirage.News does not take institutional positions or sides, and all views, positions, and conclusions expressed herein are solely those of the author(s).View in full here.