World’s Fastest Supercomputers Changing Fast

world’s-fastest-supercomputers-changing-fast

Modern day computing workloads — like scientific simulations, visualization, knowledge analytics, and device mastering — are pushing supercomputing facilities, cloud providers and enterprises to rethink their computing architecture.

The processor or the community or the program optimizations on your own can’t tackle the latest requirements of scientists, engineers and data researchers. Alternatively, the info heart is the new unit of computing, and companies have to look at the full engineering stack.

The newest rankings of the world’s most strong methods present continued momentum for this total-stack method in the latest generation of supercomputers.

NVIDIA systems accelerate around 70 percent, or 355, of the systems on the Major500 record produced at the SC21 significant efficiency computing meeting this week, such as in excess of 90 percent of all new techniques. Which is up from 342 techniques, or 68 p.c, of the machines on the Top500 record produced in June.

NVIDIA also carries on to have a solid presence on the Eco-friendly500 listing of the most energy-productive devices, powering 23 of the leading 25 techniques on the record, unchanged from June. On regular, NVIDIA GPU-run methods provide three.5x larger electrical power efficiency than non-GPU methods on the listing.

Highlighting the emergence of a new era of cloud-native units, Microsoft’s GPU-accelerated Azure supercomputer ranked 10th on the listing, the first leading 10 demonstrating for a cloud-based program.

AI is revolutionizing scientific computing.  The selection of research papers leveraging HPC and device discovering has skyrocketed in recent yrs developing from about 600 ML HPC papers submitted in 2018 to practically five,000 in 2020.


The ongoing convergence of HPC and AI workloads is also underscored by new benchmarks such as HPL-AI and MLPerf HPC.

HPL-AI is an rising benchmark of converged HPC and AI workloads that works by using combined-precision math — the basis of deep mastering and lots of scientific and professional careers — whilst however delivering the full accuracy of double-precision math, which is the standard  measuring adhere for common HPC benchmarks.

And MLPerf HPC addresses a fashion of computing that speeds and augments simulations on supercomputers with AI, with the benchmark measuring functionality on a few critical workloads for HPC centers: astrophysics (Cosmoflow), weather (Deepcam) and molecular dynamics (Opencatalyst).

NVIDIA addresses the whole stack with GPU-accelerated processing, intelligent networking, GPU-optimized programs, and libraries that support the convergence of AI and HPC. This tactic has supercharged workloads and enabled scientific breakthroughs.

Let’s seem extra closely at how NVIDIA is supercharging supercomputers.

Accelerated Computing

The blended electricity of the GPU’s parallel processing abilities and about 2,500 GPU-optimized apps allows end users to velocity up their HPC positions, in quite a few conditions from weeks to hrs.

We’re frequently optimizing the CUDA-X libraries and the GPU-accelerated applications, so it’s not unconventional for people to see an x-component overall performance obtain on the similar GPU architecture.

As a result, the general performance of the most commonly utilised scientific applications — which we contact the “golden suite” — has improved 16x in excess of the past six several years, with additional developments on the way.


16x general performance on prime HPC, AI and ML apps from full-stack innovation.

And to help users speedily acquire gain of greater efficiency, we offer the most current variations of the AI and HPC computer software as a result of containers from the NGC catalog. End users just pull and operate the application on their supercomputer, in the information centre or the cloud.

Convergence of HPC and AI 

The infusion of AI in HPC helps researchers velocity up their simulations while reaching the precision they’d get with the traditional simulation technique.

That’s why an expanding variety of researchers are taking gain of AI to velocity up their discoveries.

That consists of 4 of the finalists for this year’s Gordon Bell prize, the most prestigious award in supercomputing. Companies are racing to establish exascale AI pcs to assistance this new product, which combines HPC and AI.

That power is underscored by somewhat new benchmarks, this kind of as HPL-AI and MLPerf HPC, highlighting the ongoing convergence of HPC and AI workloads.

To fuel this trend, previous week NVIDIA introduced a wide vary of state-of-the-art new libraries and computer software progress kits for HPC.

Graphs — a crucial knowledge structure in present day info science — can now be projected into deep-neural network frameworks with Deep Graph Library, or DGL, a new Python package.

NVIDIA Modulus builds and trains physics-educated machine finding out versions that can study and obey the rules of physics.

And NVIDIA launched a few new libraries:

  • ReOpt – to boost operational performance for the $10 trillion logistics sector.
  • cuQuantum – to speed up quantum computing exploration.
  • cuNumeric – to accelerate NumPy for researchers, info researchers, and machine discovering and AI scientists in the Python local community.

Weaving it all with each other is NVIDIA Omniverse — the company’s digital environment simulation and collaboration platform for 3D workflows.

Omniverse is used to simulate digital twins of warehouses, crops and factories, of physical and biological systems, of the 5G edge, robots, self-driving automobiles and even avatars.

Utilizing Omniverse, NVIDIA declared final 7 days that it will build a supercomputer, identified as Earth-two, devoted to predicting local weather adjust by making a electronic twin of the earth.

Cloud-Indigenous Supercomputing

As supercomputers choose on additional workloads across details analytics, AI, simulation and visualization, CPUs are stretched to aid a developing number of conversation duties needed to run huge and complex units.

Information processing models relieve this stress by offloading some of these procedures.

As a entirely integrated knowledge-centre-on-a-chip system, NVIDIA BlueField DPUs can offload and control details middle infrastructure jobs as an alternative of earning the host processor do the perform, enabling more powerful protection and more effective orchestration of the supercomputer.

Put together with NVIDIA Quantum InfiniBand system, this architecture provides ideal bare-metallic overall performance although natively supporting multinode tenant isolation.


NVIDIA’s Quantum InfiniBand system offers predictive, bare-metallic functionality isolation.

Many thanks to a zero-believe in tactic, these new programs are also much more protected.

BlueField DPUs isolate programs from infrastructure. NVIDIA DOCA 1.two — the latest BlueField software package system — enables following-generation dispersed firewalls and broader use of line-price data encryption. And NVIDIA Morpheus, assuming an interloper is previously within the facts heart, makes use of deep understanding-driven data science to detect intruder functions in actual time.

And all of the trends outlined over will be accelerated by new networking technology.

NVIDIA Quantum-two, also announced previous week, is a 400Gbps InfiniBand platform and consists of the Quantum-two change, the ConnectX-7 NIC, the BlueField-three DPU, as effectively as new software program for the new networking architecture.

NVIDIA Quantum-two presents the added benefits of bare-steel high general performance and protected multi-tenancy, permitting the next generation of supercomputers to be secure, cloud-indigenous and superior utilized.

Benchmark purposes: Amber, Chroma, GROMACS, MILC, NAMD, PyTorch, Quantum Espresso Random Forest FP32 , TensorFlow, VASP | GPU node: dual-socket CPUs with 4x P100, V100, or A100 GPUs.

Leave a comment

Your email address will not be published.


*