Nvidia Chips Dominate AI Training Benchmarks

Nvidia Chips Dominate AI Training Benchmarks

New data released on Wednesday, June 4, 2025, confirms that Nvidia’s (NVDA.O) newest chips are making significant advancements in the crucial area of training large artificial intelligence (AI) systems.

 

The fresh benchmarks indicate a dramatic reduction in the number of chips required to train complex large language models (LLMs). This efficiency gain is vital for the AI industry, which constantly seeks to optimize the intensive computational processes behind cutting-edge AI.

 

MLCommons Benchmarks Highlight Performance Gains

MLCommons, a non-profit organization dedicated to publishing standardized benchmark performance results for AI systems, provided the new data.

 

Comprehensive AI Training Data

The results detail how chips from various manufacturers, including Nvidia and Advanced Micro Devices (AMD.O), perform during the AI training phase. During training, AI systems are fed vast quantities of data to learn patterns and make predictions. While much of the stock market’s recent attention has shifted towards the larger market for AI inference (where AI systems handle user queries), the number of chips needed for training remains a critical competitive factor.

 

This is particularly true as companies like China’s DeepSeek claim to develop competitive chatbots using significantly fewer chips than their U.S. counterparts.

 

First Benchmarks for Large Models

Crucially, these results represent the first time MLCommons has released data specifically on how chips perform when training massive AI systems. An example provided is Meta Platforms’ (META.O) open-source AI model, Llama 3.1 405B. This model possesses a sufficiently large number of “parameters” to serve as an indicator of how chips would perform on the most complex training tasks globally, which can involve trillions of parameters. This benchmark provides a real-world proxy for the capabilities of these advanced chips.

 

Nvidia Blackwell’s Dominance in Training Speed

Nvidia’s latest generation of chips, Blackwell, demonstrated remarkable performance in these benchmarks.

 

Double the Speed of Previous Generation

Nvidia and its partners were the sole entrants to submit data for training such a large model. The results unequivocally showed that Nvidia’s new Blackwell chips are more than twice as fast as their previous generation Hopper chips on a per-chip basis. This significant leap in performance translates directly to faster AI development and reduced computational costs.

 

Dramatic Reduction in Training Time

In the fastest recorded results for Nvidia’s new chips, a cluster of 2,496 Blackwell chips completed the demanding training test in an astonishing 27 minutes. To achieve a faster time, it required more than three times that number of Nvidia’s prior generation chips. This illustrates the profound efficiency gains and computational power packed into the Blackwell architecture.

See also  SpaceX Launches Starlink Mission from Florida

 

Industry Trends: Smaller Clusters for Greater Efficiency

The AI industry is also seeing an evolution in how large-scale training tasks are managed.

 

Subsystems Over Homogeneous Clusters

During a press conference, Chetan Kapoor, Chief Product Officer for CoreWeave, a company that collaborated with Nvidia on some of the benchmark results, discussed an emerging industry trend. He noted a shift towards “stringing together smaller groups of chips into subsystems for separate AI training tasks.”

 

This approach contrasts with the traditional method of creating massive, homogeneous groups of 100,000 chips or more. Kapoor emphasized that this methodology enables ongoing acceleration and reduction in the time required to train “crazy, multi-trillion parameter model sizes,” highlighting a strategic optimization in AI infrastructure deployment. This focus on modular, efficient clusters helps to tackle the ever-growing computational demands of cutting-edge AI models.

Travel the World as a Digital Nomad
travel the world as a digital nomad

Imagine your office view changing from a cubicle wall to a sandy beach in Thailand, a bustling cafe in Lisbon, or a mountain retreat in Colombia. This dream is the Read more

Huawei Pura 80 Ultra: Dual Telephoto, Single Sensor Camera
Huawei Pura 80 Ultra: Dual Telephoto, Single Sensor Camera

Huawei has officially unveiled its flagship Pura 80 series phones in China, with the top-tier Pura 80 Ultra leading the charge with a groundbreaking camera innovation. This new device introduces Read more

How to See June’s Strawberry Micromoon
How to See June's Strawberry Micromoon

Skywatchers are in for a treat as June's Strawberry Moon prepares to dazzle the night sky. This year, the full moon will be a micromoon, meaning it will appear slightly Read more

Axial Seamount: Oregon Coast Underwater Volcano Poised to Erupt
axial seamount oregon coast underwater volcano poised to erupt

Approximately 300 miles off the picturesque coast of Oregon, a massive submarine volcano known as Axial Seamount is exhibiting compelling signs that it is poised for an eruption—its first since Read more

Powerful X2.7 Solar Flare Erupts from Sun
powerful x2.7 solar flare erupts from sun

The Sun recently displayed its immense power with the emission of a significant solar flare. Powerful X2.7 Solar Flare Erupts from Sun This strong burst of Read more

Android Auto 14.7 Beta: New Bright Theme & Gemini AI
Android Auto 14.7 Beta: New Bright Theme & Gemini AI

Google continues to refine its in-car infotainment and driving assistance system, Android Auto, with a significant new update.   The tech giant has just released Version 14.7 Beta, which promises Read more