Blackwell is 2 times more efficient than its predecessors

Nvidia has unveiled new Blackwell chips that significantly speed up the process of training large artificial intelligence (AI) models. According to new data published by the non-profit organization MLCommons, the number of chips required to train large language models has significantly decreased.
New Nvidia chips double the speed of training large-scale AI
© BoliviaInteligente

 

MLCommons specializes in comparing the performance of AI hardware. The latest report included testing results from Nvidia and AMD chips, as well as other manufacturers. At the same time, an important indicator remains the efficiency of training AI models – the stage when systems receive huge amounts of data for training. Although many now pay attention to the speed of AI when responding to users, it is the number and power of chips involved in training that remains a critical factor.

The tests were conducted using the Llama 3.1 405B* model, a complex open AI module with a huge number of parameters developed by Meta*. Training data for this model was provided only by Nvidia and its partners. The results showed that the new Blackwell chips are twice as fast as the previous Hopper series per chip.

It took 2,496 Blackwell chips to complete the test, which took 27 minutes. That’s better than the previous chips, which were more than three times as many.

Experts note a change in approaches in the industry: instead of huge homogeneous clusters of tens of thousands of chips, more compact subsystems are increasingly used for individual AI training tasks. This method allows for speeding up the process and reducing the time for training models with trillions of parameters.

* belongs to the Meta company, the organization is recognized as extremist, its activities are prohibited in Russia

Similar Posts