option
Home
News
Nvidia Dominates Gen AI Benchmarks, Outperforming Two Rival AI Chips

Nvidia Dominates Gen AI Benchmarks, Outperforming Two Rival AI Chips

April 16, 2025
161

Nvidia's general-purpose GPU chips have once again dominated one of the most widely recognized benchmarks for assessing chip performance in artificial intelligence, this time focusing on generative AI applications such as large language models (LLMs). The competition was relatively one-sided.

Systems from SuperMicro, Hewlett Packard Enterprise, Lenovo, and other companies, each equipped with up to eight Nvidia chips, secured the majority of the top spots in the MLPerf benchmark test organized by the MLCommons, an industry consortium. This test, which measures the speed at which machines can produce tokens, process queries, or output data samples—known as AI inference—was the fifth in a series of prediction-making benchmarks that have been conducted over the years.

This latest iteration of the MLPerf benchmark included new tests tailored to common generative AI tasks. One test evaluated the performance of chips on Meta's open-source LLM, Llama 3.1 405b, a substantial model widely used in the field. Another test introduced an interactive version of Meta's smaller Llama 2 70b, designed to simulate chatbot interactions where response time is crucial. This test specifically measures how quickly the system can generate the first token of output, reflecting the need for rapid responses to user prompts.

A third new test assessed the speed of processing graph neural networks, which handle complex relationships among entities, like those in a social network. These networks have become increasingly vital in generative AI, exemplified by Google's DeepMind unit's use of graph nets in its AlphaFold 2 model, which made significant strides in protein-folding predictions in 2021. Additionally, a fourth test gauged the speed at which LiDAR sensing data can be compiled into an automobile's road map, using a custom neural net developed by MLCommons from existing open-source technologies.

MLCommons

MLCommons

The MLPerf competition involves computers built by Lenovo, HPE, and others, adhering to stringent requirements for the accuracy of neural net outputs. Each system reports its top speed in producing output per second, with some benchmarks measuring average latency, or the time taken for a response to come back from the server.

Nvidia's GPUs excelled in nearly all tests within the closed division, where the software setup rules are the strictest.

MLCommons

MLCommons

However, AMD, with its MI300X GPU, claimed the top score in two Llama 2 70b tests, achieving 103,182 tokens per second, which was significantly better than Nvidia's newer Blackwell GPU. This winning AMD system was assembled by MangoBoost, a startup specializing in plug-in cards that enhance data transfer between GPU racks, and LLMboost, their software designed to improve generative AI performance.

Nvidia contested the comparison of AMD's results to their Blackwell scores, pointing out the need to adjust for the number of chips and computer "nodes" used in each system. Dave Salvator, Nvidia's director of accelerated computing products, emphasized in an email to ZDNET:

"MangoBoost's results do not reflect an accurate performance comparison against NVIDIA's results. AMD's testing applied 4X the number of GPUs – 32 MI300X GPUs – against 8 NVIDIA B200s, yet still only achieved a 3.83% higher result than the NVIDIA submission. NVIDIA's 8x B200 submission actually outperformed MangoBoost's x32 AMD MI300X GPUs in the Llama 2 70B server submission."

Google also entered the competition, showcasing its Trillium chip, the sixth iteration of its in-house Tensor Processing Unit (TPU). However, it significantly lagged behind Nvidia's Blackwell in a test measuring query response speed for the Stable Diffusion image-generation test.

The latest MLPerf benchmarks saw fewer competitors challenging Nvidia compared to previous rounds. Notably absent were submissions from Intel's Habana unit and Qualcomm, both of which had participated in past years.

Despite this, Intel had reason to celebrate. In the datacenter closed division, Intel's Xeon microprocessor powered seven of the top 11 systems, outperforming AMD's EPYC server microprocessor, which secured only three victories. This marks an improvement for Intel compared to previous years.

The 11th top-performing system, tasked with processing Meta's massive Llama 3.1 405b, was built by Nvidia without using an Intel or AMD microprocessor. Instead, it utilized the integrated Grace-Blackwell 200 chip, combining Nvidia's Blackwell GPU with its own Grace microprocessor in a single package.

Related article
Gen Z's Reality Check: AI Fakery & Critical Thinking Gen Z's Reality Check: AI Fakery & Critical Thinking In today's fast-paced digital world, Generation Z, having grown up with the internet at their fingertips, faces unique challenges. The rise of AI-generated content, which can be ha
Alex Hormozi Reveals Secrets to Cold Email Success in Lead Generation Alex Hormozi Reveals Secrets to Cold Email Success in Lead Generation If you're eager to boost your client acquisition strategy, mastering cold emailing could be your game-changer. Yet, many businesses find it challenging to reap substantial benefits from their cold email campaigns. In this guide, we'll dive into actionable tips inspired by experts like Alex Hormozi t
Guide to Unlocking Storytelling Magic with Gen AI Tools Guide to Unlocking Storytelling Magic with Gen AI Tools In today's fast-paced digital world, the art of storytelling has never been more crucial. Whether you're managing projects, marketing products, or simply expressing your creativity, storytelling can truly engage your audience and foster meaningful interactions. The advent of generative AI is transfo
Comments (40)
0/200
JustinScott
JustinScott April 17, 2025 at 12:00:00 AM EDT

Nvidia's chips are just crushing it in the AI world! I mean, who else can keep up with their performance in generative AI? It's like watching a one-sided race, but hey, if you're into tech, you gotta appreciate the dominance. Maybe it's time for the others to step up their game! 🚀

WillGarcía
WillGarcía April 18, 2025 at 12:00:00 AM EDT

NvidiaのチップはAI分野で圧倒的ですね!生成AIでのパフォーマンスは他社が追いつけないレベル。まるで一方的なレースを見ているようですが、テクノロジーに興味があるなら、この優位性を評価せざるを得ません。他の会社も頑張ってほしいですね!🚀

DonaldSanchez
DonaldSanchez April 17, 2025 at 12:00:00 AM EDT

Nvidia의 칩은 AI 분야에서 정말 압도적이에요! 생성 AI에서의 성능은 다른 회사들이 따라잡을 수 없는 수준이에요. 마치 일방적인 경주를 보는 것 같지만, 기술에 관심이 있다면 이 우위를 인정하지 않을 수 없어요. 다른 회사들도 힘내야겠죠! 🚀

BrianThomas
BrianThomas April 17, 2025 at 12:00:00 AM EDT

Os chips da Nvidia estão dominando o mundo da IA! Quer dizer, quem mais consegue acompanhar o desempenho deles em IA generativa? É como assistir a uma corrida unilateral, mas, ei, se você gosta de tecnologia, tem que apreciar essa dominância. Talvez seja hora dos outros aumentarem o jogo! 🚀

JustinAnderson
JustinAnderson April 17, 2025 at 12:00:00 AM EDT

¡Los chips de Nvidia están dominando el mundo de la IA! Quiero decir, ¿quién más puede seguir su rendimiento en IA generativa? Es como ver una carrera unilateral, pero, oye, si te gusta la tecnología, tienes que apreciar esta dominancia. ¡Quizás es hora de que los demás suban su juego! 🚀

JuanLopez
JuanLopez April 17, 2025 at 12:00:00 AM EDT

Nvidia's GPU chips are just unreal! They absolutely crushed it in the gen AI benchmarks. I mean, who even comes close? It's like watching a race where one car laps the others twice. Still, I wish they'd focus more on energy efficiency too. 🤓🔥

Back to Top
OR