Version 5.0 of MLPerf, which measures the performance of AI infrastructure, has been released, and NVIDIA, AMD, Intel, and others have released the results

MLCommons Releases New MLPerf Inference v5.0 Benchmark Results - MLCommons
https://mlcommons.org/2025/04/mlperf-inference-v5-0-results/

1/ We are excited to announce the latest MLCommons MLPerf Inference v5.0 benchmark suite results. This round featured robust participation from 23 submitting organizations delivering over 17,000 performance results! https://t.co/vlrJcoz25t pic.twitter.com/6V56AjkRUq
— MLCommons (@MLCommons) April 2, 2025
MLPerf Inference is a benchmark that spans both data center and edge systems to measure how fast systems can run AI and machine learning (ML) models across a variety of workloads. MLPerf Inference is an open-source, peer-reviewed benchmark suite designed to create a level playing field to drive innovation, performance, and energy efficiency across the industry.
MLPerf Inference v5.0, the latest version of MLPerf Inference, has newly implemented benchmark tests for 'Llama 3.1 405B,' 'Llama 2 70B Interactive for low-latency applications,' 'RGAT,' and 'Automotive PointPainting for 3D object detection.'
A new benchmark using 'Llama 3.1 405B' has also been introduced. Llama 3.1 405B supports input and output of up to 128,000 tokens, making it possible to incorporate up to 405 billion parameters into AI models. The Llama 3.1 405B benchmark tests three tasks: general question answering, mathematics, and code generation.
'Llama 2 70B Interactive for low-latency applications' adds low-latency requirements to the Llama 2 70B benchmark. This benchmark reflects industry trends such as conversational chatbots, next-generation inference systems, and agent systems, which require the system under test (SUT) to meet more stringent system response metrics such as time to first token (TTFT) and time per output token (TPOT).
'RGAT' is a benchmark for data centers that implements the
'Automotive PointPainting for 3D object detection' is a new benchmark targeted at edge computing devices, specifically automobiles. Announced in summer 2024 as a 'Minimum Viable Product' benchmark, it provides a proxy for an important edge computing scenario: 3D object detection in camera feeds in applications such as self-driving cars.

The MLPerf Inference v5.0 benchmark test using Llama 2 70B measures generative AI inference workloads. The rate of submissions of benchmark scores for this test has increased 2.5 times over the past year, and MLCommons has stated that 'generative AI is gaining momentum.'
Comparing test submission scores using Llama 2 70B with previous versions, the median score was 2x faster and the best score was 3.3x faster than Inference v4.0.

'Much of the ecosystem is tackling the adoption of generative AI head-on, and it's clear that the performance benchmarking feedback loop is at work,' said David Kanter, MLPerf lead at MLCommons. 'We're seeing an unprecedented influx of new accelerators, with hardware paired with new software techniques, including hardware-software alignment support for the FP4 format. These advancements are driving the community to set new records in generative AI inference performance.'
The benchmark results also include benchmark scores for six processors that are either newly released or soon to be shipped:
・AMD Instinct MI325X
・Intel Xeon 6980P
・Google TPU Trillium
・NVIDIA B200
・NVIDIA Jetson AGX Thor 128
・NVIDIA GB200
MLPerf Inference v5.0 received 17,457 performance data submissions from 23 organizations, including AMD, ASUSTeK, Broadcom, Cisco, CoreWeave, CTuning, Dell, FlexAI, Fujitsu, GATEOverflow, Giga Computing, Google, HPE, Intel, Krai, Lambda, Lenovo, MangoBoost, NVIDIA, Oracle, Quanta Cloud Technology, Supermicro, and Sustainable Metal Cloud.
Companies like NVIDIA, AMD, and Intel are touting their MLPerf Inference v5.0 scores on their blogs and newsrooms.
NVIDIA Blackwell Takes Pole Position in Latest MLPerf Inference Results | NVIDIA Blog
https://blogs.nvidia.com/blog/blackwell-mlperf-inference/
AMD Instinct GPUs Continue AI Momentum Across Indu... - AMD Community
https://community.amd.com/t5/instinct-accelerators/amd-instinct-gpus-continue-ai-momentum-across-industry/ba-p/756056
AMD InstinctTM MI325X GPUs Produce Strong Performance in MLPerf Inference v5.0 — ROCm Blogs
https://rocm.blogs.amd.com/artificial-intelligence/mi325x-accelerates-mlperf-inference/README.html
Intel Xeon Remains Only Server CPU on MLPerf - Intel Newsroom
https://newsroom.intel.com/data-center/intel-xeon-remains-only-server-cpu-mlperf
MLCommons also stated, 'The continuing growth of the submission community is a testament to the importance of accurate and reliable performance metrics to the AI community,' and expressed its gratitude to the five new collaborators in MLPerf Inference v5.0: CoreWeave, FlexAI, GATEOverflow, Lambda, and MangoBoost, and its joy at the growth of the community.
Related Posts:
in Software, Posted by logu_ii