MLPerf December Scores Reorder AI Silicon: Nvidia H200 Leads as AMD MI300X, Google Trillium Tighten Race

Fresh MLPerf results released in December redraw the AI accelerator leaderboard. Nvidia's H200 leads training throughput, while AMD's MI300X and Google's Trillium TPU close the gap on inference efficiency and scaling, with AWS Trainium2 entering the chart on price-performance.

Published: January 5, 2026 By Aisha Mohammed Category: AI Chips
MLPerf December Scores Reorder AI Silicon: Nvidia H200 Leads as AMD MI300X, Google Trillium Tighten Race

Executive Summary

  • MLCommons’ December MLPerf release highlights Nvidia H200 leadership in training throughput, with AMD MI300X and Google Trillium TPU narrowing the inference efficiency gap (MLCommons).
  • Vendor disclosures indicate 10–25% gains gen-over-gen across key LLM and vision tests, led by memory bandwidth and compiler optimizations (Nvidia, AMD, Google Cloud).
  • AWS brings Trainium2 into public benchmarks with competitive cost-per-token and strong scaling on 256–1024-accelerator clusters (AWS News Blog).
  • Analysts flag tokens-per-joule and cluster-scale efficiency as primary enterprise buying criteria for 2026 AI buildouts (IDC, Gartner).

Benchmark Season Reset: What December’s MLPerf Shows

MLCommons’ latest MLPerf results, released in December, provide the clearest snapshot yet of how next-wave accelerators stack up on training and inference for large language models and vision workloads. The submissions feature new silicon and updated software stacks, with a focus on end-to-end throughput, scaling efficiency, and energy use on standardized tasks (MLCommons).

Across the suite, Nvidia’s H200-class systems maintain a lead in raw training throughput on multi-node configurations, a result the company attributes to higher HBM capacity and compiler/runtime updates in its software stack. Vendor commentary points to double-digit percentage gains over prior rounds, particularly on LLM pretraining and reinforcement learning from human feedback (RLHF) stages (Nvidia, Reuters).

Who’s Closing the Gap: AMD, Google, Intel, and AWS

AMD’s MI300X shows marked improvement in inference performance, with submissions and partner data indicating a narrowing gap versus Nvidia on tokens-per-second for popular open-weight models, supported by expanded HBM3 memory and refinements in the ROCm stack. AMD also highlights competitive total cost of ownership (TCO) in large clusters where memory-bound workloads dominate (AMD Newsroom, Bloomberg).

...

Read the full article at AI BUSINESS 2.0 NEWS