News from the AI & ML world

DeeperML

@www.intel.com //
NVIDIA's Blackwell platform has dominated the latest MLPerf Inference V5.0 benchmarks, showcasing significant performance improvements in AI reasoning. The NVIDIA GB200 NVL72 system, featuring 72 Blackwell GPUs, achieved up to 30x higher throughput on the Llama 3.1 405B benchmark compared to the NVIDIA H200 NVL8 submission. This was driven by more than triple the performance per GPU and a 9x larger NVIDIA NVLink interconnect domain. The latest MLPerf results reflect the shift toward reasoning in AI inference.

Alongside this achievement, NVIDIA is open-sourcing the KAI Scheduler, a Kubernetes GPU scheduling solution, as part of its commitment to open-source AI innovation. Previously a core component of the Run:ai platform, KAI Scheduler is now available under the Apache 2.0 license. This solution is designed to address the unique challenges of managing AI workloads that utilize both GPUs and CPUs. According to NVIDIA, this will help in managing fluctuating GPU demands, which traditional resource schedulers struggle to handle.
Original img attribution: https://spectrum.ieee.org/media-library/image.jpg?id=59778851&width=1200&height=600&coordinates=0%2C658%2C0%2C92
ImgSrc: spectrum.ieee.o

Share: bluesky twitterx--v2 facebook--v1 threads


References :
  • NVIDIA Newsroom: Speed Demon: NVIDIA Blackwell Takes Pole Position in Latest MLPerf Inference Results
  • IEEE Spectrum: Nvidia Blackwell Ahead in AI Inference, AMD Second
  • AIwire: MLPerf v5.0 Reflects the Shift Toward Reasoning in AI Inference
  • Developer Tech News: KAI Scheduler: NVIDIA open-sources Kubernetes GPU scheduler
Classification:
  • HashTags: #NVIDIABlackwell #MLPerf #GPU
  • Company: NVIDIA
  • Target: AI Community
  • Product: Blackwell
  • Feature: High Performance
  • Type: AI
  • Severity: Informative