News from the AI & ML world

DeeperML - #blackwell

@futurumgroup.com //
NVIDIA is making significant strides in the development of AI infrastructure and solutions, partnering with key industry players to advance AI capabilities across various sectors. Dell Technologies has been awarded a contract to build the next-generation NERSC-10 supercomputer, designated Doudna, powered by the NVIDIA Vera Rubin architecture. This system, named after Nobel laureate Jennifer Doudna, will integrate AI, quantum workflows, and real-time data streaming, aiming to serve 11,000 researchers in fields such as fusion energy, astronomy, and life sciences. The Doudna supercomputer will deliver 10 times the scientific output of its predecessor while improving power efficiency by 3-5 times.

NVIDIA is also addressing the China AI market with the development of export rules-compliant chips in collaboration with AMD. This move comes as the US government has placed restrictions on chip exports to China. Despite these challenges, NVIDIA CEO Jensen Huang has emphasized the importance of maintaining a presence in China, where a significant portion of AI developers are located. Huang advocates for the adoption of American technology stacks, highlighting the strategic importance of NVIDIA's CUDA development platform in the Chinese market. These efforts underscore NVIDIA's commitment to navigating trade regulations while continuing to foster AI innovation globally.

Furthermore, Dell is expanding its AI offerings with the Pro Max Plus workstation laptop, featuring a discrete enterprise-grade NPU (Neural Processing Unit). This device enables enterprises to run large AI models locally, bypassing the cloud for enhanced control, privacy, and cost-efficiency. The Pro Max Plus can support models with up to 109 billion parameters on-device, catering to AI developers, engineers, and data scientists working with proprietary data. This initiative reflects a growing demand for secure, high-performance AI environments that can be deployed in portable, flexible form factors, further solidifying NVIDIA's role in driving advancements in AI infrastructure and accessibility.

Recommended read:
References :
  • insideAI News: NVIDIA and AMD Devising Export Rules-Compliant Chips for China AI Market
  • futurumgroup.com: Can Dell and NVIDIA’s AI Factory 2.0 Solve Enterprise-Scale AI Infrastructure Gaps?
  • TechHQ: Dell to build Nvidia Vera Rubin supercomputer for US Energy Department
  • techhq.com: Dell to build Nvidia Vera Rubin supercomputer for US Energy Department
  • futurumgroup.com: Can Dell Challenge Public Cloud AI with Its Expanded AI Factory?
  • insidehpc.com: DOE Announces “Doudnaâ€� Dell-NVIDIA Supercomputer at NERSC
  • techxplore.com: US supercomputer named after Nobel laureate Jennifer Doudna to power AI and scientific research

Aaron Klotz@tomshardware.com //
References: insideAI News , techvro.com , ServeTheHome ...
NVIDIA has recently announced a significant breakthrough in AI inference, achieving a new world record with its DGX B200 Blackwell node. This node, equipped with eight Blackwell GPUs, surpassed the 1,000 tokens per second (TPS) per user barrier while running Meta’s Llama 4 Maverick large language model. According to a report by Artificial Analysis, the DGX B200 node achieved 1,038 TPS/user, outperforming previous record holders like SambaNova, who achieved 792 TPS/user. This advancement showcases the immense capabilities of the Blackwell architecture and sets a new standard for AI performance.

NVIDIA achieved this record-breaking performance through extensive software optimizations, utilizing TensorRT and Eagle-3 techniques for speculative decoding. These optimizations resulted in a 4x performance uplift compared to Blackwell's prior best results. Further enhancements involved using FP8 data types, Attention operations, and the Mixture of Experts (MoE) AI technique. These improvements not only boosted speed but also maintained response accuracy. NVIDIA's Blackwell GPUs reached 72,000 TPS/server at their highest throughput configuration.

In addition to AI performance, NVIDIA is also revolutionizing AI data center infrastructure through a collaboration with Navitas Semiconductor. They are introducing a new 800V HVDC architecture designed to replace the aging 54V systems currently in use. This new architecture is expected to deliver up to 5% better power efficiency and 70% lower maintenance costs. The transition to 800V power enables a 45% reduction in copper wire thickness, significantly lowering material use and weight, while reducing heat and fewer losses.

Recommended read:
References :
  • insideAI News: AI Inference: NVIDIA Reports Blackwell Surpasses 1000 TPS/User Barrier with Llama 4 Maverick
  • techvro.com: NVIDIA collaborates with Navitas to usher in a new era for AI data center infrastructure. Their joint 800V HVDC architecture replaces the...
  • www.tomshardware.com: Nvidia has broken another AI world record, breaking over 1,000 TPS/user with a DGX B200 node boasting eight Blackwell GPUs inside.
  • ServeTheHome: Nvidia has broken another AI world record, breaking over 1,000 TPS/user with a DGX B200 node boasting eight Blackwell GPUs inside.

@blogs.nvidia.com //
Nvidia is significantly expanding its AI infrastructure initiatives by introducing NVLink Fusion, a technology that allows for the integration of non-Nvidia CPUs and AI accelerators with Nvidia's GPUs. This strategic move aims to provide customers with more flexible and customizable AI system configurations, broadening Nvidia's reach in the rapidly growing data center market. Key partnerships are already in place with companies like Qualcomm, Fujitsu, Marvell, and MediaTek, as well as design software firms Cadence and Synopsys, to foster a robust and open ecosystem. This approach allows Nvidia to remain central to the future of AI infrastructure, even when systems incorporate chips from other vendors.

Nvidia is also solidifying its presence in Taiwan, establishing a new office complex near Taipei that will serve as its overseas headquarters. The company is collaborating with Foxconn to build an "AI factory" in Taiwan, which will utilize 10,000 Nvidia Blackwell GPUs. This facility is intended to bolster Taiwan's AI infrastructure and support local organizations in adopting AI technologies across various sectors. TSMC, Nvidia's primary chip supplier, plans to leverage this supercomputer for research and development, aiming to develop the next generation of AI chips.

Furthermore, Nvidia is working with Taiwan's National Center for High-Performance Computing (NCHC) to develop a new AI supercomputer. This system will feature over 1,700 GPUs, GB200 NVL72 rack-scale systems, and an HGX B300 system based on the Blackwell Ultra platform, all connected via Quantum InfiniBand networking. Expected to launch later this year, the supercomputer promises an eightfold performance increase over its predecessor for AI workloads, providing researchers with enhanced capabilities to advance their projects. Academic institutions, government agencies, and small businesses in Taiwan will be able to apply for access to the supercomputer to accelerate their AI initiatives.

Recommended read:
References :
  • techvro.com: NVLink Fusion: Nvidia To Sell Hybrid Systems Using AI Chips
  • The Register - Software: Nvidia sets up shop in Taiwan with AI supers and a factory full of ambition
  • NVIDIA Newsroom: NVIDIA CEO Envisions AI Infrastructure Industry Worth ‘Trillions of Dollars’
  • AIwire: Nvidia’s Global Expansion: AI Factories, NVLink Fusion, AI Supercomputers, and More

@blogs.nvidia.com //
Cadence has unveiled the Millennium M2000 Supercomputer, a powerhouse featuring NVIDIA Blackwell systems, aimed at revolutionizing AI-driven engineering design and scientific simulations. This supercomputer integrates NVIDIA HGX B200 systems and NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs, coupled with NVIDIA CUDA-X software libraries and Cadence's optimized software. The result is a system capable of delivering up to 80 times higher performance compared to its CPU-based predecessors, marking a significant leap forward in computational capability for electronic design automation, system design, and life sciences workloads.

This collaboration between Cadence and NVIDIA is set to enable engineers to conduct massive simulations, leading to breakthroughs in various fields, including the design and development of autonomous machines, drug molecules, semiconductors, and data centers. NVIDIA's founder and CEO, Jensen Huang, highlighted the transformative potential of AI, stating that it will infuse every aspect of business and product development. Huang also announced NVIDIA's plans to acquire ten Millennium Supercomputer systems based on the NVIDIA GB200 NVL72 platform to accelerate the company’s chip design workflows, emphasizing the importance of this technology for NVIDIA's future endeavors.

In related news, the open-source OpenSearch software has launched its 3.0 version, which includes GPU acceleration to enhance AI workloads through its new OpenSearch Vector Engine. This update leverages NVIDIA GPUs to improve search performance with large-scale vector workloads and reduce index build times, aiming to address scalability issues common in vector databases. OpenSearch 3.0 also supports Anthropic PBC’s Model Context Protocol, facilitating the integration of large language models with external data. The Millennium M2000 Supercomputer harnesses accelerated software from NVIDIA and Cadence for applications including circuit simulation, computational fluid dynamics, data center design and molecular design.

Recommended read:
References :
  • NVIDIA Newsroom: Cadence Taps NVIDIA Blackwell to Accelerate AI-Driven Engineering Design and Scientific Simulation
  • www.networkworld.com: Cadence debuts Nvidia-powered supercomputer to accelerate enterprise engineering, biotech
  • insidehpc.com: Cadence Unveils Millennium M2000 Supercomputer with NVIDIA Blackwell Systems
  • Ken Yeung: ServiceNow and Nvidia Debut Apriel Nemotron 15B, an Open-Source Reasoning Model Built for Faster, Cheaper Agentic AI

@blogs.nvidia.com //
NVIDIA's Blackwell platform is set to revolutionize data center cooling with a focus on water efficiency. The new platform introduces direct-to-chip liquid cooling, which dramatically reduces water consumption compared to traditional air-cooled systems. NVIDIA claims this innovative approach offers a 300x improvement in water efficiency. This is crucial as the increasing compute power required by AI and HPC applications is driving a global shift towards larger, more power-hungry data centers and AI factories.

This cooling solution addresses the escalating energy demands and environmental concerns associated with AI infrastructure. Historically, cooling has accounted for up to 40% of a data center's electricity consumption. The Blackwell platform's liquid cooling technology captures heat directly at the source, cycling it through a coolant distribution unit and liquid-to-liquid heat exchanger, before transferring it to a facility cooling loop. This method allows data centers to operate effectively at warmer water temperatures, reducing or eliminating the need for mechanical chillers in many climates, leading to significant cost savings and reduced energy consumption.

The NVIDIA GB200 NVL72 and GB300 NVL72 systems, built on the Blackwell platform, utilize this direct-to-chip liquid cooling technology. Unlike evaporative or immersion cooling, this is a closed-loop system, meaning the coolant doesn't evaporate or require replacement due to loss from phase change, further conserving water. These rack-scale systems are designed to handle the demanding tasks of trillion-parameter large language model inference, making them ideal for running AI reasoning models while efficiently managing energy costs and heat. This advancement not only ensures optimal performance of AI servers but also promotes a more sustainable and environmentally friendly AI infrastructure.

Recommended read:
References :
  • NVIDIA Newsroom: Chill Factor: NVIDIA Blackwell Platform Boosts Water Efficiency by Over 300x
  • blogs.nvidia.com: Chill Factor: NVIDIA Blackwell Platform Boosts Water Efficiency by Over 300x
  • www.tomshardware.com: Nvidia aims to solve AI's water consumption problems with direct-to-chip cooling — claims 300X improvement with closed-loop systems

@hothardware.com //
References: hothardware.com , insideAI News ,
Nvidia's Blackwell architecture is making significant strides in both the AI and gaming sectors. The GeForce RTX 5060 Ti, priced at $429, brings the Blackwell architecture to mainstream gamers, targeting 1440p gaming with power efficiency and overclocking headroom. Reviews indicate that the RTX 5060 Ti is built around the GB206 GPU and is the full implementation of the chip, with roughly 21.9 billion transistors manufactured on TSMC's 4N node. It is arranged into 3 GPCs. It is the first of the Blackwell cards to target under $500 and should see peak performance realized in advanced AI rendering techniques like DLSS4 with multi-frame generation.

CoreWeave, a GPU cloud platform, is among the first to bring Nvidia's Grace Blackwell GB200 NVL72 systems online at scale. Companies like Cohere, IBM, and Mistral AI are leveraging these systems for model training and deployment. Cohere is using Grace Blackwell Superchips to develop secure enterprise AI applications, with reported performance increases of up to 3x in training for 100 billion-parameter models. IBM is scaling its deployment to thousands of Blackwell GPUs on CoreWeave to train its Granite open-source AI models for IBM watsonx Orchestrate. Mistral AI is also utilizing the Blackwell GPUs to build the next generation of open-source AI models, reporting a 2x improvement in performance for dense model training.

However, Nvidia faces challenges due to U.S. government restrictions on exports to China. The company is writing off $5.5 billion in GPUs as the U.S. government chokes off supply of H20s to China, highlighting geopolitical impacts on the tech industry. The U.S. government's concern stems from the potential use of these processors in Chinese supercomputers. In response, Nvidia is reportedly working to onshore the manufacturing of chips and other components, as well as the assembly of systems, to the United States.

Recommended read:
References :
  • hothardware.com: NVIDIA GeForce RTX 5060 Ti 16GB: MSRP $429 The new GeForce RTX 5060 Ti brings NVIDIA's latest Blackwell GPU architecture down below $500, to target mainstream, 1440p gamers.
  • insideAI News: NVIDIA announced that GPU cloud platform CoreWeave is among the first cloud providers to bring NVIDIA GB200 NVL72 systems online at scale, with Cohere, IBM and Mistral AI using them for model training and deployment.
  • www.tomshardware.com: Apparently, Jensen Huang's $1 million dinner with Donald Trump did not help to ease U.S. government's concerns over H20 GPUs exports to China.

NVIDIA Newsroom@NVIDIA Blog //
Nvidia has announced a major initiative to manufacture its AI supercomputers entirely within the United States. The company aims to produce up to $500 billion worth of AI infrastructure in the U.S. over the next four years, partnering with major manufacturing firms like Taiwan Semiconductor Manufacturing Co (TSMC), Foxconn, Wistron, Amkor, and SPIL. This move marks the first time Nvidia will carry out chip packaging and supercomputer assembly entirely within the United States. The company sees this effort as a way to meet the increasing demand for AI chips, strengthen its supply chain, and boost resilience.

Nvidia is commissioning over a million square feet of manufacturing space to build and test Blackwell chips in Arizona and assemble AI supercomputers in Texas. Production of Blackwell chips has already begun at TSMC’s chip plants in Phoenix, Arizona. The company is also constructing supercomputer manufacturing plants in Texas, partnering with Foxconn in Houston and Wistron in Dallas, with mass production expected to ramp up within the next 12-15 months. These facilities are designed to support the deployment of "gigawatt AI factories", data centers specifically built for processing artificial intelligence.

CEO Jensen Huang emphasized the significance of bringing AI infrastructure manufacturing to the U.S., stating that "The engines of the world’s AI infrastructure are being built in the United States for the first time." Nvidia also plans to deploy its own technologies to optimize the design and operation of the new facilities, utilizing platforms like Omniverse to simulate factory operations and Isaac GR00T to develop automated robotics systems. The company said domestic production could help drive long-term economic growth and job creation.

Recommended read:
References :
  • Reid Burke: NVIDIA is working with its manufacturing partners to design and build factories that, for the first time, will produce NVIDIA AI supercomputers entirely in the U.S.
  • The Register - Software: Nvidia wants to build and sell up to half a trillion US dollars of American-made AI supercomputer equipment over the next four years, with the help of Taiwan Semiconductor Manufacturing Co, aka TSMC, and its partners.
  • TechInformed: Nvidia has announced plans to manufacture AI supercomputers in the United States for the first time.
  • AIwire: Nvidia Begins US Production of Blackwell Chips, AI Systems to Follow
  • www.tomshardware.com: Nvidia aims to build $500 billion worth of AI servers in the USA by 2029
  • www.techrepublic.com: NVIDIA’s Vision For AI Factories – ‘Major Trend in the Data Center World’
  • www.tomshardware.com: Made in the USA: Inside Nvidia's $500 billion server gambit
  • www.theguardian.com: Jensen Huang causes stir on social media and is reported to have met founder of AI company DeepSeek The chief executive of the American chip maker Nvidia visited Beijing on Thursday, days after the US on sales of the only AI chip it was still allowed to sell to China.

NVIDIA Newsroom@NVIDIA Blog //
Nvidia has announced plans to manufacture its AI supercomputers entirely within the United States, marking the first time the company will conduct chip packaging and supercomputer assembly domestically. The move, driven by increasing global demand for AI chips and the potential impact of tariffs, aims to establish a resilient supply chain and bolster the American AI ecosystem. Nvidia is partnering with major manufacturing firms including TSMC, Foxconn, and Wistron to construct and operate these facilities.

Mass production of Blackwell chips has already commenced at TSMC's Phoenix, Arizona plant. Nvidia is constructing supercomputer manufacturing plants in Texas, partnering with Foxconn in Houston and Wistron in Dallas. These facilities are expected to ramp up production within the next 12-15 months. More than a million square feet of manufacturing space has been commissioned to build and test NVIDIA Blackwell chips in Arizona and AI supercomputers in Texas.

The company anticipates producing up to $500 billion worth of AI infrastructure in the U.S. over the next four years through these partnerships. This includes designing and building "gigawatt AI factories" to produce NVIDIA AI supercomputers completely within the US. CEO Jensen Huang stated that American manufacturing will help meet the growing demand for AI chips and supercomputers, strengthen the supply chain and improve resiliency. The White House has lauded Nvidia's decision as "the Trump Effect in action".

Recommended read:
References :
  • Reid Burke: NVIDIA to Manufacture American-Made AI Supercomputers in US for First Time
  • insideAI News: NVIDIA today said it is working with manufacturing partners to design and build factories that will produce NVIDIA AI supercomputers — i.e., “AI factories†— entirely in the United States… NVIDIA said that within four years, it plans to produce up to half a trillion dollars worth of AI infrastructure in the U.S. through partnerships ....
  • AIwire: Nvidia Begins US Production of Blackwell Chips, AI Systems to Follow
  • www.theguardian.com: Nvidia says it will build up to $500bn of US AI infrastructure as chip tariff looms
  • www.tomshardware.com: Nvidia aims to build $500 billion worth of AI servers in the USA by 2029
  • Analytics India Magazine: NVIDIA to Manufacture First American-Made AI Supercomputers
  • THE DECODER: Nvidia shifts AI production to US amid changing trade landscape
  • AI News | VentureBeat: Nvidia pledges to build its own factories in the U.S. for the first time to make AI supercomputers
  • www.cnbc.com: Nvidia to mass produce AI supercomputers in Texas as part of $500 billion U.S. push
  • the-decoder.com: Nvidia shifts AI production to US amid changing trade landscape
  • NVIDIA Newsroom: Everywhere, All at Once: NVIDIA Drives the Next Phase of AI Growth
  • blogs.nvidia.com: NVIDIA to Manufacture American-Made AI Supercomputers in US for First Time
  • CIO Dive - Latest News: Nvidia pledges to invest up to $500B in US chip manufacturing
  • analyticsindiamag.com: NVIDIA to Manufacture First American-Made AI Supercomputers

NVIDIA Newsroom@NVIDIA Blog //
Nvidia has announced plans to manufacture AI supercomputers entirely within the United States for the first time. The company is working with manufacturing partners to design and build factories that will produce NVIDIA AI supercomputers – or "AI factories" – on U.S. soil. This initiative includes a projected investment of up to $500 billion over the next four years and aims to establish a comprehensive domestic supply chain for AI infrastructure. Nvidia's partners include industry giants such as TSMC, Foxconn, Wistron, Amkor, and SPIL, deepening their ties with NVIDIA while expanding their global footprint and enhancing supply chain resilience.

The company has already commissioned over a million square feet of manufacturing space in Arizona and Texas to build and test NVIDIA Blackwell chips and assemble AI supercomputers. NVIDIA Blackwell chips have started production at TSMC’s chip plants in Phoenix, Arizona. NVIDIA is building supercomputer manufacturing plants in Texas, with Foxconn in Houston and with Wistron in Dallas. Mass production at both plants is expected to ramp up in the next 12-15 months. Amkor and SPIL will handle packaging and testing operations in Arizona. This move marks the first time Nvidia will be building AI supercomputers entirely in the US.

Jensen Huang, founder and CEO of NVIDIA, emphasized the strategic importance of this initiative. "The engines of the world's AI infrastructure are being built in the United States for the first time," Huang stated. "Adding American manufacturing helps us better meet the incredible and growing demand for AI chips and supercomputers, strengthens our supply chain, and boosts our resiliency." NVIDIA also intends to deploy its own technologies, such as Omniverse and Isaac GR00T, to optimize factory operations and automate manufacturing processes. Manufacturing NVIDIA AI chips and supercomputers for American AI factories is expected to create hundreds of thousands of jobs and drive trillions of dollars in economic security over the coming decades.

Recommended read:
References :
  • Reid Burke: NVIDIA to Manufacture American-Made AI Supercomputers in US for First Time
  • www.tomshardware.com: Nvidia aims to build $500 billion worth of AI servers in the USA by 2029
  • insidehpc.com: NVIDIA to Manufacture AI Supercomputers in U.S.
  • PCMag Middle East ai: To Avoid Tariffs, Nvidia to Build Supercomputer Factories in Texas
  • AI News | VentureBeat: Nvidia pledges to build its own factories in the U.S. for the first time to make AI supercomputers
  • Maginative: NVIDIA has announced plans to manufacture AI supercomputers entirely within the United States for the first time, committing up to half a trillion dollars to domestic production through partnerships with TSMC, Foxconn, and others.
  • insideAI News: NVIDIA today said it is working with manufacturing partners to design and build factories that will produce NVIDIA AI supercomputers — i.e., “AI factories†— entirely in the United States
  • Flipboard Tech Desk: Nvidia says it has commissioned more than a million square feet of manufacturing space to build and test AI chips in Arizona and Texas as part of an effort to move a portion of production to the U.S.
  • Policy ? Ars Technica: Amid Trump tariff chaos, Nvidia launches AI chip production on US soil
  • www.cnbc.com: Nvidia to mass produce AI supercomputers in Texas as part of $500 billion U.S. push
  • analyticsindiamag.com: NVIDIA to Manufacture First American-Made AI Supercomputers
  • The Register - Software: Nvidia joins made-in-America party, hopes to flog $500B in homegrown AI supers by 2029
  • AIwire: Nvidia Begins US Production of Blackwell Chips, AI Systems to Follow
  • www.theguardian.com: The Guardian article on Nvidia's plan to build AI infrastructure in the US.
  • TechInformed: Nvidia has announced plans to manufacture AI supercomputers in the United States for the first time. Over the next four years, the company said that… The post appeared first on .
  • the-decoder.com: Nvidia shifts AI production to US amid changing trade landscape
  • NVIDIA Newsroom: Everywhere, All at Once: NVIDIA Drives the Next Phase of AI Growth
  • blogs.nvidia.com: NVIDIA to Manufacture American-Made AI Supercomputers in US for First Time
  • CIO Dive - Latest News: Article on Nvidia's investment of up to $500 billion in US chip manufacturing.
  • blogs.nvidia.com: Every company and country wants to grow and create economic opportunity — but they need virtually limitless intelligence to do so.
  • insidehpc.com: Nvidia's plans to manufacture AI supercomputers in the US is a move to avoid tariffs and strengthen its supply chain.
  • www.tomshardware.com: Nvidia's ambitious plan to establish a domestic AI supply chain in the US looks challenging but could reshape the AI industry.
  • techdator.net: NVIDIA Teams Up with U.S. Companies to Build AI Supercomputers at Home
  • THE DECODER: Nvidia shifts AI production to US amid changing trade landscape
  • www.techrepublic.com: During a keynote at Data Center World 2025, a NVIDIA exec spoke about AI factories, purpose-built for gigawatt-scale data centers.
  • techinformed.com: Nvidia to manufacture AI supercomputers in the US

@www.intel.com //
NVIDIA's Blackwell platform has dominated the latest MLPerf Inference V5.0 benchmarks, showcasing significant performance improvements in AI reasoning. The NVIDIA GB200 NVL72 system, featuring 72 Blackwell GPUs, achieved up to 30x higher throughput on the Llama 3.1 405B benchmark compared to the NVIDIA H200 NVL8 submission. This was driven by more than triple the performance per GPU and a 9x larger NVIDIA NVLink interconnect domain. The latest MLPerf results reflect the shift toward reasoning in AI inference.

Alongside this achievement, NVIDIA is open-sourcing the KAI Scheduler, a Kubernetes GPU scheduling solution, as part of its commitment to open-source AI innovation. Previously a core component of the Run:ai platform, KAI Scheduler is now available under the Apache 2.0 license. This solution is designed to address the unique challenges of managing AI workloads that utilize both GPUs and CPUs. According to NVIDIA, this will help in managing fluctuating GPU demands, which traditional resource schedulers struggle to handle.

Recommended read:
References :
  • NVIDIA Newsroom: Speed Demon: NVIDIA Blackwell Takes Pole Position in Latest MLPerf Inference Results
  • IEEE Spectrum: Nvidia Blackwell Ahead in AI Inference, AMD Second
  • AIwire: MLPerf v5.0 Reflects the Shift Toward Reasoning in AI Inference
  • Developer Tech News: KAI Scheduler: NVIDIA open-sources Kubernetes GPU scheduler

@www.intel.com //
NVIDIA is making strides in both agentic AI and open-source initiatives. Jacob Liberman, director of product management at NVIDIA, explains how agentic AI bridges the gap between powerful AI models and practical enterprise applications. Enterprises are now deploying AI agents to free human workers from time-consuming and error-prone tasks, allowing them to focus on high-value work that requires creativity and strategic thinking. NVIDIA AI Blueprints help enterprises build their own AI agents.

NVIDIA has announced the open-source release of the KAI Scheduler, a Kubernetes-native GPU scheduling solution, now available under the Apache 2.0 license. Originally developed within the Run:ai platform, the KAI Scheduler is now available to the community while also continuing to be packaged and delivered as part of the NVIDIA Run:ai platform. The KAI Scheduler is designed to optimize the scheduling of GPU resources and tackle challenges associated with managing AI workloads on GPUs and CPUs.

Recommended read:
References :
  • NVIDIA Newsroom: Speed Demon: NVIDIA Blackwell Takes Pole Position in Latest MLPerf Inference Results
  • NVIDIA Technical Blog: The compute demands for large language model (LLM) inference are growing rapidly, fueled by the combination of growing model sizes, real-time latency...
  • IEEE Spectrum: Nvidia Blackwell Ahead in AI Inference, AMD Second
  • insidehpc.com: MLCommons Releases New MLPerf Inference v5.0 Benchmark Results
  • www.networkworld.com: Nvidia’s Blackwell raises the bar with new MLPerf Inference V5.0 results
  • NVIDIA Newsroom: AI is rapidly transforming how organizations solve complex challenges. The early stages of enterprise AI adoption focused on using large language models to create chatbots. Now, enterprises are using agentic AI to create intelligent systems that reason, act and execute complex tasks with a degree of autonomy.
  • NVIDIA Technical Blog: Today, NVIDIA announced the open-source release of the KAI Scheduler, a Kubernetes-native GPU scheduling solution, now available under the Apache 2.0 license.
  • insideAI News: Today, NVIDIA posted a blog announcing the open-source release of the KAI Scheduler, a Kubernetes-native GPU scheduling solution, now available under the Apache 2.0 license.
  • Developer Tech News: NVIDIA has open-sourced KAI Scheduler, a Kubernetes solution designed to optimise the scheduling of GPU resources.

Jaime Hampton@BigDATAwire //
NVIDIA's GTC 2025 showcased significant advancements in AI, marked by the unveiling of the Blackwell Ultra GPU and the Vera Rubin roadmap extending through 2027. CEO Jensen Huang emphasized a 40x AI performance leap with the Blackwell platform compared to its predecessor, Hopper, highlighting its crucial role in inference workloads. The conference also introduced open-source ‘Dynamo’ software and advancements in humanoid robotics, demonstrating NVIDIA’s commitment to pushing AI boundaries.

The Blackwell platform is now in full production, meeting incredible customer demand, and the Vera Rubin roadmap details the next generation of superchips expected in 2026. Huang also touted new DGX systems, highlighting the push towards photonic switches to handle growing data demands efficiently. Blackwell Ultra will offer 288GB of memory. NVIDIA claims the GB300 chip brings 1.5x more AI performance than the NVIDIA GB200. These advancements aim to bolster AI reasoning capabilities and energy efficiency, positioning NVIDIA to maintain its dominance in AI infrastructure.

Recommended read:
References :
  • AI News | VentureBeat: Nvidia launches Blackwell RTX Pro for workstations and servers
  • AIwire: Nvidia’s DGX AI Systems Are Faster and Smarter Than Ever
  • Analytics Vidhya: 10 NVIDIA GTC 2025 Announements that You Must Know
  • venturebeat.com: Nvidia’s GTC 2025 keynote: 40x AI performance leap, open-source ‘Dynamo’, and a walking Star Wars-inspired ‘Blue’ robot
  • BigDATAwire: Nvidia Touts Next Generation GPU Superchip and New Photonic Switches
  • www.laptopmag.com: "I'm the chief revenue destroyer": Nvidia's Jensen Huang says new Blackwell chips make previous-gen feel obsolete
  • AIwire: The GTC 2025 happening in San Jose, Calif., has become one of the marquee events in the tech world. It has grabbed the attention of everyone from industry leaders and developers to AI enthusiasts and even those who remain skeptical about AI’s potential.
  • BigDATAwire: Nvidia unveils updates to its AI infrastructure portfolio, including its next-generation datacenter GPU, the NVIDIA Blackwell Ultra.
  • BigDATAwire: Nvidia Preps for 100x Surge in Inference Workloads, Thanks to Reasoning AI Agents
  • Gradient Flow: Overview As I sat watching Jensen Huang’s keynote at Nvidia’s recent GTC, I was struck once again by how this annual event has evolved from a graphics card showcase into something far more consequential for global markets.
  • Analytics Vidhya: Nvidia’s annual GPU Technology Conference (GTC) has long been a highlight for the AI community. At this year’s event, Nvidia CEO Jensen Huang unveiled a roadmap of new products and innovations aimed at scaling up artificial intelligence.
  • NVIDIA Newsroom: AI Factories Are Redefining Data Centers and Enabling the Next Era of AI
  • AIwire: The GTC 2025 happening in San Jose, Calif., has become one of the marquee events in the tech world.
  • BigDATAwire: The boundaries between artificial intelligence and the physical world are dissolving. AI systems are becoming increasingly adept at perceiving, interacting, analyzing, and responding to their physical environments. In this AI The post appeared first on .
  • Data Phoenix: NVIDIA's new Blackwell Ultra platform delivers significantly enhanced AI computing power for reasoning and agentic AI applications.
  • Maginative: How NVIDIA Is Building the Operating System for Physical AI
  • insideAI News: Highlights from the Nvidia extravaganza with an AI-everywhere theme. We review the conference, discussing everything from the new AI compute industry landscape
  • The Next Platform: Reports on how Nvidia is turning its AI eye to the enterprise.
  • Last Week in AI: Nvidia's GTC 2025 keynote focused on the transition to AI-driven computing and the development of AI factories.
  • BigDATAwire: Reporter’s Notebook: AI Hype and Glory at Nvidia GTC 2025