News from the AI & ML world

DeeperML - #aichips

@www.marktechpost.com //
Nvidia is reportedly developing a new AI chip, the B30, specifically tailored for the Chinese market to comply with U.S. export controls. This Blackwell-based alternative aims to offer multi-GPU scaling capabilities, potentially through NVLink or ConnectX-8 SuperNICs. While earlier reports suggested different names like RTX Pro 6000D or B40, B30 could be one variant within the BXX family. The design incorporates GB20X silicon, which also powers consumer-grade RTX 50 GPUs, but may exclude NVLink support seen in prior generations due to its absence in consumer-grade GPU dies.

Nvidia has also introduced Fast-dLLM, a training-free framework designed to enhance the inference speed of diffusion large language models (LLMs). Diffusion models, explored as an alternative to autoregressive models, promise faster decoding through simultaneous multi-token generation, enabled by bidirectional attention mechanisms. However, their practical application is limited by inefficient inference, largely due to the lack of key-value (KV) caching, which accelerates performance by reusing previously computed attention states. Fast-dLLM aims to address this by bringing KV caching and parallel decoding capabilities to diffusion LLMs, potentially surpassing autoregressive systems.

During his keynote speech at GTC 2025, Nvidia CEO Jensen Huang emphasized the accelerating pace of artificial intelligence development and the critical need for optimized AI infrastructure. He stated Nvidia would shift to the Blackwell architecture for future China-bound chips, discontinuing Hopper-based alternatives following the H20 ban. Huang's focus on AI infrastructure highlights the industry's recognition of the importance of robust and scalable systems to support the growing demands of AI applications.

Share: bluesky twitterx--v2 facebook--v1 threads


References :
  • thenewstack.io: This article discusses Jensen Huang's keynote speech at GTC 2025, where he emphasized the acceleration of artificial intelligence development and outlined five key takeaways regarding optimizing AI infrastructure.
  • MarkTechPost: This article discusses NVIDIA's Fast-dLLM, a training-free framework that brings KV caching and parallel decoding to diffusion LLMs. It aims to improve inference speed in diffusion models, potentially surpassing autoregressive systems.
  • www.tomshardware.com: This article discusses the development of Nvidia's B30 AI chip specifically for the Chinese market. It highlights the potential inclusion of NVLink for multi-GPU scaling and the creation of high-performance clusters.
  • www.marktechpost.com: NVIDIA has introduced Llama Nemotron Nano VL, a vision-language model (VLM) designed to address document-level understanding tasks with efficiency and precision.
Classification:
@www.artificialintelligence-news.com //
Nvidia is actively working to maintain its presence in the rapidly expanding Chinese AI chip market, even amidst tightening U.S. export controls and rising domestic competition from companies like Huawei. To navigate these challenges, Nvidia is developing a cheaper, stripped-down AI chip based on its Blackwell architecture, specifically tailored for the Chinese market. This new chip is designed to comply with U.S. regulations while offering competitive performance, aiming to regain some of the market share Nvidia has been losing.

This strategic gamble involves significant cost reductions and performance trade-offs. The new processor, expected to be priced between $6,500 and $8,000, will use the RTX Pro 6000D foundation with GDDR7 memory, forgoing the high-bandwidth memory found in premium variants and Taiwan Semiconductor's advanced CoWoS packaging technology. This decision is driven by the need to meet export restrictions and lower manufacturing complexity, following a $5.5 billion writedown Nvidia absorbed after the prohibition of its H20 sales in China.

Huawei's growing influence in the Chinese AI chip landscape is a key factor driving Nvidia's strategy. Huawei's Ascend 910C and 910B processors are gaining adoption among major domestic tech firms, challenging Nvidia's dominance, particularly in inference applications. The competitive pressure extends to complete infrastructure solutions, with Huawei's CloudMatrix 384 rack system directly challenging Nvidia's Blackwell GB200 NVL72 configuration. The company hopes the new budget option will help them survive the growth of the internal competion,

Share: bluesky twitterx--v2 facebook--v1 threads


References :
  • www.artificialintelligence-news.com: Nvidia is preparing to go head-to-head with Huawei to maintain its relevance in the booming AI chip market of China.
  • The Rundown AI: PLUS: OpenAI's o3 model finds novel security vulnerability
  • insidehpc.com: Report: NVIDIA and AMD Devising Export Rules-Compliant Chips for China AI Market
  • insideAI News: One of the biggest stories this week in the HPC-AI world involves — surprise! — NVIDIA.
  • PCMag Middle East ai: Though he 'trusts' President Trump, Nvidia CEO Jensen Huang says banning sales of its H20 GPUs to China means that $50 billion market 'is effectively closed to US industry.' Facing billions in lost sales, Nvidia CEO Jensen Huang is urging the US to let it sell AI chips to China, though he stopped short of directly criticizing President Trump. The Trump administration has been trying to curb China's access to cutting-edge AI chips, so last month it restricted Nvidia …
  • www.tomshardware.com: Nvidia reportedly developing new AI chip for China that meets export controls – B30 could include NVLink for creation of high-performance clusters
Classification: