News from the AI & ML world

DeeperML - #aiinfrastructure

Allyson Vasquez@NVIDIA Technical Blog //
References: Data Phoenix , AIwire , TechCrunch ...
NVIDIA's GTC 2025 is shaping up to be a major event for AI enthusiasts, packed with networking opportunities, live demos, and discussions on the latest AI innovations. Data Phoenix is highlighting the event as a key gathering, featuring meetups, networking receptions, and hands-on sessions alongside the main conference. They are also co-hosting and supporting key events like the INFRA@GTC Networking Reception and AI Demo Jam.

VAST Data plans to showcase its data platform for enterprise Retrieval Augmented Generation (RAG) use cases at the conference. Microsoft and NVIDIA have also announced a partnership to integrate RTX Neural Shaders into a DirectX preview in April, bringing more AI capabilities to game development. This integration will allow developers to leverage Tensor cores in RTX GPUs to accelerate neural networks within a game's graphics pipeline.

Recommended read:
References :
  • Data Phoenix: DATAPHOENIX: NVIDIA GTC 2025 & Top AI Events of the Week! (March 17-23)
  • AIwire: VAST Fleshes Out Data Platform for Enterprise RAG Use Cases
  • BigDATAwire: NVIDIA GTC 2025: What to Expect From the Ultimate AI Event?
  • TechCrunch: Nvidia GTC 2025: What to expect from this year’s show
  • AIwire: Provides an overview of what to expect from NVIDIA's GTC 2025 event.
  • www.tomsguide.com: Nvidia CEO Jensen Huang is taking center stage in the Nvidia GTC 2025 Keynote to show off what's next in AI — here's the latest news.
  • insidehpc.com: SAN JOSE, Calif. – March 18, 2025– Data infrastructure company NetApp (NASDAQ: NTAP) today made an agentic AI announcement that taps the NVIDIA AI Data Platform reference design.Â
  • AIwire: At its GTC event in San Jose today, Nvidia unveiled updates to its AI infrastructure portfolio, including its next-generation datacenter GPU, the NVIDIA Blackwell Ultra. Expanding on the Blackwell architecture introduced last year, Nvidia is integrating its new 300-series GPUs into two DGX systems: the NVIDIA DGX GB300 and the NVIDIA DGX B300.
  • www.tomshardware.com: Nvidia announces Blackwell Ultra B300 —1.5X faster than B200 with 288GB HBM3e and 15 PFLOPS dense FP4
  • www.laptopmag.com: Nvidia's Jensen Huang says new Blackwell chips make previous-gen feel obsolete
  • NVIDIA Newsroom: Full Steam Ahead: NVIDIA-Certified Program Expands to Enterprise Storage for Faster AI Factory Deployment
  • BigDATAwire: Nvidia unveils updates to its AI infrastructure portfolio, including its next-generation datacenter GPU, the NVIDIA Blackwell Ultra.
  • venturebeat.com: Nvidia’s GTC 2025 keynote: 40x AI performance leap, open-source ‘Dynamo’, and a walking Star Wars-inspired ‘Blue’ robot
  • BigDATAwire: Nvidia used its GTC conference today to introduce new GPU superchips, including the second generation of its current Grace Blackwell chip, as well as the next generation, dubbed the Vera The post appeared first on .
  • IBM - Announcements: IBM Taps NVIDIA AI Data Platform Technologies to Accelerate AI at Scale
  • Analytics Vidhya: Nvidia CEO Jensen Huang's keynote at GTC 2025 is detailed, emphasizing the advancements in AI and accelerated computing, including the introduction of the Blackwell Ultra GB300 chips and the upcoming Blackwell B300 series.
  • AIwire: Nvidia CEO Jensen Huang's keynote highlighted advancements in AI and predictions for industry evolution in the coming years. The keynote showcased Nvidia's next-generation graphics architectures: Blackwell Ultra and Vera Rubin, designed for managing sophisticated AI processes.
  • Analytics Vidhya: Nvidia’s GTC 2025 Announcements That Shook the Stock Market
  • AIwire: Nvidia touts next-generation GPU superchip and new photonic switches, unveiling Blackwell Ultra and Vera Rubin architectures.
  • Gradient Flow: Nvidia’s AI Vision: GTC 2025 and the Road Ahead
  • OODAloop: Nvidia is releasing what it’s calling an AI foundation model for humanoid robotics.
  • Data Phoenix: Nvidia introduces the Blackwell Ultra to support the rise of AI reasoning, agents, and physical AI
  • insideAI News: In what is becoming an annual tradition for the @HPCpodcast, we present “Live from Nvidia GTC 2025,â€� covering highlights from the Nvidia extravaganza with an AI-everywhere theme.
  • BigDATAwire: Reports from Nvidia’s GPU Technology Conference (GTC) 2025, a weeklong event in San Jose, California that be remembered for a long time, if not for the content

Jaime Hampton@BigDATAwire //
NVIDIA's GTC 2025 showcased significant advancements in AI, marked by the unveiling of the Blackwell Ultra GPU and the Vera Rubin roadmap extending through 2027. CEO Jensen Huang emphasized a 40x AI performance leap with the Blackwell platform compared to its predecessor, Hopper, highlighting its crucial role in inference workloads. The conference also introduced open-source ‘Dynamo’ software and advancements in humanoid robotics, demonstrating NVIDIA’s commitment to pushing AI boundaries.

The Blackwell platform is now in full production, meeting incredible customer demand, and the Vera Rubin roadmap details the next generation of superchips expected in 2026. Huang also touted new DGX systems, highlighting the push towards photonic switches to handle growing data demands efficiently. Blackwell Ultra will offer 288GB of memory. NVIDIA claims the GB300 chip brings 1.5x more AI performance than the NVIDIA GB200. These advancements aim to bolster AI reasoning capabilities and energy efficiency, positioning NVIDIA to maintain its dominance in AI infrastructure.

Recommended read:
References :
  • AI News | VentureBeat: Nvidia launches Blackwell RTX Pro for workstations and servers
  • AIwire: Nvidia’s DGX AI Systems Are Faster and Smarter Than Ever
  • Analytics Vidhya: 10 NVIDIA GTC 2025 Announements that You Must Know
  • venturebeat.com: Nvidia’s GTC 2025 keynote: 40x AI performance leap, open-source ‘Dynamo’, and a walking Star Wars-inspired ‘Blue’ robot
  • BigDATAwire: Nvidia Touts Next Generation GPU Superchip and New Photonic Switches
  • www.laptopmag.com: "I'm the chief revenue destroyer": Nvidia's Jensen Huang says new Blackwell chips make previous-gen feel obsolete
  • AIwire: The GTC 2025 happening in San Jose, Calif., has become one of the marquee events in the tech world. It has grabbed the attention of everyone from industry leaders and developers to AI enthusiasts and even those who remain skeptical about AI’s potential.
  • BigDATAwire: Nvidia unveils updates to its AI infrastructure portfolio, including its next-generation datacenter GPU, the NVIDIA Blackwell Ultra.
  • BigDATAwire: Nvidia Preps for 100x Surge in Inference Workloads, Thanks to Reasoning AI Agents
  • Gradient Flow: Overview As I sat watching Jensen Huang’s keynote at Nvidia’s recent GTC, I was struck once again by how this annual event has evolved from a graphics card showcase into something far more consequential for global markets.
  • Analytics Vidhya: Nvidia’s annual GPU Technology Conference (GTC) has long been a highlight for the AI community. At this year’s event, Nvidia CEO Jensen Huang unveiled a roadmap of new products and innovations aimed at scaling up artificial intelligence.
  • NVIDIA Newsroom: AI Factories Are Redefining Data Centers and Enabling the Next Era of AI
  • AIwire: The GTC 2025 happening in San Jose, Calif., has become one of the marquee events in the tech world.
  • BigDATAwire: The boundaries between artificial intelligence and the physical world are dissolving. AI systems are becoming increasingly adept at perceiving, interacting, analyzing, and responding to their physical environments. In this AI The post appeared first on .
  • Data Phoenix: NVIDIA's new Blackwell Ultra platform delivers significantly enhanced AI computing power for reasoning and agentic AI applications.
  • Maginative: How NVIDIA Is Building the Operating System for Physical AI
  • insideAI News: Highlights from the Nvidia extravaganza with an AI-everywhere theme. We review the conference, discussing everything from the new AI compute industry landscape
  • The Next Platform: Reports on how Nvidia is turning its AI eye to the enterprise.
  • Last Week in AI: Nvidia's GTC 2025 keynote focused on the transition to AI-driven computing and the development of AI factories.
  • BigDATAwire: Reporter’s Notebook: AI Hype and Glory at Nvidia GTC 2025

@tomshardware.com //
Nvidia has unveiled its next-generation data center GPU, the Blackwell Ultra, at its GTC event in San Jose. Expanding on the Blackwell architecture, the Blackwell Ultra GPU will be integrated into the DGX GB300 and DGX B300 systems. The DGX GB300 system, designed with a rack-scale, liquid-cooled architecture, is powered by the Grace Blackwell Ultra Superchip, combining 36 NVIDIA Grace CPUs and 72 NVIDIA Blackwell Ultra GPUs. Nvidia officially revealed its Blackwell Ultra B300 data center GPU, which packs up to 288GB of HBM3e memory and offers 1.5X the compute potential of the existing B200 solution.

The Blackwell Ultra GPU promises a 70x speedup in AI inference and reasoning compared to the previous Hopper-based generation. This improvement is achieved through hardware and networking advancements in the DGX GB300 system. Blackwell Ultra is designed to meet the demand for test-time scaling inference with a 1.5X increase in the FP4 compute. Nvidia's CEO, Jensen Huang, suggests that the new Blackwell chips render the previous generation obsolete, emphasizing the significant leap forward in AI infrastructure.

Recommended read:
References :
  • AIwire: Nvidia’s DGX AI Systems Are Faster and Smarter Than Ever
  • www.tomshardware.com: Nvidia officially revealed its Blackwell Ultra B300 data center GPU, which packs up to 288GB of HBM3e memory and offers 1.5X the compute potential of the existing B200 solution.
  • BigDATAwire: Nvidia's GTC 2025 conference showcased the new Blackwell Ultra GPUs and updates to its AI infrastructure portfolio.
  • www.laptopmag.com: Blackwell Ultra and Rubin Ultra are Nvidia's newest additions to the growing list of AI superchips
  • BigDATAwire: Nvidia used its GTC conference today to introduce new GPU superchips, including the second generation of its current Grace Blackwell chip, as well as the next generation, dubbed the Vera The post appeared first on .
  • venturebeat.com: Nvidia's GTC 2025 keynote highlighted advancements in AI infrastructure, featuring the Blackwell Ultra GB300 chips.
  • Analytics Vidhya: An overview of Nvidia's GTC 2025 announcements, including new GPUs and advancements in AI hardware.
  • AI News: NVIDIA Dynamo: Scaling AI inference with open-source efficiency
  • www.tomshardware.com: Nvidia unveils DGX Station workstation PCs with GB300 Blackwell Ultra inside
  • BigDATAwire: Nvidia Preps for 100x Surge in Inference Workloads, Thanks to Reasoning AI Agents
  • Data Phoenix: Nvidia introduces the Blackwell Ultra to support the rise of AI reasoning, agents, and physical AI
  • The Next Platform: This article discusses Nvidia's new advancements in AI, and how the company is looking to capture market share and the challenges they face.

Maximilian Schreiner@THE DECODER //
OpenAI has announced it will adopt Anthropic's Model Context Protocol (MCP) across its product line. This surprising move involves integrating MCP support into the Agents SDK immediately, followed by the ChatGPT desktop app and Responses API. MCP is an open standard introduced last November by Anthropic, designed to enable developers to build secure, two-way connections between their data sources and AI-powered tools. This collaboration between rivals marks a significant shift in the AI landscape, as competitors typically develop proprietary systems.

MCP aims to standardize how AI assistants access, query, and interact with business tools and repositories in real-time, overcoming the limitation of AI being isolated from systems where work happens. It allows AI models like ChatGPT to connect directly to the systems where data lives, eliminating the need for custom integrations for each data source. Other companies, including Block, Apollo, Replit, Codeium, and Sourcegraph, have already added MCP support, and Anthropic's Chief Product Officer Mike Krieger welcomes OpenAI's adoption, highlighting MCP as a thriving open standard with growing integrations.

Recommended read:
References :
  • AI News | VentureBeat: The open source Model Context Protocol was just updated — here’s why it’s a big deal
  • Runtime: Why AI infrastructure companies are lining up behind Anthropic's MCP
  • : OpenAI and Anthropic Play Nice – It’s A Big Deal For Agents
  • THE DECODER: OpenAI adopts competitor Anthropic's standard for AI data access
  • Simon Willison's Weblog: OpenAI Agents SDK You can now connect your Model Context Protocol servers to Agents: We’re also working on MCP support for the OpenAI API and ChatGPT desktop app—we’ll share some more news in the coming months. — Tags: , , , , , ,
  • Analytics Vidhya: To improve AI interoperability, OpenAI has announced its support for Anthropic’s Model Context Protocol (MCP), an open-source standard designed to streamline the integration between AI assistants and various data systems.
  • THE DECODER: Anthropic and Databricks close 100 million dollar deal for AI agents
  • Analytics India Magazine: Databricks and Anthropic Partner to Bring AI Models to Businesses
  • www.itpro.com: Databricks and Anthropic are teaming up on agentic AI development – here’s what it means for customers
  • Runtime: Model Context Protocol (MCP) was introduced last November by Anthropic, which called it "an open standard that enables developers to build secure, two-way connections between their data sources and AI-powered tools."
  • The Tech Basic: OpenAI has formed a partnership with its competitor, Anthropic, to implement the Model Context Protocol (MCP) tool.
  • www.techrepublic.com: OpenAI Agents Now Support Rival Anthropic’s Protocol, Making Data Access ‘Simpler, More Reliable’

Chris McKay@Maginative //
NVIDIA's GTC 2025 event showcased significant advancements in AI infrastructure, highlighting the Blackwell Ultra and Rubin architectures, along with several related technologies and partnerships. Jensen Huang, Nvidia CEO, delivered a keynote address outlining the company’s vision for the AI-powered future, emphasizing improvements in processor performance, network design, and memory capabilities. The Blackwell Ultra GPUs are being integrated into DGX systems to meet the rising demands of AI workloads, especially in inference and reasoning.

NVIDIA is also expanding its offerings beyond chips with the introduction of desktop AI supercomputers for developers. The DGX Station, powered by the GB300 Blackwell Ultra Superchip, aims to bring data center-level AI capabilities to a compact form factor. Nvidia introduced Dynamo, an open-source inference software engineered to maximize token revenue generation for AI factories deploying reasoning AI models. The presentation emphasized a clear roadmap for data center computing, advancements in AI reasoning capabilities, and bold moves into robotics and autonomous vehicles.

Recommended read:
References :
  • Analytics Vidhya: 10 NVIDIA GTC 2025 Announements that You Must Know
  • BigDATAwire: Nvidia Touts Next Generation GPU Superchip and New Photonic Switches
  • www.tomshardware.com: Nvidia unveils DGX Station workstation PCs with GB300 Blackwell Ultra inside
  • Gradient Flow: Nvidia’s AI Vision: GTC 2025 and the Road Ahead
  • BigDATAwire: Nvidia Cranks Up the DGX Performance with Blackwell Ultra
  • Data Phoenix: NVIDIA's new Blackwell Ultra platform delivers significantly enhanced AI computing power for reasoning and agentic AI applications. Introduced yesterday at GTC, Blackwell Ultra is expected to be adopted by Nvidia's cloud and manufacturing partners.
  • BigDATAwire: Nvidia Preps for 100x Surge in Inference Workloads, Thanks to Reasoning AI Agents
  • Analytics Vidhya: Nvidia’s GTC 2025 Announcements That Shook the Stock Market
  • eWEEK: NVIDIA Shows More AI Infrastructure at GTC 2025: ‘Every Single Layer of Computing Has Been Transformed’
  • insideAI News: @HPCpodcast: Live from GTC 2025, Among the Crowds for the New AI Compute Landscape
  • AIwire: Nvidia Touts Next Generation GPU Superchip and New Photonic Switches

Matt Milano@WebProNews //
References: THE DECODER , Maginative , techstrong.ai ...
OpenAI has inked a significant five-year, $11.9 billion agreement with CoreWeave, a cloud infrastructure provider specializing in AI workloads. This substantial investment aims to secure the necessary GPU compute capacity for OpenAI's increasingly demanding AI models. The deal also includes OpenAI acquiring a $350 million stake in CoreWeave, demonstrating a deeper strategic partnership between the two companies as CoreWeave prepares for its IPO.

This move signifies OpenAI's ongoing quest for ever greater AI compute capabilities and a diversification of its cloud infrastructure strategy beyond Microsoft Azure. CoreWeave, backed by NVIDIA, operates 32 AI data centers housing over 250,000 GPUs and is positioning itself as a key player in the AI infrastructure space. The partnership also strengthens CoreWeave’s market position as it attempts to reduce its reliance on Microsoft, which accounted for 62% of its $1.92 billion revenue in 2024.

Recommended read:
References :
  • THE DECODER: OpenAI's $12 billion Coreweave deal fuels its unrelenting quest for more AI compute
  • Maginative: OpenAI has signed a five-year, $11.9 billion deal with CoreWeave to secure AI infrastructure, marking a major shift in its compute strategy and adding complexity to its partnership with Microsoft.
  • WebProNews: CoreWeave Reportedly Scores $12 Billion OpenAI Cloud Contract
  • techstrong.ai: CoreWeave has signed a five-year, $11.9 billion deal with OpenAI to provide artificial intelligence (AI) infrastructure services ahead of its much-anticipated initial public offering.
  • AiThority: Details the agreement between CoreWeave and OpenAI to deliver AI infrastructure.
  • eWEEK: The AI cloud provider has secured an $11.9 billion deal with OpenAI ahead of its IPO.
  • Verdict: CoreWeave signs $11.9bn deal with OpenAI
  • The Next Platform: Wouldn’t it be funny if all of that money that Microsoft spent last year paying neocloud upstart CoreWeave was just to support ever-embiggening AI training workloads at OpenAI as it makes its GPT models smarter?
  • Tech Monitor: CoreWeave wins contract from OpenAI, aimed at delivering crucial AI infrastructure to boost ChatGPT developer’s computing capabilities.
  • insidehpc.com: Coreweave is reportedly providing AI infrastructure services to OpenAI.
  • Data Phoenix: This is an article about OpenAI investing billions of dollars into CoreWeave, a cloud computing company, to provide more AI infrastructure.

staff@insideAI News //
References: insideAI News , insidehpc.com , Verdict ...
OpenAI and Oracle are collaborating on a massive AI data center project in Texas, known as 'Stargate', which will be equipped with Nvidia's GB200 Blackwell chips. The data center, located in Abilene, Texas, aims to become a transformative benchmark for next-generation AI innovation.

This ambitious project involves the purchase of 64,000 Nvidia GB200 chips, to be installed in phases with an initial rollout of 16,000 chips planned for completion within the next six months. The data center is expected to house all 64,000 of the chips by the end of 2026.

The GB200 Superchip, which combines a Grace CPU with two enhanced B200 GPUs and priced between $60,000-$70,000 per chip, will provide the necessary computing power for the facility. An OpenAI spokesperson confirmed their collaboration with Oracle on the design and delivery of the data center, noting that Oracle will oversee the acquisition and operation of the supercomputer being constructed. This venture is part of a larger $100 billion Stargate infrastructure initiative.

Recommended read:
References :
  • insideAI News: Report: 64,000 Nvidia GB200s for Stargate AI Data Center in Texas
  • insidehpc.com: Report: 64,000 Nvidia GB200s for Stargate AI Data Center in Texas
  • eWEEK: OpenAI, Oracle’s ‘Stargate’ Takes a Billion-Dollar Leap with Nvidia’s AI Chips
  • Verdict: OpenAI, Oracle to equip Texas data centre with Nvidia AI chips
  • OODAloop: Stargate to Require 64,000 Nvidia GPUs by 2026
  • John Werner: Details on OpenAI and Oracle Corp planning to equip Texas data center with Nvidia AI chips.

staff@insideAI News //
References: insideAI News , insidehpc.com , OODAloop ...
Celestial AI, the creator of the Photonic Fabric optical interconnect technology, has announced a successful Series C1 funding round, securing $250 million. The round was led by Fidelity Management & Research Company, bringing the total capital raised by the company to over $515 million. Celestial AI's Photonic Fabric platform aims to speed up data transfer inside servers using light, with plans for new data centers in North America and France. This investment underscores the growing demand for advanced AI models and the development of AI infrastructure.

The company's Photonic Fabric technology allows AI compute to be networked seamlessly, from within processor packages to servers across multiple racks. Celestial AI offers a full suite of products, that include connectivity, switching and packaging solutions which serve as the foundation for optical scale-up networks for accelerated computing. Celestial AI's valuation has now reached $2.5 billion, reflecting investor confidence in its potential to revolutionize AI infrastructure. New investors include funds and accounts managed by BlackRock, Maverick Silicon, Tiger Global Management and Lip-Bu Tan, as well as participation from existing investors including AMD Ventures, Koch Disruptive Technologies (KDT), Temasek, Temasek’s wholly-owned subsidiary Xora Innovation, Porsche Automobil Holding SE and The Engine Ventures.

Recommended read:
References :
  • insideAI News: SANTA CLARA, CA – March 11, 2025 – Celestial AI, creator of the Photonic Fabric optical interconnect, today announced that it has raised $250 million in its Series C1 funding round led by Fidelity Management & Research Company, bringing the total capital raised to date to more than $515 million.
  • insidehpc.com: SANTA CLARA, CA – March 11, 2025 – Celestial AI, creator of the Photonic Fabric optical interconnect, today announced that it has raised $250 million in its Series C1 funding round led by Fidelity Management & Research Company, bringing the total capital raised to date to more than $515 million.
  • Crunchbase News: Optical interconnectivity startup Celestial AI raised a $250 million Series C1 round led by Fidelity Management & Research Co. at a reported $2.5 billion valuation.
  • OODAloop: Celestial AI, a startup that seeks to use light to speed up data transfer inside servers, has raised $250 million to better position itself in a race to lower AI computing power requirements.
  • SiliconANGLE: Celestial AI Inc., a startup that develops optical technology for linking chips, has raised $250 million in funding at a $2.5 billion valuation.
  • Verdict: Celestial AI raises $250m in Series C1 funding

Harsh Mishra@Analytics Vidhya //
DeepSeek AI has been making significant contributions to the open-source community, particularly in the realm of AI model efficiency and accessibility. They recently launched the Fire-Flyer File System (3FS), a high-performance distributed file system tailored for AI training and inference workloads. This system is designed to address the challenges of managing large-scale, concurrent data access, a common bottleneck in traditional file systems. 3FS leverages modern SSDs and RDMA networks, offering a shared storage layer that facilitates the development of distributed applications by bypassing limitations seen in more traditional, locality-dependent file systems.

DeepSeek's commitment extends to data processing and model optimization. They have introduced the Smallpond framework for data processing and released quantized DeepSeek-R1 models, optimized for deployment-ready reasoning tasks. The quantized models, including Llama-8B, Llama-70B, Qwen-1.5B, Qwen-7B, Qwen-14B, and Qwen-32B, are available as a Hugging Face collection with evaluations, benchmarks, and setup instructions. These models maintain competitive reasoning accuracy while unlocking significant inference speedups.

Recommended read:
References :
  • Analytics Vidhya: DeepSeek #OpenSourceWeek Day 5: Launch of 3FS and Smallpond Framework
  • MarkTechPost: DeepSeek AI Releases Fire-Flyer File System (3FS): A High-Performance Distributed File System Designed to Address the Challenges of AI Training and Inference Workload
  • Neural Magic: Quantized DeepSeek-R1 Models: Deployment-Ready Reasoning Models
  • MarkTechPost: DeepSeek AI Releases Smallpond: A Lightweight Data Processing Framework Built on DuckDB and 3FS
  • www.itpro.com: ‘Awesome for the community’: DeepSeek open sourced its code repositories, and experts think it could give competitors a scare

Jaime Hampton@AIwire //
Cerebras Systems is significantly expanding its AI infrastructure to challenge Nvidia's dominance in the AI market. The company is deploying over a thousand of its wafer-scale AI accelerator chips across six new data centers in North America and France. This expansion aims to provide ultrafast AI inference capabilities, promising faster speeds and cost reductions compared to traditional GPU-based setups.

These new data centers will process an impressive 40 million tokens per second, with 85% of the capacity located in the United States. Facilities are already operational in Santa Clara, Stockton, and Dallas, and further expansion includes sites in Minneapolis (Q2 2025), Oklahoma City and Montreal (Q3), and Atlanta and France (Q4). Cerebras is also partnering with Hugging Face to provide developers with easy access to its AI inference service, marking a major distribution channel for open-source models like Llama 3.

Recommended read:
References :
  • The Register - Software: Plus, startup's inference service makes debut on Hugging Face Cerebras has begun deploying more than a thousand of its dinner-plate sized-accelerators across North America and parts of France as the startup looks to establish itself as one of the largest and fastest suppliers of AI inference services. 
  • THE DECODER: Cerebras Systems plans to strengthen its AI inference capabilities by building new data centers across North America and Europe.
  • venturebeat.com: Cerebras just announced 6 new AI datacenters that process 40M tokens per second — and it could be bad news for Nvidia
  • AIwire: Cerebras Scales AI Inference with Hugging Face Partnership and Datacenter Expansion

Ellie Ramirez-Camara@Data Phoenix //
References: Data Phoenix , BigDATAwire , BigDATAwire ...
Nvidia is making significant strides in the realm of AI agents, highlighted at this year's GTC 2025 conference. CEO Jensen Huang emphasized the transformative impact of agentic AI and reasoning models, predicting that these technologies will revolutionize industries and automate processes. To support this shift, Nvidia unveiled the Blackwell Ultra platform, designed to handle the demanding requirements of AI reasoning, agentic AI, and physical AI applications. The platform, which includes the GB300 NVL72 rack-scale solution and the HGX B300 NVL16 system, offers substantial performance improvements over previous generations, with the GB300 NVL72 delivering 1.5x more AI performance.

In addition to hardware advancements, Nvidia launched NVIDIA Dynamo, an open-source inference framework, to optimize reasoning AI services across thousands of GPUs. This framework is designed to maximize token revenue generation for AI factories deploying reasoning AI models by orchestrating and accelerating inference communication across GPU clusters. Major cloud providers and server manufacturers are expected to offer Blackwell Ultra-based products starting in the second half of 2025. These developments position Nvidia as a key player in the emerging landscape of AI agents and reasoning models, promising to drive significant advancements in AI capabilities and applications.

Recommended read:
References :
  • Data Phoenix: Nvidia introduces the Blackwell Ultra to support the rise of AI reasoning, agents, and physical AI
  • BigDATAwire: Nvidia Preps for 100x Surge in Inference Workloads, Thanks to Reasoning AI Agents
  • AIwire: Jensen Huang Charts Nvidia’s AI-Powered Future
  • BigDATAwire: The Rise of Intelligent Machines: Nvidia Accelerates Physical AI Progress

@openai.com //
OpenAI has recently partnered with the US National Laboratories to provide its AI models for applications in national security and scientific research. This move aims to leverage the power of artificial intelligence in critical areas, enhancing capabilities in both research and security sectors. The collaboration underscores the growing recognition of AI's potential to address complex challenges and drive innovation across various domains.

France is making significant investments to strengthen its position as an AI hub. President Emmanuel Macron announced that foreign and local companies will invest €109 billion in AI projects within the country. This financial commitment includes €20 billion from Brookfield, with additional financing from the UAE potentially reaching €50 billion. California State University just made a massive move in higher ed that might set the tone for how colleges nationwide adopt AI. The 23-campus system, serving more than 460,000 students and 63,000 staff and faculty, is rolling out a specialized version of ChatGPT—called ChatGPT Edu—to all of them.

Recommended read:
References :

@www.cnbc.com //
Meta is significantly increasing its investment in artificial intelligence, with CEO Mark Zuckerberg pledging "hundreds of billions" of dollars in long-term spending. This strategic move comes as Meta reports a strong fourth quarter, boasting a 21% year-over-year revenue increase to $48.4 billion and a 49% jump in net income to $20.8 billion. Zuckerberg views this massive investment in AI infrastructure as a crucial "strategic advantage" for Meta's future, enabling them to compete effectively and serve their billions of users. This move is in part a response to the emergence of new competitors like DeepSeek.

Meta's Reality Labs, while still operating at a loss of $4.97 billion in Q4, has shown positive signs with revenue up 1% year-over-year to $1.1 billion. Furthermore, internal memos reveal that Reality Labs surpassed nearly all sales and user targets for 2024, experiencing a 40% overall sales growth. Meta is particularly focused on developing open-source AI models, aiming to make Llama 4 the most competitive in the world. This open-source strategy is seen as a way to allow Meta to innovate and compete with established AI leaders, despite recent market anxieties regarding DeepSeek.

Recommended read:
References :
  • africa.businessinsider.com: In an internal memo, Andrew Bosworth says Meta's Reality Labs beat nearly all its sales and user targets for 2024, growing sales ">40% YoY overall" (Jyoti Mann/Business Insider)
  • techcrunch.com: Mark Zuckerberg says spending heavily on AI infrastructure is a "strategic advantage" and vows Meta will invest "hundreds of billions" in AI over the long term (Charles Rollet/TechCrunch)
  • www.cnbc.com: Meta's Reality Labs posts Q4 revenue up 1% YoY to $1.1B, vs. $1.1B est., and a $4.97B operating loss, vs. $5.4B est.; Reality Labs has lost $60B+ since 2020 (Jonathan Vanian/CNBC)