News from the AI & ML world
Jaime Hampton@AIwire
//
Cerebras Systems is significantly expanding its AI infrastructure to challenge Nvidia's dominance in the AI market. The company is deploying over a thousand of its wafer-scale AI accelerator chips across six new data centers in North America and France. This expansion aims to provide ultrafast AI inference capabilities, promising faster speeds and cost reductions compared to traditional GPU-based setups.
These new data centers will process an impressive 40 million tokens per second, with 85% of the capacity located in the United States. Facilities are already operational in Santa Clara, Stockton, and Dallas, and further expansion includes sites in Minneapolis (Q2 2025), Oklahoma City and Montreal (Q3), and Atlanta and France (Q4). Cerebras is also partnering with Hugging Face to provide developers with easy access to its AI inference service, marking a major distribution channel for open-source models like Llama 3.
ImgSrc: www.aiwire.net
References :
- The Register - Software: Plus, startup's inference service makes debut on Hugging Face Cerebras has begun deploying more than a thousand of its dinner-plate sized-accelerators across North America and parts of France as the startup looks to establish itself as one of the largest and fastest suppliers of AI inference services. 
- THE DECODER: Cerebras Systems plans to strengthen its AI inference capabilities by building new data centers across North America and Europe.
- venturebeat.com: Cerebras just announced 6 new AI datacenters that process 40M tokens per second — and it could be bad news for Nvidia
- AIwire: Cerebras Scales AI Inference with Hugging Face Partnership and Datacenter Expansion
Classification:
- HashTags: #AIinfrastructure #DataCenters #Cerebras
- Company: Cerebras Systems
- Target: AI Developers
- Product: Wafer-Scale AI Accelerators
- Feature: AI Inference
- Type: AI
- Severity: Informative