Jaime Hampton@AIwire
//
China's multi-billion-dollar AI infrastructure boom is now facing a significant downturn, according to a new report. The rush to build AI datacenters, fueled by the rise of generative AI and encouraged by government incentives, has resulted in billions of dollars in idle infrastructure. Many newly built facilities are now sitting empty, with some reports indicating that up to 80% of China’s new computing resources remain unused.
The "DeepSeek Effect" is a major factor in this reversal. DeepSeek's AI models, particularly the Deepseek v3, have demonstrated impressive efficiency in training, reducing the demand for large-scale datacenter deployments. Smaller players are abandoning plans to pretrain large models because DeepSeek’s open-source models match ChatGPT-level performance at a fraction of the cost, leading to a collapse in demand for training infrastructure just as new facilities were ready to come online. Recommended read:
References :
Jaime Hampton@AIwire
//
DeepSeek's innovative AI models are reshaping China's AI data center infrastructure, leading to a market disruption and potentially underutilized resources. The company's DeepSeek-V3 model has demonstrated performance that rivals ChatGPT but at a significantly reduced cost. This has altered the demand for extensive GPU clusters used in traditional AI training, shifting the focus towards hardware prioritizing low-latency, particularly near tech hubs. This has resulted in increased speculation as well as experienced players who are now posed with the challenge of the DeepSeek V3.
The open-source nature of DeepSeek’s model is also allowing smaller players to compete without the need for extensive pretraining, which is undermining the demand for large data centers. DeepSeek-V3, which runs at 20 tokens per second on a Mac Studio, poses a new challenge for existing AI models. Chinese AI startups are now riding DeepSeek's momentum and building an ecosystem that is revolutionizing the AI landscape. This narrows the technology divide between China and the United States. Recommended read:
References :
Dashveenjit Kaur@AI News
//
Chinese AI startup DeepSeek is shaking up the global technology landscape with its latest large language model, DeepSeek-V3-0324. This new model has been lauded for matching the performance of American AI models, while boasting significantly lower development costs. According to Lee Kai-fu, CEO of Chinese startup 01.AI, the gap between Chinese and American AI capabilities has narrowed dramatically, with China even ahead in some specific areas.
DeepSeek-V3-0324 features enhanced reasoning capabilities and improved performance in multiple benchmarks, particularly in mathematics. The model scored 59.4 on the American Invitational Mathematics Examination (AIME), a significant improvement over its predecessor. Häme University lecturer Kuittinen Petri noted DeepSeek's achievements were realized with just a fraction of the resources available to competitors like OpenAI. This breakthrough has been attributed to DeepSeek’s focus on algorithmic efficiency and novel approaches to model architecture, allowing them to overcome restrictions on access to the latest silicon. This disruption is not going unnoticed, when DeepSeek launched its R1 model in January, America’s Nasdaq plunged 3.1%, while the S&P 500 fell 1.5%. While DeepSeek claimed a $5.6 million training cost, this represented only the marginal cost of the final training run. SemiAnalysis estimates DeepSeek's actual hardware investment at closer to $1.6 billion, with hundreds of millions in operating costs. The developments present opportunities and challenges for the. Recommended read:
References :
Dashveenjit Kaur@AI News
//
DeepSeek, a Chinese AI startup, is causing a stir in the AI industry with its new large language model, DeepSeek-V3-0324. Released with little fanfare on the Hugging Face AI repository, the 641-gigabyte model is freely available for commercial use under an MIT license. Early reports indicate it can run directly on consumer-grade hardware, such as Apple’s Mac Studio with the M3 Ultra chip, especially in a 4-bit quantized version that reduces the storage footprint to 352GB. This innovation challenges the previous notion that Silicon Valley held a chokehold on the AI industry.
China's focus on algorithmic efficiency over hardware superiority has allowed companies like DeepSeek to flourish despite restrictions on access to the latest silicon. DeepSeek's R1 model, launched earlier this year, already rivaled OpenAI's ChatGPT-4 at a fraction of the cost. Now the DeepSeek-V3-0324 features enhanced reasoning capabilities and improved performance. This has sparked a gold rush among Chinese tech startups, rewriting the playbook for AI development and allowing smaller companies to believe they have a shot in the market. Recommended read:
References :
Ryan Daws@AI News
//
DeepSeek, a Chinese AI company, has released DeepSeek V3-0324, an updated AI model that demonstrates impressive performance. The model is now running at 20 tokens per second on a Mac Studio. This model is said to contain 685 billion parameters and its cost-effectiveness challenges the dominance of American AI models, signaling that China continues to innovate in AI despite chip restrictions. Reports from early testers show improvements over previous versions and the model tops non-reasoning AI models in open-source first.
This new model runs on consumer-grade hardware, specifically Apple's Mac Studio with the M3 Ultra chip, diverging from the typical data center requirements for AI. It is freely available for commercial use under the MIT license. According to AI researcher Awni Hannun, the model runs at over 20 tokens per second on a 512GB M3 Ultra. The company has made no formal announcement, just an empty README file and the model weights themselves. This stands in contrast to the carefully orchestrated product launches by Western AI companies. Recommended read:
References :
Madeline Clarke@techrepublic.com
//
References:
Jon Keegan
, www.techrepublic.com
,
Ant Group has announced a significant breakthrough in artificial intelligence, claiming to have slashed AI costs by 20% using Chinese-made chips. The company's Ling-Plus and Ling-Lite models reportedly match or outperform leading AI models, demonstrating China's increasing ability to innovate around US export controls. This achievement marks a potential leap forward in China’s AI development efforts, signaling a move towards self-reliance in AI development and reduced dependence on foreign technologies.
Ant Group leveraged chips from Chinese tech giants Alibaba and Huawei to train its AI model, reaching performance levels comparable to those obtained with Nvidia’s H800 chips. While Ant Group continues to utilize Nvidia’s hardware for certain AI development tasks, the company is now relying increasingly on alternatives — particularly chips from AMD and Chinese manufacturers — for its latest models. This strategic pivot reflects a broader trend within China’s tech industry, driven in part by tightening U.S. sanctions that limit access to Nvidia’s most advanced GPUs. Recommended read:
References :
Ryan Daws@AI News
//
DeepSeek V3-0324 has emerged as a leading AI model, topping benchmarks for non-reasoning AI in an open-source breakthrough. This milestone signifies a significant advancement in the field, as it marks the first time an open weights model has achieved the top position among non-reasoning models. The model's performance surpasses proprietary counterparts and edges it closer to proprietary reasoning models, highlighting the growing viability of open-source solutions for latency-sensitive applications. DeepSeek V3-0324 represents a new era for open-source AI, offering a powerful and adaptable tool for developers and enterprises.
DeepSeek-V3 now runs at 20 tokens per second on Apple’s Mac Studio, presenting a challenge to OpenAI’s cloud-dependent business model. The 685-billion-parameter model, DeepSeek-V3-0324, is freely available for commercial use under the MIT license. This achievement, coupled with its cost efficiency and performance, signals a shift in the AI sector, where open-source frameworks increasingly compete with closed systems. Early testers report significant improvements over previous versions, positioning DeepSeek's new model above Claude Sonnet 3.5 from Anthropic. Recommended read:
References :
Matthias Bastian@THE DECODER
//
Baidu has launched two new AI models, ERNIE 4.5 and ERNIE X1, designed to compete with DeepSeek's R1 model. The company is making these models freely accessible to individual users through the ERNIE Bot platform, ahead of the initially planned schedule. ERNIE 4.5 is a multimodal foundation model, integrating text, images, audio, and video to enhance understanding and content generation across various data types. This model demonstrates significant improvements in language understanding, reasoning, and coding abilities.
ERNIE X1 is Baidu's first model specifically designed for complex reasoning tasks, excelling in logical inference, problem-solving, and structured decision-making suitable for applications in finance, law, and data analysis. Baidu claims that ERNIE X1 matches DeepSeek R1’s performance at half the cost. ERNIE 4.5 has shown performance on par with models like DeepSeek-R1, but at approximately half the deployment cost. Recommended read:
References :
Matthias Bastian@THE DECODER
//
Baidu has released two new large language models, ERNIE 4.5 and ERNIE X1, claiming they outperform OpenAI's GPT-4.5 and DeepSeek-R1. These models are more cost-effective, offering high quality at a fraction of the price. ERNIE 4.5 is a multimodal foundation model that integrates text, images, audio, and video, enhancing its ability to understand and generate different kinds of content. ERNIE X1 is a deep-thinking reasoning model with multimodal capabilities, excelling in tasks requiring advanced reasoning.
Baidu has made both models freely accessible to individual users via the ERNIE Bot platform, ahead of schedule. For enterprise users and developers, ERNIE 4.5 is available via APIs on Baidu AI Cloud's Qianfan platform, with ERNIE X1 set to follow. Baidu also plans to integrate the models into its existing products, including Baidu Search and the Wenxiaoyan app. This move positions Baidu as a competitive force in the AI landscape, challenging Western AI companies. Recommended read:
References :
Nitika Sharma@Analytics Vidhya
//
China's Manus AI, developed by Monica, is generating buzz as an invite-only multi-agent AI product. This AI agent is designed to autonomously tackle complex, real-world tasks by operating as a multi-agent system. It utilizes a planner optimized for strategic reasoning, and an executor driven by Claude 3.5 Sonnet, incorporating code execution, web browsing, and multi-file code management.
The AI agent has sparked considerable global attention, igniting discussions about its technological and ethical implications, as well as its potential impact on the AI landscape. Manus reportedly outperformed OpenAI's o3-powered Deep Research agent on benchmarks, as showcased on the Manus website, leading some to believe it is among the most effective autonomous agents currently available. However, there is some skepticism due to it appearing to be a Claude wrapper with a jailbreak and tools optimized for the GAIA benchmark. Recommended read:
References :
Matthias Bastian@THE DECODER
//
Chinese AI company DeepSeek is making waves in the global AI market with its high profit margins and low pricing. The company makes $200 million per year at 85% or greater profit margins, even while charging $2.19 per million tokens on its R1 model, about 25 times less than OpenAI. DeepSeek's financial data suggests a theoretical peak revenue could exceed operating costs by six times when using optimal R1 model pricing.
The company's success has prompted Tencent to unveil its own AI platform, Hunyuan Turbo S, designed specifically to compete with DeepSeek. Although Hunyuan Turbo S is the clear winner in certain cases, it still falls behind DeepSeek-R1-Zero in several instances. DeepSeek uses smart resource management and a dynamic resource allocation system which keeps costs down. Recommended read:
References :
|
BenchmarksBlogsResearch Tools |