News from the AI & ML world
Ryan Daws@AI News
//
Alibaba's Qwen team has launched QwQ-32B, a 32-billion parameter AI model, designed to rival the performance of much larger models like DeepSeek-R1, which has 671 billion parameters. This new model highlights the effectiveness of scaling Reinforcement Learning (RL) on robust foundation models. QwQ-32B leverages continuous RL scaling to demonstrate significant improvements in areas like mathematical reasoning and coding proficiency.
The Qwen team successfully integrated agent capabilities into the reasoning model, allowing it to think critically, use tools, and adapt its reasoning based on environmental feedback. The model has been evaluated across a range of benchmarks, including AIME24, LiveCodeBench, LiveBench, IFEval, and BFCL, designed to assess its mathematical reasoning, coding proficiency, and general problem-solving capabilities. QwQ-32B is available as open-weight on Hugging Face and on ModelScope under an Apache 2.0 license, allowing for both commercial and research uses.
ImgSrc: www.artificiali
References :
- AI News | VentureBeat: Alibaba's new open source model QwQ-32B matches DeepSeek-R1 with way smaller compute requirements
- Analytics Vidhya: In the world of large language models (LLMs) there is an assumption that larger models inherently perform better. Qwen has recently introduced its latest model, QwQ-32B, positioning it as a direct competitor to the massive DeepSeek-R1 despite having significantly fewer parameters.
- AI News: The Qwen team at Alibaba has unveiled QwQ-32B, a 32 billion parameter AI model that demonstrates performance rivalling the much larger DeepSeek-R1. This breakthrough highlights the potential of scaling Reinforcement Learning (RL) on robust foundation models.
- www.infoworld.com: Alibaba Cloud on Thursday launched QwQ-32B, a compact reasoning model built on its latest large language model (LLM), Qwen2.5-32b, one it says delivers performance comparable to other large cutting edge models, including Chinese rival DeepSeek and OpenAI’s o1, with only 32 billion parameters.
- THE DECODER: Alibaba's latest AI model demonstrates how reinforcement learning can create efficient systems that match the capabilities of much larger models.
- bdtechtalks.com: Alibaba’s QwQ-32B reasoning model matches DeepSeek-R1, outperforms OpenAI o1-mini
- Last Week in AI: Alibaba’s New QwQ 32B Model is as Good as DeepSeek-R1
- Last Week in AI: LWiAI Podcast #202 - Qwen-32B, Anthropic's $3.5 billion, LLM Cognitive Behaviors
- Last Week in AI: #202 - Qwen-32B, Anthropic's $3.5 billion, LLM Cognitive Behaviors
Classification: