News from the AI & ML world

DeeperML - #alibaba

@bdtechtalks.com //
Alibaba has recently launched Qwen-32B, a new reasoning model, which demonstrates performance levels on par with DeepMind's R1 model. This development signifies a notable achievement in the field of AI, particularly for smaller models. The Qwen team showcased that reinforcement learning on a strong base model can unlock reasoning capabilities for smaller models that enhances their performance to be on par with giant models.

Qwen-32B not only matches but also surpasses models like DeepSeek-R1 and OpenAI's o1-mini across key industry benchmarks, including AIME24, LiveBench, and BFCL. This is significant because Qwen-32B achieves this level of performance with only approximately 5% of the parameters used by DeepSeek-R1, resulting in lower inference costs without compromising on quality or capability. Groq is offering developers the ability to build FAST with Qwen QwQ 32B on GroqCloud™, running the 32B parameter model at ~400 T/s. This model is proving to be very competitive in reasoning benchmarks and is one of the top open source models being used.

The Qwen-32B model was explicitly designed for tool use and adapting its reasoning based on environmental feedback, which is a huge win for AI agents that need to reason, plan, and adapt based on context (outperforms R1 and o1-mini on the Berkeley Function Calling Leaderboard). With these capabilities, Qwen-32B shows that RL on a strong base model can unlock reasoning capabilities for smaller models that enhances their performance to be on par with giant models.

Recommended read:
References :
  • Last Week in AI: LWiAI Podcast #202 - Qwen-32B, Anthropic's $3.5 billion, LLM Cognitive Behaviors
  • Groq: A Guide to Reasoning with Qwen QwQ 32B
  • Last Week in AI: #202 - Qwen-32B, Anthropic's $3.5 billion, LLM Cognitive Behaviors
  • Sebastian Raschka, PhD: This article explores recent research advancements in reasoning-optimized LLMs, with a particular focus on inference-time compute scaling that have emerged since the release of DeepSeek R1.
  • Analytics Vidhya: China is rapidly advancing in AI, releasing models like DeepSeek and Qwen to rival global giants.
  • Last Week in AI: Alibaba’s New QwQ 32B Model is as Good as DeepSeek-R1
  • Maginative: Despite having far fewer parameters, Qwen’s new QwQ-32B model outperforms DeepSeek-R1 and OpenAI’s o1-mini in mathematical benchmarks and scientific reasoning, showcasing the power of reinforcement learning.

@bdtechtalks.com //
References: Groq , Analytics Vidhya , bdtechtalks.com ...
Alibaba's Qwen team has unveiled QwQ-32B, a 32-billion-parameter reasoning model that rivals much larger AI models in problem-solving capabilities. This development highlights the potential of reinforcement learning (RL) in enhancing AI performance. QwQ-32B excels in mathematics, coding, and scientific reasoning tasks, outperforming models like DeepSeek-R1 (671B parameters) and OpenAI's o1-mini, despite its significantly smaller size. Its effectiveness lies in a multi-stage RL training approach, demonstrating the ability of smaller models with scaled reinforcement learning to match or surpass the performance of giant models.

The QwQ-32B is not only competitive in performance but also offers practical advantages. It is available as open-weight under an Apache 2.0 license, allowing businesses to customize and deploy it without restrictions. Additionally, QwQ-32B requires significantly less computational power, running on a single high-end GPU compared to the multi-GPU setups needed for larger models like DeepSeek-R1. This combination of performance, accessibility, and efficiency positions QwQ-32B as a valuable resource for the AI community and enterprises seeking to leverage advanced reasoning capabilities.

Recommended read:
References :
  • Groq: A Guide to Reasoning with Qwen QwQ 32B
  • Analytics Vidhya: Qwen’s QwQ-32B: Small Model with Huge Potential
  • Maginative: Alibaba's Latest AI Model, QwQ-32B, Beats Larger Rivals in Math and Reasoning
  • bdtechtalks.com: Alibaba’s QwQ-32B reasoning model matches DeepSeek-R1, outperforms OpenAI o1-mini
  • Last Week in AI: LWiAI Podcast #202 - Qwen-32B, Anthropic's $3.5 billion, LLM Cognitive Behaviors

Ryan Daws@AI News //
Alibaba's Qwen team has launched QwQ-32B, a 32-billion parameter AI model, designed to rival the performance of much larger models like DeepSeek-R1, which has 671 billion parameters. This new model highlights the effectiveness of scaling Reinforcement Learning (RL) on robust foundation models. QwQ-32B leverages continuous RL scaling to demonstrate significant improvements in areas like mathematical reasoning and coding proficiency.

The Qwen team successfully integrated agent capabilities into the reasoning model, allowing it to think critically, use tools, and adapt its reasoning based on environmental feedback. The model has been evaluated across a range of benchmarks, including AIME24, LiveCodeBench, LiveBench, IFEval, and BFCL, designed to assess its mathematical reasoning, coding proficiency, and general problem-solving capabilities. QwQ-32B is available as open-weight on Hugging Face and on ModelScope under an Apache 2.0 license, allowing for both commercial and research uses.

Recommended read:
References :
  • AI News | VentureBeat: Alibaba's new open source model QwQ-32B matches DeepSeek-R1 with way smaller compute requirements
  • Analytics Vidhya: In the world of large language models (LLMs) there is an assumption that larger models inherently perform better. Qwen has recently introduced its latest model, QwQ-32B, positioning it as a direct competitor to the massive DeepSeek-R1 despite having significantly fewer parameters.
  • AI News: The Qwen team at Alibaba has unveiled QwQ-32B, a 32 billion parameter AI model that demonstrates performance rivalling the much larger DeepSeek-R1. This breakthrough highlights the potential of scaling Reinforcement Learning (RL) on robust foundation models.
  • www.infoworld.com: Alibaba Cloud on Thursday launched QwQ-32B, a compact reasoning model built on its latest large language model (LLM), Qwen2.5-32b, one it says delivers performance comparable to other large cutting edge models, including Chinese rival DeepSeek and OpenAI’s o1, with only 32 billion parameters.
  • THE DECODER: Alibaba's latest AI model demonstrates how reinforcement learning can create efficient systems that match the capabilities of much larger models.
  • bdtechtalks.com: Alibaba’s QwQ-32B reasoning model matches DeepSeek-R1, outperforms OpenAI o1-mini
  • Last Week in AI: Alibaba’s New QwQ 32B Model is as Good as DeepSeek-R1
  • Last Week in AI: LWiAI Podcast #202 - Qwen-32B, Anthropic's $3.5 billion, LLM Cognitive Behaviors
  • Last Week in AI: #202 - Qwen-32B, Anthropic's $3.5 billion, LLM Cognitive Behaviors

Ryan Daws@AI News //
Alibaba's Qwen team has introduced QwQ-32B, a 32 billion parameter AI model that rivals the performance of the much larger DeepSeek-R1. This achievement showcases the potential of scaling Reinforcement Learning (RL) on robust foundation models. The Qwen team has successfully integrated agent capabilities into the reasoning model, enabling it to think critically and utilize tools. This highlights that scaled reinforcement learning can lead to significant advancements in AI performance without necessarily requiring immense computational resources.

QwQ-32B demonstrates that RL scaling can dramatically enhance model intelligence without requiring massive parameter counts. QwQ-32B leverages RL techniques through a reward-based, multi-stage training process. This enables deeper reasoning capabilities, typically associated with much larger models. QwQ-32B has achieved performance comparable to DeepSeek-R1 which underscores the potential of RL to bridge the gap between model size and performance.

Recommended read:
References :
  • AI News | VentureBeat: Alibaba’s new open source model QwQ-32B matches DeepSeek-R1 with way smaller compute requirements
  • MarkTechPost: Qwen Releases QwQ-32B: A 32B Reasoning Model that Achieves Significantly Enhanced Performance in Downstream Task
  • Analytics Vidhya: Discussion on Qwen Chat, noting QwQ-32B’s capabilities.
  • AI News: Alibaba Qwen QwQ-32B: Scaled reinforcement learning showcase
  • Simon Willison's Weblog: QwQ-32B: Embracing the Power of Reinforcement Learning
  • Analytics Vidhya: QwQ-32B Vs DeepSeek-R1: Can a 32B Model Challenge a 671B Parameter Model?
  • MarkTechPost: Alibaba Released Babel: An Open Multilingual Large Language Model LLM Serving Over 90% of Global Speakers
  • www.infoworld.com: Alibaba says its new AI model rivals DeepSeeks’s R-1, OpenAI’s o1
  • IEEE Spectrum: QwQ, DeepSeek-R1 32B, and Sky-T1-R had the highest overthinking scores, and they weren’t any more successful at resolving tasks than nonreasoning models.
  • THE DECODER: Alibaba's QwQ-32B is an efficient reasoning model that rivals much larger AI systems
  • eWEEK: Alibaba unveils QwQ-32B, an AI model rivaling OpenAI and DeepSeek with 98% lower compute costs. A game-changer in AI efficiency, boosting Alibaba’s market position.
  • SiliconANGLE: Alibaba shares jump on new open-source QwQ-32B reasoning model
  • Last Week in AI: Alibaba’s New QwQ 32B Model is as Good as DeepSeek-R1 , Judge Denies Musk’s Request to Block OpenAI’s For-Profit Plan, Alexa Plus’ AI upgrades cost $19.99, and more!
  • Last Week in AI: Alibaba released Qwen-32B, Anthropic raised $3.5 billion,DeepMind introduced BigBench Extra Hard, and more!
  • bdtechtalks.com: Alibaba's QwQ-32B is a new large reasoning model (LRM) with high performance on key benchmarks, improved efficiency and open-source access.
  • Groq: With a community of over one million developers who build FAST, Groq can’t help but want to keep up.
  • Maginative: Information on how Alibaba's Latest AI Model, QwQ-32B, Beats Larger Rivals in Math and Reasoning
  • Analytics Vidhya: Small Model with Huge Potential.
  • Towards AI: Performance Analysis Between QWQ-32B and DeepSeek-R1 and How to Run QWQ-32B Locally on Your Machine

@oodaloop.com //
References: oodaloop.com , ReadWrite , eWEEK ...
Alibaba's CEO, Eddie Wu, has announced that the company's primary objective is now the pursuit of Artificial General Intelligence (AGI). This strategic shift was communicated to investors during an earnings call where Wu emphasized that Alibaba aims to develop models that "extend the boundaries of intelligence." The decision highlights the increasing influence of AI within the technology sector, particularly in the context of Alibaba's traditional e-commerce business, which includes services like AliExpress and Taobao.

This focus on AGI, a type of AI that matches or surpasses human cognitive capabilities, signifies Alibaba's commitment to innovation in the rapidly evolving AI landscape. Wu believes that if AGI is achieved, the AI industry could potentially become the world's largest. The company has already been active in AI development with its Qwen large language models, with the latest version unveiled in January. Alibaba's revenue has seen an 8% year-over-year increase, marking progress in its AI-driven strategies.

Recommended read:
References :
  • oodaloop.com: Alibaba Group Holding Ltd.’s chief executive officer said the pursuit of artificial general intelligence is now the company’s “primary objective.â€�
  • ReadWrite: The CEO of the Chinese e-commerce company Alibaba is reported to have told investors that the “primary objectiveâ€� and focus…
  • www.eweek.com: Alibaba CEO’s focus on theoretical technology comes after a good few years for the e-commerce giant and Qwen maker.
  • eWEEK: Alibaba Commits to AGI, Making AI “The World’s Largest Industryâ€�
  • OODAloop: Alibaba CEO Wu Says AGI Is Now Company’s ‘Primary Objective’
  • readwrite.com: The CEO of the Chinese e-commerce company Alibaba is reported to have told investors that the “primary objectiveâ€� and focus…