Ben Dickson@AI News | VentureBeat
//
DeepSeek, a Chinese AI company, has achieved a breakthrough in AI reward modeling that promises to enhance the reasoning and responsiveness of AI systems. Collaborating with Tsinghua University researchers, DeepSeek developed a technique called "Inference-Time Scaling for Generalist Reward Modeling," demonstrating improved performance compared to existing methods and competitive results against established public reward models. This innovation aims to improve how AI systems learn from human preferences, a key factor in developing more useful and aligned artificial intelligence.
DeepSeek's new approach involves a dual method combining Generative Reward Modeling (GRM) and Self-Principled Critique Tuning (SPCT). GRM provides flexibility in handling various input types and enables scaling during inference time, offering a richer representation of rewards through language compared to previous scalar approaches. SPCT, a learning method, fosters scalable reward-generation behaviors in GRMs through online reinforcement learning. One of the paper's authors explained that this combination allows principles to be generated based on the input query and responses, adaptively aligning the reward generation process. The SPCT technique addresses challenges in creating generalist reward models capable of handling broader tasks. These challenges include input flexibility, accuracy, inference-time scalability, and learning scalable behaviors. By creating self-guiding critiques, SPCT promises more scalable intelligence for enterprise LLMs, particularly in open-ended tasks and domains where current models struggle. DeepSeek has also released models like DeepSeek-V3 and DeepSeek-R1, which have achieved performance close to, and sometimes exceeding, leading proprietary models while using fewer training resources. These advancements signal that cutting-edge AI is not solely the domain of closed labs and highlight the importance of efficient model architecture, training algorithms, and hardware integration. References :
Classification:
Matthew S.@IEEE Spectrum
//
Recent research indicates that AI models, particularly large language models (LLMs), can struggle with overthinking and analysis paralysis, impacting their efficiency and success rates. A study has found that reasoning LLMs sometimes overthink problems, which leads to increased computational costs and a reduction in their overall performance. This issue is being addressed through various optimization techniques, including scaling inference-time compute, reinforcement learning, and supervised fine-tuning, to ensure models use only the necessary amount of reasoning for tasks.
The size and training methods of these models play a crucial role in their reasoning abilities. For instance, Alibaba's Qwen team introduced QwQ-32B, a 32-billion-parameter model that outperforms much larger rivals in key problem-solving tasks. QwQ-32B achieves superior performance in math, coding, and scientific reasoning using multi-stage reinforcement learning, despite being significantly smaller than DeepSeek-R1. This advancement highlights the potential of reinforcement learning to unlock reasoning capabilities in smaller models, rivaling the performance of giant models while requiring less computational power. References :
Classification:
|
BenchmarksBlogsResearch Tools |