News from the AI & ML world

DeeperML - #llms

Ryan Daws@AI News //
References: On my Om , Shelly Palmer , bsky.app ...
Anthropic has announced that its AI assistant Claude can now search the web. This enhancement allows Claude to provide users with more up-to-date and relevant responses by expanding its knowledge base beyond its initial training data. It may seem like a minor feature update, but it's not. It is available to paid Claude 3.7 Sonnet users by toggling on "web search" in their profile settings.

This integration emphasizes transparency, as Claude provides direct citations when incorporating information from the web, enabling users to easily fact-check sources. Claude aims to streamline the information-gathering process by processing and delivering relevant sources in a conversational format. Anthropic believes this update will unlock new use cases for Claude across various industries, including sales, finance, research, and shopping.

Recommended read:
References :
  • On my Om: You can now use Claude to search the internet to provide more up-to-date and relevant responses. With web search, Claude has access to the latest events and information, boosting its accuracy on tasks that benefit from the most recent data.
  • Shelly Palmer: Most heavy LLM users will tell you that ChatGPT is the GOAT, but they prefer Claude for writing. Why wasn't Claude the GOAT?
  • AI News: Anthropic has announced its AI assistant Claude can now search the web, providing users with more up-to-date and relevant responses.
  • bsky.app: Simon Willison's notes on the new web search feature for Claude
  • venturebeat.com: VentureBeat article on Anthropic giving Claude real-time web search
  • Analytics Vidhya: Claude AI Now Supports Web Search ğŸŒ
  • Maginative: Anthropic Finally Adds Search Capabilities to Its AI Assistant
  • bsky.app: Anthropic ships a new "web search" feature for their Claude consumer apps today, here are my notes - it's frustrating that they don't share details on whether the underlying index is their own or run by a partner
  • Ken Yeung: Intercom is doubling down on AI-driven customer support with a significant expansion of its Fin agent.
  • THE DECODER: Anthropic's new 'think tool' lets Claude take notes to solve complex problems
  • www.producthunt.com: The "think" tool from Claude
  • www.techradar.com: The ultimate AI search face-off - I pitted Claude's new search tool against ChatGPT Search, Perplexity, and Gemini, the results might surprise you
  • www.tomsguide.com: Claude 3.7 Sonnet now supports real-time web searching — but there's a catch

@the-decoder.com //
Perplexity AI has launched Deep Research, an AI-powered research tool aimed at competing with OpenAI and Google Gemini. Using DeepSeek-R1, Perplexity is offering comprehensive research reports at a much lower cost than OpenAI, with 500 queries per day for $20 per month compared to OpenAI's $200 per month for only 100 queries. The new service automatically conducts dozens of searches and analyzes hundreds of sources to produce detailed reports in one to two minutes.

Perplexity claims Deep Research performs 8 searches and consults 42 sources to generate a 1,300-word report in under 3 minutes. The company says that Deep Research tool works particularly well for finance, marketing, and technology research. The service is launching first on web browsers, with iOS, Android, and Mac versions planned for later release. Perplexity CEO Aravind Srinivas stated he wants to keep making it faster and cheaper for the interest of humanity.

Recommended read:
References :
  • the-decoder.com: Perplexity uses Deepseek-R1 to offer Deep Research 10 times cheaper than OpenAI
  • www.analyticsvidhya.com: Enhancing Multimodal RAG with Deepseek Janus Pro
  • www.marktechpost.com: DeepSeek AI Introduces CODEI/O: A Novel Approach that Transforms Code-based Reasoning Patterns into Natural Language Formats to Enhance LLMs’ Reasoning Capabilities
  • venturebeat.com: Perplexity just made AI research crazy cheap—what that means for the industry
  • Analytics Vidhya: The landscape of AI-powered research just became even more competitive with the launch of Perplexity’s Deep Research. Previously, OpenAI and Google Gemini were leading the way in this space, and now Perplexity has joined the ranks.
  • iHLS: New York State Bans DeepSeek AI App Over Security Concerns
  • NextBigFuture.com: Does DeepSeek Impact the Future of AI Data Centers?
  • THE DECODER: Perplexity's Deep Research utilizes DeepSeek-R1 for generating comprehensive research reports.
  • www.ghacks.net: Perplexity AI has unveiled its latest feature, the 'Deep Research' tool, designed to enhance users' ability to conduct comprehensive research on complex topics.
  • PCMag Middle East ai: Perplexity Launches a Free 'Deep Research' AI Tool
  • bsky.app: Perplexity follows OpenAI with the release of its Deep Research.
  • techstrong.ai: Perplexity AI Launches a Deep Research Tool to Help Humans Research, Deeply
  • Data Phoenix: Perplexity has launched Deep Research, a free AI-powered research tool that can analyze hundreds of sources in minutes to create comprehensive reports across various domains, promising to save users significant research time.
  • eWEEK: Perplexity 1776 Model Fixes DeepSeek-R1’s “Refusal to Respond to Sensitive Topicsâ€�

@Google DeepMind Blog //
References: Google DeepMind Blog , AI News ,
ARC Prize has launched ARC-AGI-2, its toughest AI benchmark yet, accompanied by the announcement of their 2025 competition with $1 million in prizes. ARC-AGI-2 aims to push the limits of general and adaptive AI. As AI progresses beyond narrow tasks to general intelligence, these challenges aim to uncover capability gaps and actively guide innovation. ARC-AGI-2 is designed to be relatively easy for humans, who can solve every task in under two attempts, yet hard or impossible for AI, focusing on areas like symbolic interpretation, compositional reasoning, and contextual rule application.

The benchmark includes datasets with varying visibility and includes the following characteristics: symbolic interpretation, compositional reasoning and contextual rule application. Most existing benchmarks focus on superhuman capabilities, testing advanced, specialised skills. The competition challenges AI developers to attain an 85% accuracy rating on ARC-AGI-2’s private evaluation dataset.

Recommended read:
References :
  • Google DeepMind Blog: FACTS Grounding: A new benchmark for evaluating the factuality of large language models
  • AI News: ARC Prize launches its toughest AI benchmark yet: ARC-AGI-2
  • eWEEK: New AI Benchmark ARC-AGI-2 ‘Significantly Raises the Bar for AI’

Divya@gbhackers.com //
Researchers from Duke University and Carnegie Mellon University have successfully jailbroken several leading AI language models, including OpenAI’s o1/o3, DeepSeek-R1, and Google’s Gemini 2.0 Flash. The team developed a novel attack method called Hijacking Chain-of-Thought (H-CoT), which exploits the reasoning processes of these models to bypass safety mechanisms designed to prevent harmful outputs. This research highlights significant security vulnerabilities in advanced AI systems and raises concerns about their potential misuse.

The researchers introduced the Malicious-Educator benchmark, which utilizes seemingly harmless educational prompts to mask dangerous requests. They found that all tested models failed to consistently recognize these contextual deceptions. For example, DeepSeek-R1 proved particularly susceptible to financial crime queries, providing actionable money laundering steps in a high percentage of test cases. The team has shared mitigation strategies with affected vendors.

Recommended read:
References :
  • gbhackers.com: Researchers Jailbreak OpenAI o1/o3, DeepSeek-R1, and Gemini 2.0 Flash Models
  • Talkback Resources: GitHub - dukeceicenter/jailbreak-reasoning-openai-o1o3-deepseek-r1 [mal]
  • The Register - Software: How nice that state-of-the-art LLMs reveal their reasoning ... for miscreants to exploit

@Google DeepMind Blog //
Researchers are making strides in understanding how AI models think. Anthropic has developed an "AI microscope" to peek into the internal processes of its Claude model, revealing how it plans ahead, even when generating poetry. This tool provides a limited view of how the AI processes information and reasons through complex tasks. The microscope suggests that Claude uses a language-independent internal representation, a "universal language of thought", for multilingual reasoning.

The team at Google DeepMind introduced JetFormer, a new Transformer designed to directly model raw data. This model, capable of both understanding and generating text and images seamlessly, maximizes the likelihood of raw data without depending on any pre-trained components. Additionally, a comprehensive benchmark called FACTS Grounding has been introduced to evaluate the factuality of large language models (LLMs). This benchmark measures how accurately LLMs ground their responses in provided source material and avoid hallucinations, aiming to improve trust and reliability in AI-generated information.

Recommended read:
References :
  • Google DeepMind Blog: FACTS Grounding: A new benchmark for evaluating the factuality of large language models
  • THE DECODER: Anthropic's AI microscope reveals how Claude plans ahead when generating poetry

Matthew S.@IEEE Spectrum //
References: IEEE Spectrum , Composio
Recent research has revealed that AI reasoning models, particularly Large Language Models (LLMs), are prone to overthinking, a phenomenon where these models favor extended internal reasoning over direct interaction with the problem's environment. This overthinking can negatively impact their performance, leading to reduced success rates in resolving issues and increased computational costs. The study highlights a crucial challenge in training AI models: finding the optimal balance between reasoning and efficiency.

The study, conducted by researchers, tasked leading reasoning LLMs with solving problems in benchmark. The results indicated that reasoning models overthought nearly three times as often as their non-reasoning counterparts. Furthermore, the more a model overthought, the fewer problems it successfully resolved. This suggests that while enhanced reasoning capabilities are generally desirable, excessive internal processing can be detrimental, hindering the model's ability to arrive at correct and timely solutions. This raises questions about how to effectively train models to utilize just the right amount of reasoning, avoiding the pitfalls of "analysis paralysis."

Recommended read:
References :
  • IEEE Spectrum: It’s Not Just Us: AI Models Struggle With Overthinking
  • Composio: CoT Reasoning Models – Which One Reigns Supreme in 2025?

@github.com //
References: github.com , LessWrong
Latent Adversarial Training (LAT) has emerged as a promising method for enhancing the safety of Large Language Models (LLMs). A recent study compared LAT to standard Supervised Safety Fine-Tuning (SSFT) and Embedding Space Adversarial Training (AT) and found that LAT encodes refusal behavior in a more distributed way across the model's latent space. This means that instead of relying on a few specific elements, refusal is woven into the model's overall structure, potentially making it more resilient. The study investigated this by generating refusal vectors using each method.

The results indicated that refusal vectors computed from the LAT model were more effective at triggering refusal ablation attacks across multiple models, lowering refusal rates when compared to the other approaches. However, paradoxically, the models trained with LAT maintained the highest refusal rates and were more robust overall against these attacks. This is likely because LAT allows the models to explore a wider range of responses through hidden layer perturbations creating a more comprehensive understanding of refusal. However, the researchers also highlight a potential downside as the more robust encoding of refusal behaviour could be exploited by malicious actors leading to more effective refusal attacks.

Recommended read:
References :
  • github.com: Latent Adversarial Training (LAT) Improves the Representation of Refusal
  • LessWrong: Latent Adversarial Training (LAT) Improves the Representation of Refusal

Ali Azhar@AIwire //
References: AIwire
Microsoft is reportedly developing its own large language models (LLMs), internally called MAI, to compete directly with OpenAI, a company in which Microsoft has invested billions. The move signifies a potential shift in Microsoft's AI strategy, aiming to reduce its dependence on external partners and decrease the costs associated with using OpenAI's models. Microsoft is experimenting with integrating these LLMs into its existing AI products, particularly those based on Microsoft Copilot, as well as Microsoft Teams and Azure cloud services.

Microsoft plans to release the new LLMs as an API for external developers by the end of the year, enabling them to integrate these models into their own applications. This initiative reflects a desire to optimize AI infrastructure and create models tailored specifically for enterprise applications. While Microsoft maintains close ties with OpenAI, the development of in-house LLMs suggests a growing ambition to control its AI destiny and offer more competitive solutions within the rapidly evolving AI landscape.

Recommended read:
References :
  • AIwire: Rival or Partner? Microsoft Develops Its Own LLMs to Compete with OpenAI

@www.eweek.com //
References: www.eweek.com
OpenAI has been actively advancing its AI capabilities while also focusing on safety and real-world applicability. The company developed SWE-Lancer, a benchmark designed to evaluate how well large language models (LLMs) can perform in software engineering tasks. This test assessed how much money LLMs, including Claude 3.5 Sonnet and GPT-4o, could earn by completing jobs on platforms like Upwork. While the models showed promise, researchers found that they still struggled to solve the majority of tasks, highlighting the challenges of applying AI to complex real-world scenarios.

In addition to practical applications, OpenAI is dedicated to AI safety research. The company uses prompt evaluation techniques to combat potential misuse, specifically focusing on preventing AI from aiding in bio-weapon research. They have also expanded the accessibility of the Operator AI agent to multiple countries, including India, further integrating AI-powered automation into daily tasks. These efforts demonstrate OpenAI's commitment to both innovation and responsible development in the rapidly evolving field of artificial intelligence.

Recommended read:
References :
  • www.eweek.com: OpenAI created SWE-Lancer, a benchmark test of how much LLMs could earn from doing software engineering gig work.