News from the AI & ML world

DeeperML - #aiinnovation

staff@insideAI News //
References: AiThority , insideAI News , Dataconomy ...
IBM has launched watsonx AI Labs, a developer-first innovation hub located in New York City. The new lab is designed to accelerate the adoption of AI at scale by connecting IBM's enterprise resources and expertise with AI developers focused on building AI applications for business. Located in Manhattan at IBM's new offices at One Madison, watsonx AI Labs aims to connect IBM’s network of engineering labs, bringing together IBM researchers and engineers in a collaborative hub for co-creating and advancing agentic AI solutions.

The watsonx AI Labs is intended to co-create generative AI solutions with IBM clients, nurture AI talent within New York City, and advance enterprise AI implementations. IBM plans to work with startups, scale-ups, and enterprises to discover AI value through this initiative. New York City has a growing AI ecosystem, with more than 2,000 AI startups and an AI workforce that grew by almost 25% from 2022 to 2023. Since 2019, over 1,000 AI-related companies in New York City have collectively raised $27 billion in funding.

As part of its investment in AI and commitment to the local startup ecosystem, IBM also announced the acquisition of Seek AI. Seek AI is a New York City-based startup that specializes in building AI agents that leverage enterprise data, providing businesses with a natural language interface to query and analyze corporate data stores. Seek AI's expertise will be integrated into watsonx AI Labs, helping businesses leverage agentic AI to extract value from their data and improve data analysis and summarization capabilities.

Recommended read:
References :
  • AiThority: IBM today announced watsonx AI Labs, a new, developer-first innovation hub in New York City, designed to supercharge AI builders and accelerate AI adoption at scale. watsonx AI Labs connects IBM’s enterprise resources and expertise with the next generation of AI developers in order to build breakthrough AI applications for business. Located in the heart of Manhattan at IBM’s new […] The post appeared first on .
  • insideAI News: IBM Unveils watsonx AI Labs in New York City
  • IBM - Announcements: New AI initiative will co-create gen AI solutions with IBM clients, nurture NYC talent, advance enterprise AI implementations
  • Dataconomy: IBM announced Monday its acquisition of Seek AI, a natural language AI platform for enterprise data inquiries, for an undisclosed amount.
  • The Register - Software: IBM Watson zombie brand shuffles forward with new AI lab in NYC
  • www.lifewire.com: IBM Acquires Seek AI to Fuel Enterprise Innovation in NYC
  • www.cio.com: IBM acquires Seek AI, launches Watsonx Labs to scale enterprise AI
  • aithority.com: IBM today announced watsonx AI Labs, a new, developer-first innovation hub in New York City, designed to supercharge AI builders and accelerate AI adoption at scale.

Ken Yeung@Ken Yeung //
References: Ken Yeung , eWEEK
Microsoft is aggressively pushing AI innovation to the edge, a key theme highlighted at Microsoft Build 2025. The company is developing and integrating AI capabilities into various platforms, aiming to create smarter, faster experiences across devices. This initiative involves not only expanding cloud capabilities but also embedding AI agents into browsers, websites, the operating system, and everyday workflows. Microsoft envisions a future of AI-powered productivity where human workers partner with autonomous agents to streamline tasks and enhance efficiency.

Microsoft is also making strides in AI-driven weather forecasting with its latest AI model, Aurora. This model promises accurate 10-day forecasts in seconds, a significant improvement over traditional models that take hours. Aurora isn't limited to weather; it can also handle any Earth system with available data, opening possibilities for forecasting air pollution, cyclones, and other environmental factors. While the model "doesn't know the laws of physics," its data-driven approach delivers detailed and quick forecasts, demonstrating the potential of AI in revolutionizing how we understand and predict environmental changes.

A core component of Microsoft's AI strategy is the integration of the Model Context Protocol (MCP) into Windows 11. This integration aims to transform the operating system into an "agentic" platform, where AI agents can securely interact with apps and system tools to carry out tasks across files and services. MCP acts as a standardized communication protocol, facilitating seamless interaction between AI agents, applications, and services. With security measures in place, MCP allows for powerful AI integrations while mitigating risks, enabling new forms of AI-driven experiences for Windows 11 users.

Recommended read:
References :
  • Ken Yeung: Microsoft Pushes AI to the Edge
  • eWEEK: Microsoft’s Big Bet on AI Agents: Model Context Protocol in Windows 11

Kevin Okemwa@windowscentral.com //
OpenAI and Microsoft are reportedly engaged in high-stakes negotiations to revise their existing partnership, a move prompted by OpenAI's aspirations for an initial public offering (IPO). The discussions center around redefining the terms of their strategic alliance, which has seen Microsoft invest over $13 billion in OpenAI since 2019. A key point of contention is Microsoft's desire to secure guaranteed access to OpenAI's AI technology beyond the current contractual agreement, set to expire in 2030. Microsoft is reportedly willing to sacrifice some equity in OpenAI to ensure long-term access to future AI models.

These negotiations also entail OpenAI potentially restructuring its for-profit arm into a Public Benefit Corporation (PBC), a move that requires Microsoft's approval as the startup's largest financial backer. The PBC structure would allow OpenAI to pursue commercial goals and attract further capital, paving the way for a potential IPO. However, the non-profit entity would retain overall control. OpenAI reportedly aims to reduce Microsoft's revenue share from 20% to a share of 10% by 2030, a year when the company forecasts $174B in revenue.

Tensions within the partnership have reportedly grown as OpenAI pursues agreements with Microsoft competitors and targets overlapping enterprise customers. One senior Microsoft executive expressed concern over OpenAI's attitude, stating that they seem to want Microsoft to "give us money and compute and stay out of the way." Despite these challenges, Microsoft remains committed to the partnership, recognizing its importance in the rapidly evolving AI landscape.

Recommended read:
References :
  • the-decoder.com: Microsoft could sacrifice some OpenAI shares - but wants to secure access to AI technology
  • www.techradar.com: OpenAI and Microsoft in talks to revise terms and renew partnership, FT reports
  • The Rundown AI: OpenAI, Microsoft's 'high-stakes' negotiations
  • www.computerworld.com: OpenAI’s IPO aspirations prompt rethink of Microsoft alliance
  • www.windowscentral.com: OpenAI wants Microsoft to provide money and compute and stay out of the way as it renegotiates multi-billion-dollar partnership
  • The Tech Portal: According to media reports, OpenAI and Microsoft are now negotiating to redefine… Content originally published on

Kevin Okemwa@windowscentral.com //
OpenAI and Microsoft are reportedly engaged in high-stakes negotiations to revise their existing partnership, a move prompted by OpenAI's aspirations for an initial public offering (IPO). The discussions center around redefining the terms of their strategic alliance, which has seen Microsoft invest over $13 billion in OpenAI since 2019. A key point of contention is Microsoft's desire to secure guaranteed access to OpenAI's AI technology beyond the current contractual agreement, set to expire in 2030. Microsoft is reportedly willing to sacrifice some equity in OpenAI to ensure long-term access to future AI models.

These negotiations also entail OpenAI potentially restructuring its for-profit arm into a Public Benefit Corporation (PBC), a move that requires Microsoft's approval as the startup's largest financial backer. The PBC structure would allow OpenAI to pursue commercial goals and attract further capital, paving the way for a potential IPO. However, the non-profit entity would retain overall control. OpenAI reportedly aims to reduce Microsoft's revenue share from 20% to a share of 10% by 2030, a year when the company forecasts $174B in revenue.

Tensions within the partnership have reportedly grown as OpenAI pursues agreements with Microsoft competitors and targets overlapping enterprise customers. One senior Microsoft executive expressed concern over OpenAI's attitude, stating that they seem to want Microsoft to "give us money and compute and stay out of the way." Despite these challenges, Microsoft remains committed to the partnership, recognizing its importance in the rapidly evolving AI landscape.

Recommended read:
References :
  • the-decoder.com: OpenAI is planning a major overhaul of its corporate structure as it prepares for a possible IPO, but its most important backer, Microsoft, is setting conditions—and the relationship remains tense.
  • www.techradar.com: OpenAI and Microsoft could be revising their agreement as OpenAI explores the future of listing as a public company.
  • www.windowscentral.com: Microsoft is reportedly in the process of reworking its partnership, willing to give up some equity to continue accessing sophisticated AI models beyond 2030.
  • The Rundown AI: OpenAI, Microsoft's 'high-stakes' negotiations

Carl Franzen@AI News | VentureBeat //
OpenAI is facing increased scrutiny regarding its operational structure, leading to a notable reversal in its plans. The company, initially founded as a nonprofit, will now retain the nonprofit's governance control, ensuring that the original mission remains at the forefront. This decision comes after "constructive dialogue" with the Attorney Generals of Delaware and California, suggesting a potential legal challenge if OpenAI had proceeded with its initial plan to convert fully into a profit-maximizing entity. The company aims to maintain its commitment to developing Artificial General Intelligence (AGI) for the benefit of all humanity, and CEO Sam Altman insists that OpenAI is "not a normal company and never will be."

As part of this restructuring, OpenAI will transition its for-profit arm, currently an LLC, into a Public Benefit Corporation (PBC). This move aims to balance the interests of shareholders with the company's core mission. The nonprofit will remain a large shareholder in the PBC, giving it the resources to support its beneficial objectives. The company is getting rid of the capped profit structure, which may allow the company to be more aggressive in the marketplace. Bret Taylor, Chairman of the Board of OpenAI, emphasized that the company will continue to be overseen and controlled by the nonprofit. This updated plan demonstrates a commitment to the original vision of OpenAI while adapting to the demands of funding AGI development, which Altman estimates will require "hundreds of billions of dollars of compute."

Further demonstrating its commitment to advancing AI technology, OpenAI is reportedly acquiring Windsurf (formerly Codeium) for $3 billion. While specific details of the acquisition are not provided, it's inferred that Windsurf's coding capabilities will be integrated into OpenAI's AI models, potentially enhancing their coding abilities. The acquisition aligns with OpenAI's broader strategy of pushing the boundaries of AI capabilities and making them accessible to a wider audience. This move may improve the abilities of models like the o-series (rewarding verifiable math, science, and code solutions) and agentic o3 models (rewarding tool use), which the industry is pushing forward aggressively with new training approaches.

Recommended read:
References :

Alexey Shabanov@TestingCatalog //
Meta is actively expanding the capabilities of its standalone Meta AI app, introducing new features focused on enhanced personalization and functionality. The company is developing a "Discover AIs" tab, which could serve as a hub for users to explore and interact with various AI assistants, potentially including third-party or specialized models. This aligns with Meta’s broader strategy to integrate personalized AI agents across its apps and hardware. Meta launched a dedicated Meta AI app powered by Llama 4 that focuses on offering more natural voice conversations and can leverage user data from Facebook and Instagram to provide tailored responses.

Meta is also testing a "reasoning" mode, suggesting the company aims to provide more transparent and advanced explanations in its AI assistant's responses. While the exact implementation remains unclear, the feature could emphasize structured logic or chain-of-thought capabilities, similar to developments in models from OpenAI and Google DeepMind. This would give users greater insight into how the AI derives its answers, potentially boosting trust and utility for complex queries.

Further enhancing user experience, Meta is working on new voice settings, including "Focus on my voice" and "Welcome message." "Focus on my voice" could improve the AI's ability to isolate and respond to the primary user's speech in environments with multiple speakers. The "Welcome message" feature might offer a customizable greeting or onboarding experience when the assistant is activated. These features are particularly relevant for Meta’s hardware ambitions, such as its Ray-Ban smart glasses and future AR devices, where voice interaction plays a critical role. To ensure privacy, Meta is also developing Private Processing for AI tools on WhatsApp, allowing users to leverage AI in a secure way.

Recommended read:
References :
  • Engineering at Meta: We are inspired by the possibilities of AI to help people be more creative, productive, and stay closely connected on WhatsApp, so we set out to build a new technology that allows our users around the world to use AI in a privacy-preserving way. We’re sharing an early look into Private Processing, an optional capability
  • TestingCatalog: Discover Meta AI's latest features: "Discover AIs" tab, "reasoning" mode, and new voice settings. Enhance your AI experience with personalized and advanced interactions.
  • Data Phoenix: Meta just launched a standalone Meta AI app powered by Llama 4 that focuses on offering more natural voice conversations and can leverage user data from Facebook and Instagram to provide tailored responses.
  • SiliconANGLE: Meta announces standalone AI app for personalized assistance

AiRabbit@AI Rabbit Blog //
References: AI Rabbit Blog
Open-source AI chatbots are gaining popularity as viable alternatives to proprietary options like ChatGPT. Platforms such as LibreChat and openwebui offer increased flexibility and a broader range of features. LibreChat, in particular, supports diverse models and tools, including MCP and custom APIs, empowering users to construct versatile AI agents tailored to their specific needs. Setting up these chatbots often involves using Docker and configuring MCP services, allowing for a customizable and powerful AI experience.

xAI is actively developing updates for its Grok AI model, with Grok 3.5 expected to bring significant upgrades in model capabilities. Furthermore, Grok 4 is planned for release later this year, demonstrating xAI's commitment to rapid iteration and improvement. These advancements aim to close the feature gap with leading AI products, offering users a more competitive and comprehensive AI solution.

New features are also on the horizon for Grok, including a Vision feature in voice mode, which will allow users to share their camera or grant Grok access to it, similar to functionalities in ChatGPT and Gemini. Memory reference capabilities are being developed for Grok on the web, enabling the AI to recall and reference previous conversations. An image editing tool is also in the works, allowing users to edit images using Grok's generative AI capabilities, demonstrating a focus on versatility and enhanced user interaction.

Recommended read:
References :
  • AI Rabbit Blog: LibreChat - The Open Source Answer to ChatGPT & CustomGPTs

@cloud.google.com //
Google is doubling down on empowering AI innovation with new enhancements to its Google Kubernetes Engine (GKE). Unveiled at Google Cloud Next 2025, these updates focus on simplifying AI adoption, scaling AI workloads efficiently, and optimizing AI inference performance. The enhancements aim to address the challenges of deploying generative AI by providing tools for infrastructure selection, intelligent load balancing, and cost reduction, all while leveraging the power of Kubernetes. These advancements reflect the increasing demand for AI inference capabilities as businesses seek to solve real-world problems with AI.

Google has introduced several key features to streamline AI inference on GKE, including GKE Inference Quickstart, GKE TPU serving stack, and GKE Inference Gateway. GKE Inference Quickstart helps users select the optimal accelerator, model server, and scaling configuration, providing insights into instance types, model compatibility, and performance benchmarks. The GKE TPU serving stack, with support for Tensor Processing Units (TPUs) and vLLM, enables seamless portability across GPUs and TPUs. Furthermore, the GKE Inference Gateway introduces AI-aware scaling and load balancing techniques, resulting in significant improvements in serving costs, tail latency, and throughput.

These GKE enhancements are designed to equip organizations for the agentic AI era, where multiple AI agents collaborate to accomplish tasks across various systems. Google is also offering tools like the Agent Development Kit (ADK), Agent Garden, and Agent Engine on Vertex AI to build and deploy custom agents. Google Cloud WAN, the company's internal advanced networking technology, is now available to customers, providing a high-performance, secure, and reliable network infrastructure for AI workloads. These efforts demonstrate Google Cloud's commitment to providing an open, comprehensive platform for production AI, enabling businesses to harness the power of AI with ease and efficiency.

Recommended read:
References :
  • Practical Technology: Google reveals new Kubernetes and GKE enhancements for AI innovation
  • AI & Machine Learning: Details the new GKE inference capabilities that reduce costs, tail latency and increase throughput.
  • www.itpro.com: Google Cloud Next 2025: Targeting easy AI
  • AI & Machine Learning: Kubernetes, your AI superpower: How Google Kubernetes Engine powers AI innovation
  • cloud.google.com: Delivering an application-centric, AI-powered cloud for developers and operators
  • Runtime: Google promotes k8s for AI; IBM says use a mainframe

Ali Azhar@AIwire //
References: AIwire
Nvidia is strategically expanding its AI capabilities with recent acquisitions, signaling a push towards full-stack AI control. The company is reportedly in advanced talks to acquire Lepton AI, a startup specializing in renting out Nvidia-powered servers for AI development. This move, along with the acquisition of synthetic data startup Gretel, demonstrates Nvidia's ambition to move beyond hardware and offer comprehensive AI solutions.

Nvidia's acquisition strategy aims to enhance its cloud-based AI offerings and counter competition from cloud providers developing their own AI chips. The company's interest in Lepton AI and the acquisition of Gretel, known for privacy-safe AI training data, are key steps in its strategy to become a full-stack enabler of AI development. These acquisitions are aimed at integrating into the AI development pipeline and providing more complete solutions for AI development.

Recommended read:
References :
  • AIwire: Nvidia showcased its latest advancements in AI and accelerated computing at GTC 2025.

Jaime Hampton@AIwire //
References: AI News , Sify , AIwire ...
DeepSeek's innovative AI models are reshaping China's AI data center infrastructure, leading to a market disruption and potentially underutilized resources. The company's DeepSeek-V3 model has demonstrated performance that rivals ChatGPT but at a significantly reduced cost. This has altered the demand for extensive GPU clusters used in traditional AI training, shifting the focus towards hardware prioritizing low-latency, particularly near tech hubs. This has resulted in increased speculation as well as experienced players who are now posed with the challenge of the DeepSeek V3.

The open-source nature of DeepSeek’s model is also allowing smaller players to compete without the need for extensive pretraining, which is undermining the demand for large data centers. DeepSeek-V3, which runs at 20 tokens per second on a Mac Studio, poses a new challenge for existing AI models. Chinese AI startups are now riding DeepSeek's momentum and building an ecosystem that is revolutionizing the AI landscape. This narrows the technology divide between China and the United States.

Recommended read:
References :
  • AI News: DeepSeek disruption: Chinese AI innovation narrows global technology divide
  • Sify: DeepSeek’s AI Revolution: Creating an Entire AI Ecosystem
  • Composio: Deepseek v3-0324 vs. Claude 3.7 Sonnet
  • AIwire: Report: China’s Race to Build AI Datacenters Has Hit a Wall
  • Quinta?s weblog: DeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAI

Dashveenjit Kaur@AI News //
References: venturebeat.com , AI News , Nordic APIs ...
Chinese AI startup DeepSeek is shaking up the global technology landscape with its latest large language model, DeepSeek-V3-0324. This new model has been lauded for matching the performance of American AI models, while boasting significantly lower development costs. According to Lee Kai-fu, CEO of Chinese startup 01.AI, the gap between Chinese and American AI capabilities has narrowed dramatically, with China even ahead in some specific areas.

DeepSeek-V3-0324 features enhanced reasoning capabilities and improved performance in multiple benchmarks, particularly in mathematics. The model scored 59.4 on the American Invitational Mathematics Examination (AIME), a significant improvement over its predecessor. Häme University lecturer Kuittinen Petri noted DeepSeek's achievements were realized with just a fraction of the resources available to competitors like OpenAI. This breakthrough has been attributed to DeepSeek’s focus on algorithmic efficiency and novel approaches to model architecture, allowing them to overcome restrictions on access to the latest silicon.

This disruption is not going unnoticed, when DeepSeek launched its R1 model in January, America’s Nasdaq plunged 3.1%, while the S&P 500 fell 1.5%. While DeepSeek claimed a $5.6 million training cost, this represented only the marginal cost of the final training run. SemiAnalysis estimates DeepSeek's actual hardware investment at closer to $1.6 billion, with hundreds of millions in operating costs. The developments present opportunities and challenges for the.

Recommended read:
References :
  • venturebeat.com: DeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAI
  • AI News: DeepSeek disruption: Chinese AI innovation narrows global technology divide
  • Sify: DeepSeek’s AI Revolution: Creating an Entire AI Ecosystem
  • Nordic APIs: ChatGPT vs. DeepSeek: A Side-by-Side Comparison
  • Composio: Deepseek v3-0324 vs. Claude 3.7 Sonnet

Ryan Daws@AI News //
References: SiliconANGLE , venturebeat.com , AI News ...
DeepSeek, a Chinese AI company, has released DeepSeek V3-0324, an updated AI model that demonstrates impressive performance. The model is now running at 20 tokens per second on a Mac Studio. This model is said to contain 685 billion parameters and its cost-effectiveness challenges the dominance of American AI models, signaling that China continues to innovate in AI despite chip restrictions. Reports from early testers show improvements over previous versions and the model tops non-reasoning AI models in open-source first.

This new model runs on consumer-grade hardware, specifically Apple's Mac Studio with the M3 Ultra chip, diverging from the typical data center requirements for AI. It is freely available for commercial use under the MIT license. According to AI researcher Awni Hannun, the model runs at over 20 tokens per second on a 512GB M3 Ultra. The company has made no formal announcement, just an empty README file and the model weights themselves. This stands in contrast to the carefully orchestrated product launches by Western AI companies.

Recommended read:
References :
  • SiliconANGLE: DeepSeek today released an improved version of its DeepSeek-V3 large language model under a new open-source license.
  • venturebeat.com: DeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAI
  • AI News: Chinese AI innovation is reshaping the global technology landscape, challenging assumptions about Western dominance in advanced computing. Recent developments from companies like DeepSeek illustrate how quickly China has adapted to and overcome international restrictions through creative approaches to AI development.
  • AI News: DeepSeek V3-0324 tops non-reasoning AI models in open-source first
  • MarkTechPost: DeepSeek AI Unveils DeepSeek-V3-0324: Blazing Fast Performance on Mac Studio, Heating Up the Competition with OpenAI
  • Cloud Security Alliance: Cloud Security Alliance: DeepSeek: Behind the Hype and Headlines
  • Quinta?s weblog: DeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAI
  • Composio: Deepseek v3-0324 vs. Claude 3.7 Sonnet

Jesus Rodriguez@TheSequence //
OpenAI has recently launched new audio features and tools aimed at enhancing the capabilities of AI agents. The releases include updated transcription and text-to-speech models, as well as tools for building AI agents. The audio models, named gpt-4o-transcribe and gpt-4o-mini-transcribe, promise better performance than the previous Whisper models, achieving lower word error rates across multiple languages and demonstrating improvements in challenging audio conditions like varying accents and background noise. These models are built on top of language models, making them potentially vulnerable to prompt injection attacks.

OpenAI also unveiled new tools for AI agent development, featuring a Responses API, built-in web search, file search, and computer use functionalities, alongside an open-source Agents SDK. Furthermore, they introduced o1 Pro, a new reasoning model, positioned for complex reasoning tasks, comes with a high cost, priced at $150 per million input tokens and $600 per million output tokens. The gpt-4o-mini-tts text-to-speech model introduces "steerability", allowing developers to control the tone and delivery of the model.

Recommended read:
References :
  • Data Phoenix: OpenAI Launches New Tools for Building AI Agents
  • Fello AI: OpenAI's new o1 Pro pricing strategy with a substantial markup compared to previous models.
  • TheSequence: The Sequence Engineering #513: A Deep Dive Into OpenAI's New Tools for Developing AI Agents
  • AI News | VentureBeat: OpenAI’s new voice AI model gpt-4o-transcribe lets you add speech to your existing text apps in seconds
  • Windows Copilot News: Canadian Media Outlets Sue OpenAI Over Copyright Infringement
  • www.techrepublic.com: Have Some Spare Cash? You’ll Need it for OpenAI’s New API
  • bsky.app: Discussion of OpenAI's new o1-Pro API pricing and its implications for the AI community.
  • Maginative: OpenAI Unveils New Audio Models to Make AI Agents Sound More Human Than Ever
  • bsky.app: This blog post discusses OpenAI's new audio models, noting their promising features but also mentioning the issue of mixing instructions and data in the same token stream.
  • www.techrepublic.com: This article reports on OpenAI's new text-to-speech and speech-to-text tools based on GPT-4o, highlighting their capabilities and potential applications but also mentioning a possible similar path for video.
  • Analytics Vidhya: OpenAI's Audio Models: How to Access, Features, Applications, and More
  • MarkTechPost: OpenAI Introduced Advanced Audio Models ‘gpt-4o-mini-tts’, ‘gpt-4o-transcribe’, and ‘gpt-4o-mini-transcribe’: Enhancing Real-Time Speech Synthesis and Transcription Capabilities for Developers
  • Simon Willison's Weblog: OpenAI announced today, for both text-to-speech and speech-to-text. They're very promising new models, but they appear to suffer from the ever-present risk of accidental (or malicious) instruction following.
  • THE DECODER: OpenAI releases new AI voice models with customizable speaking styles
  • Composio: Finally, OpenAI gave in and launched a new agentic framework called Agents SDK.
  • Last Week in AI: Our 204th episode with a summary and discussion of last week's big AI news! Recorded on 03/21/2025 Hosted by and . Feel free to email us your questions and feedback at and/or  Read out our text newsletter and comment on the podcast at . https://discord.gg/nTyezGSKwP In this episode: Baidu launched two new multimodal models, Ernie 4.5 and Ernie X1, boasting competitive pricing and capabilities compared to Western counterparts like GPT-4.5 and DeepSeek R1. OpenAI introduced new audio models, including impressive speech-to-text and text-to-speech systems, and added O1 Pro to their developer API at high costs, reflecting efforts for more profitability. Nvidia and Apple announced significant hardware advancements, including Nvidia's future GPU plans and Apple's new Mac Studio offering that can run DeepSeek R1. DeepSeek employees are facing travel restrictions, suggesting China is treating its AI development with increased secrecy and urgency, emphasizing a wartime footing in AI competition.