News from the AI & ML world

DeeperML - #users

Michal Langmajer@Fello AI //
OpenAI has launched its latest AI model, GPT-4.5, described as the company's most advanced language model to date. This new model features substantial enhancements over its predecessors, particularly in advanced reasoning, problem-solving, and contextual understanding. GPT-4.5 is designed to offer a more natural and engaging conversational experience, with improvements including superior capabilities in handling complex reasoning tasks, enhanced creativity, and the ability to manage intricate logic problems while maintaining nuanced conversations with improved contextual recall.

However, the launch of GPT-4.5 is facing challenges due to a shortage of GPUs, according to OpenAI CEO Sam Altman. This limitation is restricting access to the priciest tiers of ChatGPT Pro subscribers and developers initially. Altman stated that OpenAI has "run out of GPUs" due to growing demand, leading to a staggered rollout. The company plans to add tens of thousands of GPUs next week and expand access to Plus, Team, Enterprise, and Edu users in the following weeks.

Recommended read:
References :
  • AI News | VentureBeat: OpenAI has announced the release of GPT-4.5, which CEO Sam Altman previously said would be the last non-chain-of-thought (CoT) model. The company said the new model “is not a frontier modelâ€� but is still its biggest large language model (LLM), with more computational efficiency.
  • Analytics Vidhya: Since the beginning of 2025, we have been seeing the launch of one amazing model after another – from DeepSeek-R1 and o3-mini to Grok 3 and Claude 3.7 Sonnet. The latest addition to this ever-expanding list of advanced AI models is the much-awaited OpenAI GPT-4.5. This new model in the GPT series brings “Vibe Checkâ€� ...
  • Fello AI: OpenAI’s GPT‑4.5 Finally Arrived: Can It Beat Grok 3 and Claude 3.7?
  • Shelly Palmer: GPT-4.5: The Last LLM
  • www.tomshardware.com: Sam Altman says that OpenAI has to stagger the release of GPT-4.5 due to GPU shortages.
  • Techstrong.ai: OpenAI’s GPT‑4.5 AI Is Ready for ‘Natural Conversation’
  • eWEEK: OpenAI Releases GPT-4.5, a “Warmâ€� Generative AI Model, for Paid Plans and APIs
  • Gradient Flow: Scaling Up, Costs Up: GPT-4.5 and the Intensifying AI Competition
  • THE DECODER: OpenAI has presented its largest language model to date. According to Mark Chen, Chief Research Officer at OpenAI, GPT 4.5 shows that the scaling of AI models has not yet reached its limits.
  • PCMag Middle East ai: OpenAI unveiled a new AI model today, GPT-4.5, but the launch did not go as planned. The company ran out of GPUs, or computing power, ahead of the reveal, CEO Sam Altman.
  • Analytics Vidhya: OpenAI has introduced GPT-4.5, an advanced AI model designed to enhance chatbot interactions with improved natural language processing, a richer knowledge base, and better contextual understanding.
  • THE DECODER: OpenAI has released GPT-4.5 as a "Research Preview". The new language model is intended to be more natural and less hallucinatory, but is significantly more expensive than its predecessors.
  • AI News | VentureBeat: OpenAI has announced the release of GPT-4.5, a research preview of its latest and most powerful large language model (LLM) for chat applications
  • Analytics Vidhya: Two days ago, on 27 Feb 2025, OpenAI dropped GPT-4.5, expectations were sky-high. But instead of a groundbreaking leap forward, we got a model prioritizing emotional intelligence over raw reasoning power.
  • AI News | VentureBeat: GPT-4.5 for enterprise: Do its accuracy and knowledge justify the cost?
  • TechCrunch: OpenAI launches GPT-4.5, its largest model to date
  • Windows Report: OpenAI released GPT-4.5, but it’s not much of an upgrade from GPT-4o
  • Analytics Vidhya: The article discusses GPT-4.5 becoming #1 on the Chatbot Arena.
  • Data Phoenix: This article discusses OpenAI's release of GPT-4.5, its strengths and weaknesses.
  • iHLS: This article covers the launch of GPT-4.5 and its capabilities.
  • THE DECODER: The article discusses OpenAI's GPT-4.5 release and its performance compared to previous versions.
  • Towards AI: TAT #142: GPT-4.5 Released — But Can It Stack Up Against Reasoning Models?
  • pub.towardsai.net: TAT #142: GPT-4.5 Released — But Can It Stack Up Against Reasoning Models?
  • LessWrong: On GPT-4.5
  • Analytics Vidhya: Top 5 Generative AI Breakthroughs of February 2025: GPT-4.5, Grok-3, and More!
  • Towards AI: GPT-4.5: The Next Evolution in AI

Dr. Hura@Digital Information World //
OpenAI has released exciting updates for ChatGPT's Advanced Voice Mode, aimed at creating more natural and engaging user interactions. The primary focus of these updates is to reduce interruptions during conversations, a common issue where the AI would interject during pauses, hindering the flow of natural dialogue. This improvement allows users to take short breaths or think without the AI prematurely responding.

The Advanced Voice Mode is now available to all ChatGPT users with paid plans. Those with the free version of the chatbot will get access to the latest Advanced Voice Mode that enables users to pause without getting interrupted or when they want to speak to the AI assistant. The system requirements include Android app version 1.2024.206 or later, and for iOS, app version 1.2024.206 or later with iOS 16.4 or later.

In addition to minimizing interruptions, the update introduces a more personable tone to ChatGPT's voice interactions. The AI is designed to be more specific, direct, creative, and engaging in its replies, making conversations feel less robotic and more human-like. These changes come amid competition from other companies launching similar AI voice assistants, such as Sesame's new tool, Maya and Miles.

Recommended read:
References :
  • Digital Information World: This will make the AI assistant more personable and interrupt users so much less.
  • gHacks Technology News: OpenAI Updates ChatGPT Voice Mode for More Natural and Engaging Interactions
  • THE DECODER: OpenAI brings native image generation to ChatGPT
  • AI News | VentureBeat: As AI-generated images become more precise and accessible, GPT-4o represents a significant step forward in the space.
  • www.tomsguide.com: OpenAI just unveiled new ChatGPT image generator powered by Sora — here's what you can do now
  • www.zdnet.com: ChatGPT finally gets a much better image generator - how to try it for free
  • How-To Geek: ChatGPT Can Finally Generate Images With Legible Text
  • www.techradar.com: OpenAI unveiled image generation for 4o – here's everything you need to know about the ChatGPT upgrade
  • Simon Willison: OpenAI's new multi-modal image output, added to GPT-4o and ChatGPT this morning, finally gave me the selfie with a bear I've always wanted
  • Analytics Vidhya: A few days ago, Gemini rolled out its image generation feature in the 2.0 Flash version, and the internet erupted with stunning examples. Now, OpenAI is stepping up to the plate, raising the bar even higher by introducing native image generation (powered by GPT-4o) in ChatGPT.
  • www.techrepublic.com: As of March, any account holder can create images using GPT-4o in ChatGPT for free. See how to make ChatGPT work for your business.
  • SiliconANGLE: OpenAI upgrades ChatGPT’s image generation capabilities
  • TestingCatalog: OpenAI Brings Advanced Image Generation to GPT-4o in ChatGPT and Sora
  • thezvi.wordpress.com: Fun With GPT-4o Image Generation
  • Simon Willison's Weblog: Introducing 4o Image Generation
  • The Tech Basic: OpenAI’s ChatGPT Now Generates Highly Detailed Images With GPT-4o
  • gHacks Technology News: ChatGPT integrates GPT-4o for more realistic and detailed image creation
  • futurism.com: OpenAI is rolling out brand new image generation capabilities today for ChatGPT. And guess what? It finally, almost, nails text.
  • www.tomsguide.com: Here's what happened when I tested ChatGPT-4o image generator as well as what I like and don't like about this model.
  • thezvi.substack.com: Fun With GPT-4o Image Generation
  • PCMag Middle East ai: OpenAI has added AI image generation capabilities to ChatGPT. Users can now select the prompt, provide prompts, and get desired images within the regular ChatGPT window.
  • www.tomsguide.com: OpenAI is rolling out a series of upgrades to ChatGPT's Advanced Voice Mode this week, and they could make a big difference to your time with the chatbot.

Matthias Bastian@THE DECODER //
Google is enhancing its Gemini AI assistant with the ability to access users' Google Search history to deliver more personalized and relevant responses. This opt-in feature allows Gemini to analyze a user's search patterns and incorporate that information into its responses. The update is powered by the experimental Gemini 2.0 Flash Thinking model, which the company launched in late 2024.

This new capability, known as personalization, requires explicit user permission. Google is emphasizing transparency by allowing users to turn the feature on or off at any time, and Gemini will clearly indicate which data sources inform its personalized answers. To test the new feature Google suggests users ask about vacation spots, YouTube content ideas, or potential new hobbies. The system then draws on individual search histories to make tailored suggestions.

Recommended read:
References :
  • Android Faithful: Google's AI tool Gemini gets a boost by working with deeper insight about you through personalization and app connections.
  • Google DeepMind Blog: Experiment with Gemini 2.0 Flash native image generation
  • THE DECODER: Google adds native image generation to Gemini language models
  • THE DECODER: Google's Gemini AI assistant can now tap into users' search histories to provide more personalized responses, marking a significant expansion of the chatbot's capabilities.
  • TestingCatalog: Discover the latest updates to Google's Gemini app, featuring the new 2.0 Flash Thinking model, enhanced personalization, and deeper integration with Google apps.
  • The Official Google Blog: Gemini gets personal, with tailored help from your Google apps
  • Search Engine Journal: Google Search History Can Now Power Gemini AI Answers
  • www.zdnet.com: Gemini might soon have access to your Google Search history - if you let it
  • The Official Google Blog: The Assistant experience on mobile is upgrading to Gemini
  • www.zdnet.com: Google launches Gemini with Personalization, beating Apple to personal AI
  • Maginative: Google to Replace Google Assistant with Gemini on Android Phones
  • www.tomsguide.com: Google is giving away Gemini's best paid features for free — here's the tools you can try now
  • MacSparky: This article reports on Google's integration of Gemini AI into its search engine and discusses the implications for users and creators.
  • Search Engine Land: This change will roll out to most devices except Android 9 or earlier (and some other devices).
  • www.zdnet.com: Gemini's new features are now available for free, extending beyond its previous paid subscriber model.
  • www.techradar.com: Discusses how Google is giving Gemini a superpower by allowing it to access your Search history, raising excitement and concerns.
  • PCMag Middle East ai: This article discusses Google's plan to replace Google Assistant with Gemini AI, highlighting the timeline for the transition and requirements for the devices.
  • The Tech Basic: This article announces Google’s plan to replace Google Assistant with Gemini, focusing on the company’s focus on advancing AI and integrating Gemini into its mobile product ecosystem.
  • Verdaily: Google Announces New Update for its AI Wizard, Gemini: Improves User Experience
  • Windows Copilot News: Google is prepping Gemini to take action inside of apps
  • www.techradar.com: Worried about DeepSeek? Well, Google Gemini collects even more of your personal data
  • Maginative: Gemini App Gets a Major Upgrade: Canvas Mode, Audio Overviews, and More
  • TestingCatalog: Google launches Canvas and Audio Overview for all Gemini users
  • Android Faithful: Google Gemini Gets A Powerful Collaborative Upgrade: Canvas and Audio Overviews Now Available

Matthias Bastian@THE DECODER //
References: TestingCatalog , THE DECODER , AiThority ...
Baidu has released its advanced AI models, ERNIE 4.5 and ERNIE X1, making them freely available to users through the ERNIE Bot platform. This move is a direct challenge to AI giants like OpenAI, Google, and DeepSeek, aiming to provide broader access to cutting-edge AI technology. By offering these models for free, Baidu seeks to accelerate user engagement and gather real-world data to refine their AI capabilities, potentially shifting the balance in the AI landscape by making sophisticated AI tools a new standard rather than a luxury.

ERNIE 4.5 is a multimodal foundation model capable of integrating and understanding text, images, audio, and video. It enhances language understanding, reasoning, generation, and memory, and can even interpret internet memes and satirical cartoons. ERNIE X1, on the other hand, is designed for reasoning-intensive tasks, excelling in logical inference, problem-solving, and structured decision-making. Baidu claims ERNIE 4.5 can match the performance of models like DeepSeek R1 at half the deployment cost, while ERNIE X1 has demonstrated strong capabilities in areas like Chinese knowledge Q&A and complex calculations.

Recommended read:
References :
  • TestingCatalog: This article discusses Baidu's ERNIE 4.5 and ERNIE X1 models, highlighting their performance and lower prices compared to DeepSeek.
  • THE DECODER: This article discusses Baidu’s new LLMs, ERNIE 4.5 and ERNIE X1, highlighting their competitive pricing and plans for open-source release in the context of the AI market.
  • Analytics Vidhya: This article discusses Baidu’s release of ERNIE 4.5 and ERNIE X1 LLMs, highlighting their claimed performance advantages over GPT-4.5 and cost-effectiveness.
  • AiThority: With the launch of ERNIE 4.5 and ERNIE X1, ERNIE Bot is made free to the public ahead of schedule, and users can access both models free of charge. As a deep-thinking reasoning model with multimodal capabilities, ERNIE X1 delivers performance on par with DeepSeek R1 at only half the price. ERNIE 4.5 is the [...]
  • techxplore.com: Chinese internet search giant Baidu released a new artificial intelligence reasoning model Sunday and made its AI chatbot services free to consumers as ferocious competition grips the sector.
  • Analytics India Magazine: China’s Baidu Launches Two New AI Models, Rivals DeepSeek R1 at Half the Price
  • TechCrunch: Chinese search engine Baidu has launched two new AI models — Ernie 4.5, the latest version of the company’s foundational model first released two years ago, as well as a new reasoning model, Ernie X1. According to Reuters, Baidu claims that Ernie X1’s performance is “on par with DeepSeek R1 at only
  • AI News: Baidu undercuts rival AI models with ERNIE 4.5 and ERNIE X1
  • AIwire: With the launch of ERNIE 4.5 and ERNIE X1, ERNIE Bot is made free to the public ahead of schedule, and users can access both models free of charge.
  • techstrong.ai: Baidu Unleashes Speedy New AI Model to Rival DeepSeek
  • AI News | VentureBeat: Baidu has also announced plans to integrate ERNIE 4.5 and ERNIE X1 into its broader ecosystem, including Baidu Search and the Wenxiaoyan app.
  • www.tomshardware.com: ERNIE 4.5 AI model by Baidu claims to match DeepSeek R1 at half the cost
  • Fello AI: Baidu’s New ERNIE 4.5 & X1 – A Free AI That Is Better Than GPT-4.5 & Costs Pennies!

Ryan Daws@AI News //
References: On my Om , Shelly Palmer , bsky.app ...
Anthropic has announced that its AI assistant Claude can now search the web. This enhancement allows Claude to provide users with more up-to-date and relevant responses by expanding its knowledge base beyond its initial training data. It may seem like a minor feature update, but it's not. It is available to paid Claude 3.7 Sonnet users by toggling on "web search" in their profile settings.

This integration emphasizes transparency, as Claude provides direct citations when incorporating information from the web, enabling users to easily fact-check sources. Claude aims to streamline the information-gathering process by processing and delivering relevant sources in a conversational format. Anthropic believes this update will unlock new use cases for Claude across various industries, including sales, finance, research, and shopping.

Recommended read:
References :
  • On my Om: You can now use Claude to search the internet to provide more up-to-date and relevant responses. With web search, Claude has access to the latest events and information, boosting its accuracy on tasks that benefit from the most recent data.
  • Shelly Palmer: Most heavy LLM users will tell you that ChatGPT is the GOAT, but they prefer Claude for writing. Why wasn't Claude the GOAT?
  • AI News: Anthropic has announced its AI assistant Claude can now search the web, providing users with more up-to-date and relevant responses.
  • bsky.app: Simon Willison's notes on the new web search feature for Claude
  • venturebeat.com: VentureBeat article on Anthropic giving Claude real-time web search
  • Analytics Vidhya: Claude AI Now Supports Web Search ğŸŒ
  • Maginative: Anthropic Finally Adds Search Capabilities to Its AI Assistant
  • bsky.app: Anthropic ships a new "web search" feature for their Claude consumer apps today, here are my notes - it's frustrating that they don't share details on whether the underlying index is their own or run by a partner
  • Ken Yeung: Intercom is doubling down on AI-driven customer support with a significant expansion of its Fin agent.
  • THE DECODER: Anthropic's new 'think tool' lets Claude take notes to solve complex problems
  • www.producthunt.com: The "think" tool from Claude
  • www.techradar.com: The ultimate AI search face-off - I pitted Claude's new search tool against ChatGPT Search, Perplexity, and Gemini, the results might surprise you
  • www.tomsguide.com: Claude 3.7 Sonnet now supports real-time web searching — but there's a catch

Chris McKay@Maginative //
OpenAI has recently unveiled new audio models based on GPT-4o, significantly enhancing its text-to-speech and speech-to-text capabilities. These new tools are intended to give AI agents a voice, enabling a range of applications, with demonstrations including the ability for an AI to read emails in character. The announcement includes the introduction of new transcription models, specifically gpt-4o-transcribe and gpt-4o-mini-transcribe, which are designed to outperform the existing Whisper model.

The text-to-speech and speech-to-text tools are based on GPT-4o. While these models show promise, some experts have noted potential vulnerabilities. Like other large language model (LLM)-driven multi-modal models, they appear susceptible to prompt-injection-adjacent issues, stemming from the mixing of instructions and data within the same token stream. OpenAI hinted it may take a similar path with video.

Recommended read:
References :
  • AI News | VentureBeat: OpenAI’s new voice AI model gpt-4o-transcribe lets you add speech to your existing text apps in seconds
  • Analytics Vidhya: OpenAI’s Audio Models: How to Access, Features, Applications, and More
  • Maginative: OpenAI Unveils New Audio Models to Make AI Agents Sound More Human Than Ever
  • bsky.app: I published some notes on OpenAI's new text-to-speech and speech-to-text models.
  • Samrat Man Singh: OpenAI announced some new audio models yesterday, including new transcription models( gpt-4o-transcribe and gpt-4o-mini-transcribe ).
  • www.techrepublic.com: The text-to-speech and speech-to-text tools are all based on GPT-4o. OpenAI hinted it may take a similar path with video.
  • MarkTechPost: Reports on OpenAI introducing advanced audio models.
  • Simon Willison's Weblog: OpenAI announced today, for both text-to-speech and speech-to-text. They're very promising new models, but they appear to suffer from the ever-present risk of accidental (or malicious) instruction following.
  • THE DECODER: OpenAI has released a new generation of audio models that let developers customize how their AI assistants speak.
  • venturebeat.com: DeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAI
  • Last Week in AI: #204 - OpenAI Audio, Rubin GPUs, MCP, Zochi

@www.eweek.com //
OpenAI's ChatGPT has achieved a major milestone, now boasting over 400 million weekly active users. This growth reflects the platform's increasing adoption by both individual consumers and enterprise clients. The number of paying enterprise users has doubled since September, reaching 2 million. This surge is attributed to users recommending ChatGPT to their workplaces, with companies recognizing its potential as a time-saving tool.

OpenAI is also expanding the availability of its "Operator" AI agent, granting access to ChatGPT Pro subscribers in countries like Australia, Brazil, Canada, India, Japan, South Korea, and the United Kingdom. This AI agent can independently operate a web browser and perform tasks for users. Additionally, OpenAI has taken action against malicious activities by removing accounts suspected of being used for surveillance and opinion-influence operations, particularly those originating from China and North Korea. These measures underscore OpenAI's commitment to broadening access while safeguarding against misuse.

Recommended read:
References :
  • siliconangle.com: OpenAI reaches 400M weekly active users, doubles enterprise customer base
  • the-decoder.com: OpenAI rolls out Operator to more countries
  • www.eweek.com: OpenAI Hits 400M Weekly Active Users for ChatGPT: “Like Cloud Servicesâ€�
  • SiliconANGLE: OpenAI reaches 400M weekly active users, doubles enterprise customer base
  • Maginative: OpenAI’s Growth Shows No Signs of Slowing, Now Serving 400 Million Users
  • Analytics India Magazine: ChatGPT Surpasses 400 Million Weekly Active Users, OpenAI Plans GPT-5 Launch
  • the-decoder.com: ChatGPT reaches more than 400 million weekly users, according to OpenAI COO Brad Lightcap. The article appeared first on .
  • eWEEK: OpenAI Hits 400M Weekly Active Users for ChatGPT: “Like Cloud Services”
  • analyticsindiamag.com: ChatGPT Surpasses 400 Million Weekly Active Users, OpenAI Plans GPT-5 Launch
  • venturebeat.com: OpenAI’s ChatGPT explodes to 400M weekly users, with GPT-5 on the way
  • techcrunch.com: OpenAI rolls out its AI agent, Operator, in several countries
  • techxmedia.com: OpenAI Hits 400M Weekly Users, Doubles Paid Subscriptions
  • Maginative: AI in Higher Ed: 7 Major Takeaways from OpenAI’s ChatGPT Student Usage Report
  • Dataconomy: 400 million users later OpenAI still has a major problem
  • The Tech Portal: OpenAI expands its ‘Operator’ AI agent in several countries, including India
  • Dataconomy: OpenAI announced on Friday the rollout of Operator, its AI agent designed to perform tasks on behalf of users, for ChatGPT Pro subscribers in Australia, Brazil, Canada, India, Japan, Singapore, South Korea, the U.K., and additional countries.
  • Shelly Palmer: OpenAI has expanded access to its Deep Research tool – originally exclusive to users with a $200 Pro subscription – to Plus, Team, Edu, and Enterprise plan subscribers.
  • AIwire: OpenAI announced it has expanded access to its Deep Research, its new AI agent designed to conduct complex research.

@laptopmag.com //
Apple is reportedly developing future Apple Watch models with integrated cameras to enhance its artificial intelligence capabilities. According to reports, Apple aims to add small cameras to both the standard Apple Watch and the Ultra version, potentially positioning them on the front screen or where the side button is located. The goal is to provide the watch with visual perception of its environment, making it smarter and more useful for users.

Adding cameras to the Apple Watch would enable new features, similar to object and text scanning currently available on iPhones. For example, the watch could identify objects, translate foreign text, or provide information about food packaging for allergy-conscious individuals. The camera system could also have applications in health monitoring, such as examining skin tone to determine sleep quality or measuring heart rate through wrist readings.

Apple has been exploring various camera designs for the Apple Watch for years, with patents showcasing different approaches. These include hiding a camera in the watch band for health tracking, a pop-up camera for occasional use, and even integrating a camera into the Digital Crown or using a flip-up screen. Despite these advancements, Apple is shuffling its executive team, replacing the head of AI with the VP of Vision Pro, suggesting a strategic shift in its AI efforts, with internal sources labeling delays with new AI rollouts as "ugly."

Recommended read:
References :
  • The Tech Basic: Apple Is Working on Adding Cameras to the Apple Watch
  • www.techrepublic.com: Apple’s Next Big Thing is AI on Smart Watches
  • www.laptopmag.com: Apple Watch's next big AI enhancement might be a built-in camera
  • Analytics India Magazine: Apple is also working on adding cameras to the AirPods.
  • THE DECODER: Apple is working on bringing visual AI capabilities to its wearable devices, according to Bloomberg. The article appeared first on .
  • PCMag Middle East ai: After missteps with Apple Intelligence, a new plan is starting to take shape and it's all about visuals—not chatbots. has had a , but Cupertino's AI plans are more ambitious than a revamped Siri.
  • gHacks Technology News: Apple is reportedly developing future Apple Watch models equipped with integrated cameras to enable advanced artificial intelligence (AI) features, aiming for a release by 2027.
  • www.laptopmag.com: Future Apple AirPods may detect health issues using AI

Vasu Jakkal@Microsoft Security Blog //
Microsoft and Google are enhancing their AI security measures and capabilities to address the evolving landscape of artificial intelligence. Microsoft is prioritizing secure AI integration, focusing on data protection and ensuring AI systems are transparent and compliant. Microsoft Secure, an online event, showcased AI innovations for the security lifecycle, providing tools for smarter, faster, and stronger security. The event highlighted how to secure data used by AI, AI apps, and AI cloud workloads, as well as how to protect AI investments from cyberthreats through data security and compliance tools.

NASA is leveraging Microsoft AI capabilities through its new Earth Copilot to democratize access to complex Earth Science data. This empowers scientists to discover patterns and gain insights from the vast amount of data collected by NASA's satellites, which can inform policy decisions and support various industries. Meanwhile, Google is developing Gemini to enable it to take actions within apps, potentially transforming AI assistants into more useful tools. Google is also releasing the GEMMA-3 models to improve the capabilities of the Gemini project.

Recommended read:
References :

S.Dyema Zandria@The Tech Basic //
Google is enhancing its Gemini AI with a new feature that allows users to create AI podcasts from research materials. This new capability, called Audio Overviews, converts research and study materials into engaging, podcast-style discussions featuring AI hosts. This aims to make learning and information consumption more accessible and enjoyable, particularly for educational purposes.

The Audio Overviews feature leverages Gemini's Deep Research capabilities. Users can input a topic, have Gemini generate a detailed report, and then convert that report into a conversational podcast with AI hosts. These hosts discuss the information in an approachable manner, similar to two friends exploring a topic. This tool is available to both free and paid Gemini Advanced users.

Recommended read:
References :
  • The Tech Basic: Google created a system that enhances educational experiences by making study-related tasks more interesting. Gemini is an AI tool that converts dull projects and assignments into exciting podcast recordings.
  • www.techrepublic.com: Google boosts Gemini with Canvas and Audio Overview, offering real-time editing and podcast-style audio insights to power creative projects.
  • The Verge: Google will let you make AI podcasts from Gemini’s Deep Research. That means you can turn the in-depth reports generated by Gemini into a conversational podcast featuring two AI “hosts.
  • Windows Copilot News: Google launched Gemini 2.0, its new AI model for practically everything
  • Google Workspace Updates: Provides a recap of Google Workspace Updates for the week of March 21, 2025, highlighting AI-powered features.
  • Stuff South Africa: New Gemini update allows the AI assistant to see through your screen and camera

@felloai.com //
Apple is facing mounting criticism over its AI implementation, with reports of its "Apple Intelligence" features causing significant issues. The AI powered summaries are distorting headlines, fabricating news, misreporting deaths and incorrectly stating events. There have been several instances of these inaccuracies with a man being falsely reported to have died, a darts player being announced as a winner before a competition, and a tennis star being inaccurately outed as gay. These errors, have sparked concerns about trust and reliability in the tech company's AI, leading some to suggest that Apple is facing an AI trust crisis.

Apple's AI push is also being scrutinized for how it handles messages. The AI is rewriting scam messages to appear legitimate and marking them as priority messages, further complicating the issue. This is compounded by notification summaries that misinterpret data, resulting in inaccurate information being delivered to users. Apple’s attempts to mitigate the problems with a UI tweak to clarify when summaries are AI-generated are seen as insufficient, with critics arguing that such measures will not stop the spread of misinformation, and are not doing enough to address the core problems of its flawed AI.

Recommended read:
References :
  • felloai.com: Apple Intelligence Has Fake News Problem – How To Turn Off Auto AI Summaries?
  • www.theguardian.com: Apple says it will update AI feature after inaccurate news alerts
  • www.them.us: Apple AI Push Notification Falsely States Tennis Star Rafael Nadal Came Out as Gay
  • www.digitimes.com: Apple faces setbacks in AI push; missteps spark BBC criticism
  • Fello AI: Apple Intelligence Has Fake News Problem – How To Turn Off Auto AI Summaries?
  • it-online.co.za: Apple disables AI feature
  • the-decoder.com: Apple pauses AI-generated news summaries after serious accuracy issues surface
  • www.theguardian.com: Apple suspends AI-generated news alert service after BBC complaint
  • PCMag Middle East ai: Apple Pauses AI Notification Summaries That Spread Fake News

@www.gadgets360.com //
OpenAI’s ChatGPT, API, and Sora services experienced a major outage, causing high error rates and inaccessibility for users globally. This disruption affected various functionalities, including text generation, API integrations, and the Sora text-to-video platform. The root cause was identified as an issue with an upstream provider, and OpenAI worked to restore services. This outage highlights the challenges and dependencies in AI infrastructure.

Recommended read:
References :
  • www.techmeme.com: OpenAI says ChatGPT is mostly recovered and APIs and Sora fully operational, after an outage led to the services "experiencing high error rates" for a few hours (Emma Roth/The Verge)
  • siliconangle.com: Outage takes ChatGPT, Sora and OpenAI’s APIs offline for many users
  • www.macrumors.com: ChatGPT Experiencing Outage
  • Search Engine Journal: Major Outage Hits OpenAI ChatGPT
  • The Verge: OpenAI says ChatGPT is mostly recovered and APIs and Sora fully operational, after an outage led to the services "experiencing high error rates" for a few hours (Emma Roth/The Verge)
  • Antonio Pequen?o IV: OpenAI’s ChatGPT Is Down—Here’s What We Know

Evelyn Blake@The Tech Basic //
Google has begun rolling out real-time interaction features to its AI assistant, Gemini, enabling live video and screen sharing. These enhancements, powered by Project Astra, allow users to engage more intuitively with their devices, marking a significant advancement in AI-assisted technology. These features are available to Google One AI Premium subscribers.

The new live video feature allows users to utilize their smartphone cameras to engage in real-time visual interactions with Gemini, enabling the AI to answer questions about what it observes. Gemini can analyze a user’s phone screen or camera feed in real-time and instantly answer questions. The screen-sharing feature enables the AI to analyze and provide insights on the displayed content, useful for navigating complex applications or troubleshooting issues. Google plans to expand access to more users soon.

Recommended read:
References :
  • The Tech Basic: Google has started releasing new AI tools for Gemini that let the assistant analyze your phone screen or camera feed in real time.
  • gHacks Technology News: Google has begun rolling out new features for its AI assistant, Gemini, enabling real-time interaction through live video and screen sharing.
  • The Verge: Google is rolling out Gemini’s real-time AI video features

Evelyn Blake@The Tech Basic //
Google has started rolling out new AI tools for Gemini, allowing the assistant to analyze your phone screen or camera feed in real time. These features are powered by Project Astra and are available to Google One AI Premium subscribers. The update transforms Gemini into a visual helper, enabling users to point their camera at an object and receive descriptions or suggestions from the AI.

These features are part of Google's Project Astra initiative, which aims to enhance AI's ability to understand and interact with the real world in real-time. Gemini can now analyze your screen in real-time through a "Share screen with Live" button and analyze your phone's camera. Early adopters have tested the screen-reading tool, and Google plans to expand access to more users soon. With Gemini's live video and screen sharing functionalities, Google is positioning itself ahead in the competitive landscape of AI assistants.

Recommended read:
References :
  • The Tech Basic: Google has started releasing new AI tools for Gemini that let the assistant analyze your phone screen or camera feed in real time.
  • gHacks Technology News: Google rolls out Project Astra-powered features in Gemini AI
  • www.techradar.com: Gemini can now see your screen and judge your tabs

@blogs.microsoft.com //
References: IEEE Spectrum , IEEE Spectrum ,
Anthropic, Google DeepMind, and OpenAI are at the forefront of developing AI agents with the ability to interact with computers in a human-like manner. These agents are designed to perform a range of tasks, including web searches, form completion, and button clicks, enabling them to order groceries, request rides, or book flights. The models employ chain-of-thought reasoning to decompose complex instructions into manageable steps, requesting user input when necessary and seeking confirmation before executing final actions.

To address safety concerns such as prompt injection attacks, developers are implementing restrictions, such as preventing the agents from logging into websites or entering payment information. Anthropic was the first to unveil this functionality in October, with its Claude chatbot now capable of "using computers the way humans do." Google DeepMind is developing Mariner, built on top of Google’s Gemini 2 language model and OpenAI launched its computer-use agent (CUA), called Operator.

Recommended read:
References :
  • IEEE Spectrum: IEEE Spectrum discusses the development of AI agents that can use computers like humans, highlighting models from Anthropic, Google DeepMind, and OpenAI.
  • IEEE Spectrum: Article discussing OpenAI's computer-use agent, called Operator, and its ability to work with websites.
  • www.anthropic.com: Anthropic was the first to unveil this new functionality, with an announcement in October that its Claude chatbot can now “use computers the way humans do.â€�

Merin Susan@Analytics India Magazine //
OpenAI is facing internal and external scrutiny regarding the ethical implications of its AI technologies. Employees have voiced concerns about a potential military deal with the startup Anduril, fearing damage to OpenAI's reputation due to its association with a weapons manufacturer. One employee noted that the company seemed to be downplaying the implications of working with a weapons manufacturer. Another employee said that they were concerned the deal would hurt OpenAI’s reputation.

OpenAI's technologies, specifically ChatGPT, have also come under scrutiny regarding their potential impact on mental health. Research indicates that specific types of ChatGPT usage, particularly "personal conversations" involving emotional expression, may be linked to increased loneliness among users. A study found that users who were more prone to emotional attachment were more likely to report increased loneliness in response to frequent personal conversations with the chatbot. Interestingly, the research also highlights that most people use ChatGPT for practical purposes rather than seeking emotional support.

Recommended read:
References :
  • Windows Copilot News: OpenAI employees question the ethics of military deal with startup Anduril
  • Analytics India Magazine: ChatGPT Use Linked to Increased Loneliness, Finds OpenAI Study.
  • AIwire: Since its launch in November 2022, OpenAI’s ChatGPT has become the most widely used AI chatbot globally. Its rapid adoption places it alongside major search engines and social media platforms, establishing it as a key player in the digital landscape. According to some estimates, ChatGPT usage has surpassed 400 million weekly active users.
  • Platformer: New research from OpenAI shows that heavy chatbot usage is correlated with loneliness and reduced socialization. Will AI companies learn from social networks' mistakes?

@singularityhub.com //
OpenAI models, including the recently released GPT-4o, are facing scrutiny due to their vulnerability to "jailbreaks." Researchers have demonstrated that targeted attacks can bypass the safety measures implemented in these models, raising concerns about their potential misuse. These jailbreaks involve manipulating the models through techniques like "fine-tuning," where models are retrained to produce responses with malicious intent, effectively creating an "evil twin" capable of harmful tasks. This highlights the ongoing need for further development and robust safety measures within AI systems.

The discovery of these vulnerabilities poses significant risks for applications relying on the safe behavior of OpenAI's models. The concern is that, as AI capabilities advance, the potential for harm may outpace the ability to prevent it. This risk is particularly urgent as open-weight models, once released, cannot be recalled, underscoring the need to collectively define an acceptable risk threshold and take action before that threshold is crossed. A bad actor could disable safeguards and create the “evil twin” of a model: equally capable, but with no ethical or legal bounds.

Recommended read:
References :
  • www.artificialintelligence-news.com: Recent research has highlighted potential vulnerabilities in OpenAI models, demonstrating that their safety measures can be bypassed by targeted attacks. These findings underline the ongoing need for further development in AI safety systems.
  • www.datasciencecentral.com: OpenAI models, although advanced, are not completely secure from manipulation and potential misuse. Researchers have discovered vulnerabilities that can be exploited to retrain models for malicious purposes, highlighting the importance of ongoing research in AI safety.
  • Blog (Main): OpenAI models have been found vulnerable to manipulation through "jailbreaks," prompting concerns about their safety and potential misuse in malicious activities. This poses a significant risk for applications relying on the models’ safe behavior.
  • SingularityHub: This article discusses Anthropic's new system for defending against AI jailbreaks and its successful resistance to hacking attempts.

@www.reuters.com //
Meta is expanding its artificial intelligence research into the realm of humanoid robotics, aiming to develop AI-driven software and sensors. This initiative focuses on creating intelligent machines that can interact with the physical world, potentially powering consumer robots. The company's efforts are concentrated on "embodied AI," which combines intelligence with real-world interactions, enabling robots to move, sense, and make decisions in three-dimensional environments.

Meta is not initially planning to release its own branded robots. Instead, the company is concentrating on developing AI-powered software and sensor technology that can be utilized by other robotics manufacturers. This strategy positions Meta alongside tech giants like Tesla, Apple, and Google, all of which are also investing in the robotics sector. Meta is also prioritizing user data protection by using source code analysis to detect and prevent unauthorized data scraping across its platforms, including Facebook, Instagram, and Reality Labs.

Recommended read:
References :

@www.ghacks.net //
Recent security analyses have revealed that the iOS version of DeepSeek, a widely-used AI chatbot developed by a Chinese company, transmits user data unencrypted to servers controlled by ByteDance. This practice exposes users to potential data interception and raises significant privacy concerns. The unencrypted data includes sensitive information such as organization identifiers, software development kit versions, operating system versions, and user-selected languages. Apple's App Transport Security (ATS), designed to enforce secure data transmission, has been globally disabled in the DeepSeek app, further compromising user data security.

Security experts from NowSecure recommend that organizations remove the DeepSeek iOS app from managed and personal devices to mitigate privacy and security risks, noting that the Android version of the app exhibits even less secure behavior. Several U.S. lawmakers are advocating for a ban on the DeepSeek app on government devices, citing concerns over potential data sharing with the Chinese government. This mirrors previous actions against other Chinese-developed apps due to national security considerations. New York State has already banned government employees from using the DeepSeek AI app amid these concerns.

Recommended read:
References :
  • cset.georgetown.edu: China’s ability to launch DeepSeek’s popular chatbot draws US government panel’s scrutiny
  • PCMag Middle East ai: House Bill Proposes Ban on Using DeepSeek on Government-Issued Devices
  • Information Security Buzz: Recent security analyses have found that the iOS version of DeepSeek transmits user data unencrypted.
  • www.ghacks.net: Security analyses revealed unencrypted data transmission by DeepSeek's iOS app.
  • iHLS: Article about New York State banning the DeepSeek AI app.

Matthias Bastian@THE DECODER //
Google is launching a new experimental feature called 'Daily Listen' that uses AI to create personalized podcast-style summaries within its Discover feed. This feature curates roughly five-minute audio episodes based on a user's search history and news preferences. The system pulls together information on topics and stories the user follows, delivering it within the Google app on Android and iOS devices. Daily Listen is currently available in the US for users who have opted into Search Labs experiments, with a limited number of spots being available. The personalized podcast appears below the search bar, with a date and "Made for you" label.

Once enabled, the AI-generated episode will appear the following day in the Discover feed, providing a brief audio overview of the user’s favorite topics. The podcast includes a full-screen player with audio playback controls, a text transcript, and related stories corresponding to each section. The technology is similar to Google’s NotebookLM audio overviews, but while NotebookLM is based on a user prompt, Daily Listen examines topics and stories the user already follows, with a focus on delivering helpful, although fallible, information.

Recommended read:
References :
  • THE DECODER: Google's AI-powered 'Daily Listen' feature creates personalized podcast summaries in Discover feed
  • PCMag Middle East ai: Google AI Feature Turns Your Search History, News Preferences Into a Podcast
  • 9to5google.com: Google AI Feature Turns Your Search History, News Preferences Into a Podcast
  • Gadgets 360: Google Releases ‘Daily Listen’ AI-Generated, Personalised Podcasts via Discover Feed: Report
  • OODAloop: Google can turn your Discover feed into an AI-generated podcast

Chris McKay@Maginative //
Google is enhancing its NotebookLM tool with interactive mind maps, a feature designed to help users visualize and navigate complex information from uploaded sources. These mind maps present document topics as branching diagrams, allowing users to explore connections and ask questions about specific areas by clicking on nodes. This visual approach aims to transform how users interact with their content, moving beyond linear reading to a more intuitive exploration of interconnected concepts.

LlamaIndex, a framework for building knowledge-driven AI agents, has also been integrated with Google Cloud's Gen AI Toolbox for Databases. This integration empowers developers to construct sophisticated AI agents with customizable workflows. LlamaIndex offers pre-built agent architectures for common use cases, along with tools to tailor the behavior of AI agents to specific requirements, which will benefit those using Gen AI Toolbox for Databases.

Recommended read:
References :
  • AI & Machine Learning: Gen AI Toolbox for Databases announces LlamaIndex integration
  • Maginative: Google Adds Interactive Mind Maps to NotebookLM
  • www.techradar.com: Google’s NotebookLM adds Mind Maps to its string of research tools to help you learn faster than ever

@timesofindia.indiatimes.com //
Recent developments highlight both the expanding influence and the regulatory hurdles faced by the AI company DeepSeek. In South Korea, the government has halted downloads of DeepSeek's applications, citing concerns over data privacy. This action has removed the company's apps from both the Apple and Google mobile app marketplaces, though their website remains accessible.

Simultaneously, DeepSeek's AI technology is rapidly integrating into China's transportation sector, extending from electric vehicles (EVs) to e-scooters. Major automakers, including BYD, Geely, and Chery Automobile, are incorporating DeepSeek's AI into their vehicles, offering features like preliminary self-driving capabilities. E-scooter brands like Segway-Ninebot and Niu Technologies are also integrating DeepSeek for enhanced features such as AI-powered content creation, data analytics, and driver assistance systems, reflecting what some industry observers are calling "DeepSeek fever" due to its cost-effective AI integration.

Perplexity has released "1776," a modified version of DeepSeek-R1. This model addresses the original version's limitations by mitigating censorship on sensitive topics, particularly those related to Chinese history and geopolitics. The modifications were made using post-training techniques to ensure more open and contextually accurate responses, making the modified model available on Perplexity's Sonar AI platform and GitHub.

Recommended read:
References :

@shellypalmer.com //
OpenAI has recently launched a new beta feature in ChatGPT called 'Tasks'. This functionality allows users to schedule future actions and reminders, expanding ChatGPT's capabilities beyond basic conversations. Users can now set up both one-time and recurring tasks. These tasks could include daily summaries of news or sports scores, scheduled searches, or even conditional alerts based on specific criteria. The implementation uses the iCal VEVENT format for scheduling and is integrated through a tool in the system prompt.

The Tasks feature can be accessed through the "4o with scheduled tasks" model option on the ChatGPT website. Tasks can be managed through chat threads or a dedicated Tasks section, which simplifies modification and cancellation. Notifications are delivered across web, desktop, and mobile platforms upon task completion. Although it's still in beta, this feature aims to make ChatGPT a more versatile AI assistant, capable of automating daily tasks and workflows. Some users have reported early beta bugs with push notifications and date formats.

Recommended read:
References :

@www.verdict.co.uk //
OpenAI is shifting its strategy by integrating its o3 technology, rather than releasing it as a standalone AI model. CEO Sam Altman announced this change, stating that GPT-5 will be a comprehensive system incorporating o3, aiming to simplify OpenAI's product offerings. This decision follows the testing of advanced reasoning models, o3 and o3 mini, which were designed to tackle more complex tasks.

Altman emphasized the desire to make AI "just work" for users, acknowledging the complexity of the current model selection process. He expressed dissatisfaction with the 'model picker' feature and aims to return to "magic unified intelligence". The company plans to unify its AI models, eliminating the need for users to manually select which GPT model to use.

This integration strategy also includes the upcoming release of GPT-4.5, which Altman describes as their last non-chain-of-thought model. A key goal is to create AI systems capable of using all available tools and adapting their reasoning time based on the task at hand. While GPT-5 will be accessible on the free tier of ChatGPT with standard intelligence, paid subscriptions will offer a higher level of intelligence incorporating voice, search, and deep research capabilities.

Recommended read:
References :
  • www.verdict.co.uk: The Microsoft-backed AI company plans not to release o3 as an independent AI model.
  • sherwood.news: This article discusses OpenAI's 50 rules for AI model responses, emphasizing the loosening of restrictions and potential influence from the anti-DEI movement.
  • thezvi.substack.com: This article explores the controversial decision by OpenAI to loosen restrictions on its AI models.
  • thezvi.wordpress.com: This article details three recent events involving OpenAI, including the release of its 50 rules and the potential impact of the anti-DEI movement.
  • www.artificialintelligence-news.com: This blog post critically examines OpenAI's new AI model response rules.