News from the AI & ML world

DeeperML - #dataanalytics

Chris McKay@Maginative //
Snowflake is aggressively expanding its footprint in the cloud data platform market, moving beyond its traditional data warehousing focus to become a comprehensive AI platform. This strategic shift was highlighted at Snowflake Summit 2025, where the company showcased its vision of empowering business users with advanced AI capabilities for data exploration and analysis. A key element of this transformation is the recent acquisition of Crunchy Data, a move that brings enterprise-grade PostgreSQL capabilities into Snowflake’s AI Data Cloud. This acquisition is viewed as both a defensive and offensive maneuver in the competitive landscape of cloud-native data intelligence platforms.

The acquisition of Crunchy Data for a reported $250 million marks a significant step in Snowflake’s strategy to enable more complex data pipelines and enhance its AI-driven data workflows. Crunchy Data's expertise in PostgreSQL, a well-established open-source database, provides Snowflake with a FedRAMP-compliant, developer-friendly, and AI-ready database solution. Snowflake intends to provide enhanced scalability, operational governance, and performance tooling for its wider enterprise client base by incorporating Crunchy Data's technology. This strategy is meant to address the need for safe and scalable databases for mission-critical AI applications and also places Snowflake in closer competition with Databricks.

Furthermore, Snowflake introduced new AI-powered services at the Summit, including Snowflake Intelligence and Cortex AI, designed to make business data more accessible and actionable. Snowflake Intelligence enables users to query data in natural language and take actions based on the insights, while Cortex AISQL embeds AI operations directly into SQL. These initiatives, coupled with the integration of Crunchy Data’s PostgreSQL capabilities, indicate Snowflake's ambition to be the operating system for enterprise AI. By integrating such features, Snowflake is trying to transform from a simple data warehouse to a fully developed platform for AI-native apps and workflows, setting the stage for further expansion and innovation in the cloud data space.

Recommended read:
References :
  • futurumgroup.com: Is Snowflake’s Crunchy Data Acquisition a Game-Changer in the AI Data Platform Race?
  • BigDATAwire: Why Snowflake Bought Crunchy Data
  • Maginative: The Biggest Announcements from Snowflake Summit 2025
  • futurumgroup.com: Brad Shimmin, Vice President & Practice Lead, Data and Analytics at Futurum shares his/her insights on Snowflake’s acquisition of Crunchy.
  • www.bigdatawire.com: Monday brought the first surprise from Snowflake Summit 25: the acquisition of Crunchy Data for a reported $250 million.
  • futurumgroup.com: Snowflake Summit ’25: Accelerating AI with Unified Data & Compute
  • Maginative: Snowflake has announced its acquisition of Crunchy Data, a company specializing in enterprise-grade PostgreSQL solutions.
  • BigDATAwire: Agentic AI Spurs Data Stack Updates at Snowflake Summit

Chris McKay@Maginative //
Snowflake has announced the acquisition of Crunchy Data, a leading provider of enterprise-grade PostgreSQL solutions. This strategic move is designed to enhance Snowflake's AI Data Cloud by integrating robust PostgreSQL capabilities, making it easier for developers to build and deploy AI applications and agentic systems. The acquisition brings approximately 100 employees from Crunchy Data into Snowflake, signaling a significant expansion of Snowflake's capabilities in the database realm. This positions Snowflake to better compete with rivals like Databricks in the rapidly evolving AI infrastructure market, driven by the increasing demand for databases that can power AI agents.

This acquisition comes amidst a "PostgreSQL gold rush," as major platforms recognize the critical role of the data layer in feeding AI agents. Just weeks prior, Databricks acquired Neon, another Postgres startup, and other companies like Salesforce and ServiceNow have also made acquisitions in the data management space. Snowflake's SVP of Engineering, Vivek Raghunathan, highlighted the massive $350 billion market opportunity, underscoring the trend where AI agents, rather than humans, are increasingly driving database usage. PostgreSQL's popularity among developers and its suitability for rapid, automated provisioning make it an ideal choice for AI agent demands.

Crunchy Data brings enterprise-grade operational database capabilities that complement Snowflake's existing strengths. While Snowflake has excelled in analytical workloads involving massive datasets, it has been comparatively weaker on the transactional side, where real-time data storage and retrieval are essential. Crunchy Data's expertise in enterprise and regulated markets, including federal agencies and financial institutions, aligns well with Snowflake's existing customer base. The integration of Crunchy Data's PostgreSQL capabilities will enable Snowflake to provide a more comprehensive solution for organizations looking to leverage AI in their operations.

Recommended read:
References :
  • Maginative: Snowflake has announced its acquisition of Crunchy Data, a company specializing in enterprise-grade PostgreSQL solutions.
  • www.infoworld.com: Snowflake acquires Crunchy Data for enterprise-grade PostgreSQL to counter Databricks’ Neon buy
  • BigDATAwire: Monday brought the first surprise from Snowflake Summit 25: the acquisition of Crunchy Data for a reported $250 million.
  • www.bigdatawire.com: Monday brought the first surprise from Snowflake Summit 25: the acquisition of Crunchy Data for a reported $250 million.

Berry Zwets@Techzine Global //
Snowflake has unveiled a significant expansion of its AI capabilities at its annual Snowflake Summit 2025, solidifying its transition from a data warehouse to a comprehensive AI platform. CEO Sridhar Ramaswamy emphasized that "Snowflake is where data does more," highlighting the company's commitment to providing users with advanced AI tools directly integrated into their workflows. The announcements showcase a broad range of features aimed at simplifying data analysis, enhancing data integration, and streamlining AI development for business users.

Snowflake Intelligence and Cortex AI are central to the company's new AI-driven approach. Snowflake Intelligence acts as an agentic experience that enables business users to query data using natural language and take actions based on the insights they receive. Cortex Agents, Snowflake’s orchestration layer, supports multistep reasoning across both structured and unstructured data. A key advantage is governance inheritance, which automatically applies Snowflake's existing access controls to AI operations, removing a significant barrier to enterprise AI adoption.

In addition to Snowflake Intelligence, Cortex AISQL allows analysts to process images, documents, and audio within their familiar SQL syntax using native functions. Snowflake is also addressing legacy data workloads with SnowConvert AI, a new tool designed to simplify the migration of data, data warehouses, BI reports, and code to its platform. This AI-powered suite includes a migration assistant, code verification, and data validation, aiming to reduce migration time by half and ensure seamless transitions to the Snowflake platform.

Recommended read:
References :
  • www.bigdatawire.com: Snowflake Widens Analytics and AI Reach at Summit 25
  • WhatIs: AI tools highlight latest swath of Snowflake capabilities
  • Techzine Global: Snowflake enters enterprise PostgreSQL market with acquisition of Crunchy Data
  • www.infoworld.com: Snowflake ( ) has introduced a new multi-modal data ingestion service — Openflow — designed to help enterprises solve challenges around data integration and engineering in the wake of demand for generative AI and agentic AI use cases.
  • MarkTechPost: San Francisco, CA – The data cloud landscape is buzzing as Snowflake, a heavyweight in data warehousing and analytics, today pulled the wraps off two potentially transformative AI solutions: Cortex AISQL and Snowflake Intelligence. Revealed at the prestigious Snowflake Summit, these innovations are engineered to fundamentally alter how organizations interact with and derive intelligence from
  • Maginative: Snowflake Acquires Crunchy Data to Bring PostgreSQL to Its AI Cloud
  • www.marktechpost.com: The data cloud landscape is buzzing as Snowflake, a heavyweight in data warehousing and analytics, today pulled the wraps off two potentially transformative AI solutions: Cortex AISQL and Snowflake Intelligence.
  • www.zdnet.com: Snowflake's new AI agents make it easier for businesses to make sense of their data
  • BigDATAwire: Snowflake Widens Analytics and AI Reach at Summit 25
  • Maginative: At Snowflake Summit 25, the cloud data giant unveiled its most ambitious product vision yet: transforming from a data warehouse into an AI platform that puts advanced capabilities directly into business users’ hands.
  • siliconangle.com: Snowflake up its AI game, Circle IPO blasts off, and Elon splits with Trump

Heng Chi@AI Accelerator Institute //
AI is revolutionizing data management and analytics across various platforms. Amazon Web Services (AWS) is facilitating the development of high-performance data pipelines for AI and Natural Language Processing (NLP) applications, utilizing services like Amazon S3, AWS Lambda, AWS Glue, and Amazon SageMaker. These pipelines are essential for ingesting, processing, and providing output for training, inference, and decision-making at a large scale, leveraging AWS's scalability, flexibility, and cost-efficiency. AWS's auto-scaling options, seamless integration with ML and NLP workflows, and pay-as-you-go pricing model make it a preferred choice for businesses of all sizes.

Microsoft is simplifying data visualization with its new AI-powered tool, Data Formulator. This open-source application, developed by Microsoft Research, uses Large Language Models (LLMs) to transform data into interesting charts and graphs, even for users without extensive data manipulation and visualization knowledge. Data Formulator differentiates itself with its intuitive user interface and hybrid interactions, bridging the gap between visualization ideas and their actual creation. By supplementing natural language inputs with drag-and-drop interactions, it allows users to express visualization intent, with the AI handling the complex transformations in the background.

Yandex has released Yambda, the world's largest publicly available event dataset, to accelerate recommender systems research and development. This dataset contains nearly 5 billion anonymized user interaction events from Yandex Music, offering a valuable resource for bridging the gap between academic research and industry-scale applications. Yambda addresses the scarcity of large, openly accessible datasets in the field of recommender systems, which has traditionally lagged behind other AI domains due to the sensitive nature and commercial value of behavioral data. Additionally, Dremio is collaborating with Confluent’s TableFlow to provide real-time analytics on Apache Iceberg data, enabling users to stream data from Kafka into queryable tables without manual pipelines, accelerating insights and reducing ETL complexity.

Recommended read:
References :
  • insideAI News: NVIDIA and AMD Devising Export Rules-Compliant Chips for China AI Market
  • futurumgroup.com: Can Dell and NVIDIA’s AI Factory 2.0 Solve Enterprise-Scale AI Infrastructure Gaps?
  • TechHQ: Dell to build Nvidia Vera Rubin supercomputer for US Energy Department
  • techhq.com: Dell to build Nvidia Vera Rubin supercomputer for US Energy Department
  • futurumgroup.com: Can Dell Challenge Public Cloud AI with Its Expanded AI Factory?
  • insidehpc.com: DOE Announces “Doudnaâ€� Dell-NVIDIA Supercomputer at NERSC
  • techxplore.com: US supercomputer named after Nobel laureate Jennifer Doudna to power AI and scientific research
  • AI Accelerator Institute: Building efficient data pipelines for AI and NLP applications in AWS
  • www.dremio.com: Using Dremio with Confluent’s TableFlow for Real-Time Apache Iceberg Analytics
  • www.marktechpost.com: Yandex Releases Yambda: The World’s Largest Event Dataset to Accelerate Recommender Systems

@cloud.google.com //
Google Cloud is enhancing its text-to-SQL capabilities using the Gemini AI model. This technology aims to improve the speed and accuracy of data access for organizations that rely on data-driven insights for decision-making. SQL, a core component of data access, is being revolutionized by Gemini's ability to generate SQL directly from natural language, also known as text-to-SQL. This advancement promises to boost productivity for developers and analysts while also empowering non-technical users to interact with data more easily.

Gemini's text-to-SQL capabilities are already integrated into several Google Cloud products, including BigQuery Studio, Cloud SQL Studio (supporting Postgres, MySQL, and SQL Server), AlloyDB Studio, and Cloud Spanner Studio. Users can find text-to-SQL features within the SQL Editor, SQL Generation tool, and the "Help me code" functionality. Additionally, AlloyDB AI offers a direct natural language interface to the database, currently available as a public preview. These integrations leverage Gemini models accessible through Vertex AI, providing a foundation for advanced text-to-SQL functionalities.

Current state-of-the-art LLMs like Gemini 2.5 possess reasoning skills that enable them to translate intricate natural language queries into functional SQL code, complete with joins, filters, and aggregations. However, challenges arise when applying this technology to real-world databases and user questions. To address these challenges, Google Cloud is developing methods to provide business-specific context, understand user intent, manage SQL dialect differences, and complement LLMs with additional techniques to offer accurate and certified answers. These methods include context building, table retrieval, LLM-as-a-judge techniques, and LLM prompting and post-processing, which will be explored further in future blog posts.

Recommended read:
References :
  • AI & Machine Learning: Organizations depend on fast and accurate data-driven insights to make decisions, and SQL is at the core of how they access that data.
  • www.tomsguide.com: Google's adding more accessibility features to Chrome and Android — and they're powered by Gemini

@www.bigdatawire.com //
References: BigDATAwire , Maginative
Google Cloud Next 2025 featured a series of announcements focused on enhancing data analytics capabilities within Google Cloud, particularly through advancements to BigQuery. These enhancements center around a vision for AI-native data analytics, aiming to make data work more conversational, contextual, and intelligent. Key innovations include the introduction of unified governance in BigQuery, AI-powered agents for data engineering and data science tasks, and the integration of Gemini, Google's flagship foundation model, to drive these intelligent capabilities. These developments are designed to simplify data management, improve data quality, and accelerate the generation of AI-driven insights for businesses.

The new intelligent unified governance in BigQuery is designed to help organizations discover, understand, and leverage their data assets more effectively. This includes a universal, AI-powered data catalog that natively integrates Dataplex, BigQuery sharing, security, and metastore capabilities. The unified governance brings together business, technical, and runtime metadata, providing end-to-end data-to-AI lineage, data profiling, insights, and secure sharing. A universal semantic search allows users to find the right data by asking questions in natural language. These advancements promise to transform governance from a burden into a powerful tool for data activation, simplifying data and AI management.

A significant aspect of these enhancements is the introduction of specialized AI agents within BigQuery and Looker. These agents are tailored to different user roles, such as data engineers and business analysts, assisting with tasks like building data pipelines, model development, and querying data in plain English. Powered by Gemini, these agents provide suggestions based on information collected through the new BigQuery Knowledge Engine, which understands schema relationships, business terms, and query history. These agents are designed to make more data available to more people without them doing more work, ultimately transforming how data professionals interact with data.

Recommended read:
References :
  • BigDATAwire: Google Cloud Cranks Up the Analytics at Next 2025
  • Maginative: Google’s Vision for AI-Native Data Analytics Comes Into Focus