Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

Containerize legacy Spring Boot application using Amazon Q Developer CLI and MCP server

Light has two identities that are impossible to see at the same time

New vision model from Cohere runs on two GPUs, beats top-tier VLMs on visual tasks

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • OpenAI (GPT-4 / GPT-4o)
    • Anthropic (Claude 3)
    • Google DeepMind (Gemini)
    • Meta (LLaMA)
    • Cohere (Command R)
    • Amazon (Titan)
    • IBM (Watsonx)
    • Inflection AI (Pi)
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • AI Experts
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • The TechLead
    • Matt Wolfe AI
    • Andrew Ng
    • OpenAI
    • Expert Blogs
      • François Chollet
      • Gary Marcus
      • IBM
      • Jack Clark
      • Jeremy Howard
      • Melanie Mitchell
      • Andrew Ng
      • Andrej Karpathy
      • Sebastian Ruder
      • Rachel Thomas
      • IBM
  • AI Tools
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
  • AI Policy
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
  • Industry AI
    • Finance AI
    • Healthcare AI
    • Education AI
    • Energy AI
    • Legal AI
LinkedIn Instagram YouTube Threads X (Twitter)
Advanced AI News
IBM

Salesforce, IBM and Snowflake push zero copy as AI’s enterprise foundation. Here’s why it matters

By Advanced AI EditorAugust 1, 2025No Comments8 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


(pixabay)

The progression from AI experimentation to production-grade enterprise deployment has consistently stumbled on a fundamental challenge  – data architecture. While organizations invest heavily in models and use cases, the underlying data infrastructure often becomes the limiting factor that prevents AI initiatives from scaling beyond proof-of-concept stages.

An industry roundtable featuring technical leaders from IBM, Snowflake, and Salesforce highlighted how zero copy architecture is emerging as a critical enabler for enterprise AI transformation. The discussion revealed that traditional approaches to data unification – characterized by extensive copying, movement, and reconstruction of datasets – are fundamentally incompatible with the requirements of agentic AI systems.

The data afterthought problem

Edward Calvesbert, Vice President of Product Management for IBM’s watsonx platform, identified a pattern that resonates across enterprise AI implementations:

The biggest challenge for gen AI initiatives and progressing them from experimentation to production is really having data as an afterthought. So much of the attention on the initial focus is on the models, of course, on the use cases, which is critical as well. But data is somewhat of an afterthought to that conversation.

This points to a deeper architectural problem. Many AI projects begin with retrieval-augmented generation (RAG) systems built on small, curated knowledge bases – technical documentation or internal policies that are universally accessible within an organization. The simplicity of these initial implementations masks the complexity that emerges when projects attempt to incorporate additional content types, confidential information, and enterprise-scale datasets. Calvesbert explained:

When you start getting into additional types of content, additional types of questions and answers, potentially confidential information within the organization, the data management problem just gets a lot more complicated.

The challenge intensifies when similarity searching, which works adequately for limited document sets, encounters the scale requirements of millions of documents.

Zero copy architecture represents a fundamental shift from data gravity – where analytics must move to where data resides – to data fluidity, where data can be accessed in place without physical movement or duplication. Saptarshi Mukherjee, Director of Product Management at Snowflake, outlined why this architectural approach has become essential for agentic AI systems:

The foundational shift that is taking place with agentic AI is when human beings looked at data and took decision for the business, there was a limited cognitive space and capacity in an enterprise. With agentic AI, agents are able to perform workflows 24/7, looking at data much more expansively.

This expanded data access creates new security and governance challenges. Unlike human analysts who operate within cognitive limitations, AI agents can potentially access vast enterprise datasets continuously. Zero copy architecture addresses this by ensuring that “source systems are on point to make sure there is governance and security in place, there is RBAC models and access control that these agents have to adhere to.”

The enterprise AI challenge extends beyond mere data access to encompass business context preservation and real-time responsiveness. Mukherjee emphasized that AI solutions require more than raw data:

They need to understand the business context. When you’re building a RAG application, when you’re building an AI solution that is very specific to an industry, you’re looking at both data and the metadata.

Re-constructing this metadata in systems designed purely for AI consumption creates bottlenecks that zero copy architecture aims to eliminate. The approach enables business context and metadata to propagate alongside data without requiring reconstruction in downstream systems.

Real-time data access has also taken on new significance since agentic AI has hurtled onto the scene. Traditional business intelligence operated on batch processing cycles where 12-hour data latency was acceptable for dashboard and reporting use cases. Agentic systems, however, must “react to enterprise events as they are taking place,” requiring sub-second response times for many workflows.

Governance in a zero copy world

The governance implications of zero copy architecture represent both an opportunity and a challenge for enterprise implementations. Narinder Singh, VP of Product Management at Salesforce, highlighted how zero copy preserves data ownership and control:

With zero copy, the benefits come with respect to the data ownership and control and managing the lifecycle of the data, and having always access to the latest data.

This preservation of source system governance eliminates the need to reconstruct security models in downstream analytics platforms. As Mukherjee explained through a practical example involving automobile service center optimization:

We had to preserve and propagate the ACLs for the RAG applications, to honor the governance and security model. What zero copy does is an AI agent in Snowflake, for example, is accessing data in Salesforce. We don’t have to reconstruct that security and the governance model.

However, zero copy implementation must address the intersection of different governance paradigms. Calvesbert noted the complexity that arises when unstructured data governance – typically organized around document-level permissions – meets structured data policies designed to maximize access while protecting specific sensitive elements.

The practical implementation of zero copy architecture relies heavily on open standards and interoperability protocols. The roundtable participants emphasized the role of formats like Apache Iceberg and emerging communication protocols in enabling cross-platform data access without duplication.

Calvesbert positioned these developments within a broader architectural vision: “I really think it’s definitional Lake House architecture. I really think it communicates the value proposition of open table formats.” The benefits extend beyond technical implementation to address cost, compliance, and agility requirements that drive enterprise technology decisions.

The Model Context Protocol (MCP) emerged as a significant discussion point, though the panelists cautioned against viewing it as a complete solution. “If you look at how a lot of the data vendors are implementing MCP, it’s really just a different protocol for traditional kind of JDBC and SQL,” Calvesbert observed. The real value lies in semantic enrichment and use case-specific data aggregation that enables large language models to identify and utilize appropriate tools within defined workflow contexts.

Enterprise zero copy implementations face several common misconceptions that can derail projects. Singh identified a prevalent assumption that data unification alone solves AI readiness: 

Often, the common misconception I’ve seen is that we help solve all of the problem, make the data AI ready, and we should be able to get to the outcomes that we want to achieve.

This approach can create additional data silos rather than eliminating them, particularly when organizations pursue centralized data lake strategies without addressing fundamental issues around quality, consistency, and semantic harmonization. The result can be “more data chaos in terms of having multiple copies of data or having multiple ways in which agents are able to look at the data.”

Where to begin?

For organizations beginning their zero copy journey, the panelists emphasized the importance of establishing clear architectural principles before pursuing incremental implementations. Calvesbert stressed the value of articulating a clear destination:

The most important thing is to have a clear vision of the destination… a good time horizon, five plus years.

This architectural vision should encompass commodity cloud object storage, open data and table formats, shared metadata catalogs, and governance systems that can be centrally designed but locally enforced. These principles provide a framework for evaluating incremental decisions across different lines of business and projects.

Singh advocated for a use case-driven approach: 

Always start small and grow from there… starting with a specific use case, start with that use case around which you layer your data strategy.

This methodology allows organizations to build foundational capabilities while demonstrating tangible business value.

Zero copy architecture represents more than a technical optimization – it embodies a fundamental shift in how enterprises architect their data infrastructure for AI-driven operations. The technology aims to resolve the tension between data accessibility and governance that has constrained enterprise AI implementations.

As Mukherjee concluded: 

Unification, propagation of context. These are very foundational principles… You’re going to solve a problem that’s going to require these different data products to come together, to bring your AI solutions to life.

The success of zero copy implementations will ultimately depend on organizations’ ability to balance technical innovation with practical governance requirements, ensuring that agentic AI systems can access the data they need while operating within appropriate security and compliance boundaries.

My take

Zero copy architecture represents more than technical evolution – it embodies the maturation of enterprise AI from experimental curiosity to operational necessity. The roundtable discussion highlighted a critical disconnect, that while enterprises rush to deploy generative AI capabilities, many are discovering that their data infrastructure was designed for human-scale analysis, not 24/7 autonomous agents operating across enterprise datasets.

The governance challenge is particularly acute. Traditional data management assumes human oversight and cognitive limitations that naturally constrain access patterns. Agentic systems demolish these assumptions, creating new requirements for granular, context-aware access controls that can operate at machine speed while preserving business intent.

What’s most interesting about the zero copy approach is its recognition that data architecture and AI strategy are inseparable. Organizations that continue to treat data infrastructure as a secondary consideration will find themselves building impressive AI demonstrations that cannot scale beyond proof-of-concept stages. This roundtable indicates that data foundations need to be structured for agentic consumption from the outset, which could be a daunting prospect for companies that are reluctant to change the way they think about data.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleChevron (CVX) Q2 Earnings 2025
Next Article Exxon (XOM) Earnings Q2 2025
Advanced AI Editor
  • Website

Related Posts

Google, IBM make strides toward quantum computers that may revolutionize problem solving

August 1, 2025

Security updates: IBM Db2 can be attacked in various ways

August 1, 2025

Shadow AI and Poor Governance Linked to Costlier Breaches

August 1, 2025

Comments are closed.

Latest Posts

Artist Tyrrell Winston Sues New Orleans Pelicans Over Instagram Posts

Blum Staffers Speak On Closure, Spiegler Slams Art ‘Financialization’

Theatre Director and Artist Dies at 83

France to Accelerate Return of Looted Artworks—and More Art News

Latest Posts

Containerize legacy Spring Boot application using Amazon Q Developer CLI and MCP server

August 2, 2025

Light has two identities that are impossible to see at the same time

August 2, 2025

New vision model from Cohere runs on two GPUs, beats top-tier VLMs on visual tasks

August 1, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • Containerize legacy Spring Boot application using Amazon Q Developer CLI and MCP server
  • Light has two identities that are impossible to see at the same time
  • New vision model from Cohere runs on two GPUs, beats top-tier VLMs on visual tasks
  • Fundamental Research Labs nabs $30M+ to build AI agents across verticals
  • Solving years-old math problems with Gemini 2.5 Deep Think

Recent Comments

  1. NekofenKed on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  2. KexefHoalt on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  3. YovivMek on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  4. Wujuganery on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  5. TylerGlilm on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

LinkedIn Instagram YouTube Threads X (Twitter)
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.