Snowflake VP of AI Baris Gultekin on Bringing AI to Data, Agent Design, Text-2-SQL, RAG & More
By Cognitive Revolution "How AI Changes Everything"
Summary
Topics Covered
- Unstructured Data Unlocks 80-90% Enterprise Insights
- Reasoning Models Make Text-to-SQL Enterprise Reliable
- Small Models Win on Scale Over Frontier Cost
- Bring AI to Data Respects Enterprise Security
- Horizontal Apps Dominate Vertical AI Solutions
Full Transcript
Hello and welcome back to the Cognitive Revolution. Before we get started today,
Revolution. Before we get started today, a quick final reminder. If you dream of a career in AI safety research, the deadline to apply to Matt's summer 2026
program is January 18th. Listen to my recent episode with Matt's executive director, Ryan Kidd, for all of the reasons that you should consider applying and then get started at mathsprogram.org/tcr.
mathsprogram.org/tcr.
Today, my guest is Barish Golden, vice president of AI at Snowflake, the cloud-based data platform that now describes itself as the AI data cloud.
Beerish came to Snowflake along with Snowflake's current CEO Shrear Ramaswami as part of Neva, an AI powered web and personal knowledgebased search engine
that Snowflake acquired in May 2023.
Since then, he's been working at the intersection of frontier AI capabilities and hard enterprise realities, deploying these systems in environments where security, governance, and reliability
are strict requirements.
As you'll hear, Snowflake's core philosophy is to bring AI to the data rather than sending sensitive data out to model providers. And in this episode,
we unpack exactly what that looks like in practice. We cover a ton of ground,
in practice. We cover a ton of ground, including the massive ongoing unlock of unstructured data, which is making the 80 to 90% of enterprise information that
was previously trapped in PDFs and other documents queryable for the very first time. The current state of both
time. The current state of both texttosql and rag systems and why reasoning models have finally made natural language data analysis reliable
enough for business users. the
trade-offs between using frontier models versus smaller specialized models and when to use structured workflows versus letting models choose their own adventures. How data residency
adventures. How data residency requirements are shaping partnerships between cloud providers, model labs, and platforms like Snowflake. How AI coding assistants are changing the discipline
of product management by enabling rapid prototyping of working features. where
Beerish sees value acrewing in the AI stack and his prediction that horizontal applications will win out over narrower vertical solutions and finally why
Beerish takes the over on my timeline for autonomous drop-in knowledge workers and what he believes will have to happen first.
If you want to understand how large enterprise companies are deploying AI today and what's really working as they mature from the early experimentation phase to the ROI at scale phase, taking
all of the operational complexities and security and governance concerns into account. I think this conversation will
account. I think this conversation will be perfect for you. And with that, I hope you enjoy this deep dive into enterprise AI adoption and the future of
data intelligence with Barish Golden, vice president of AI at Snowflake.
Barish Golden, vice president of AI at Snowflake. Welcome to the Cognitive
Snowflake. Welcome to the Cognitive Revolution.
>> Thank you, Nathan. Thanks for having me.
>> I'm excited for this conversation.
There's going to be a lot to learn. I
think people, we have a very diverse audience. The number one profile is AI
audience. The number one profile is AI engineer. And within that profile,
engineer. And within that profile, people work at a lot of different kinds of organizations from solo entrepreneurs and consultants to startups to enterprises. So some people will
enterprises. So some people will certainly know Snowflake and will work at organizations that are customers of Snowflake. Others probably have heard of
Snowflake. Others probably have heard of it and don't really know too much of the backstory. So maybe for starters just
backstory. So maybe for starters just kind of give us the real quick Snowflake 101 and then I'd love to go into how AGI is Snowflake today. [clears throat]
>> Sure. So Snowflake is a data platform.
We call ourselves an AI data cloud. So
what that means is our customers bring a lot of their data onto Snowflake so that they can secure it, govern it and analyze large amounts of data for various insights, dashboards and the
like. And from an AI perspective,
like. And from an AI perspective, because there's a lot of gravity to data, our customers do not want to replicate data in multiple places.
Instead, they want to bring AI to run next to data. So that's a very high level overview. And AGI PIL is an
level overview. And AGI PIL is an interesting phrase for us. We're quite
practical. We serve large enterprises and the goal is to get to high quality AI agents that that create positive ROI for customers quickly. And that can
happen today and it is happening today.
Super excited about where things are.
>> I definitely want to come back to the bring AI to data strategy that you guys have in a few minutes. But to just double click a little bit on the before and after because obviously Snowflake
has been around for a while before certainly anything like the AIs that we have now were available. So what were people doing before with Snowflake and
what are the new AI use cases that have been unlocked over say the last I don't know when you would start the clock right do you start the clock at chat GPT or do we need it was that like not quite strong enough to actually make things
work but yeah there's a lot of different dimensions maybe let's start with before and after >> sure so I'll start with the before so a lot of our customers have been using Snowflake mostly for structured data
initially and and this is where they'll bring the data in and and then they'll run large scale analysis um either to um to have insights uh to power BI
dashboards for instance um or to do to to do various analytics to understand their business. The uh the value of of
their business. The uh the value of of the platform is to be able to you know bring data from all the different places to break the data silos uh so that you could run analysis across uh you know
large amounts of large amounts of data.
Um so what is happening with AI is um there's a big unlock of course of unstructured data. Um so and this this
unstructured data. Um so and this this plays out in two ways. One is um if you have you know thousands uh you know hundreds of thousands of documents for instance you can now extract structure
from these documents uh and then you can analyze them. If you have contracts for
analyze them. If you have contracts for instance and if you want to say hey what are the contracts that talk about this specific thing? how many of them do I
specific thing? how many of them do I have uh or uh you know what are the contracts that are expiring soon uh in this category. So being able to run
this category. So being able to run these analysis super easily is not very very feasible. Then there's of course uh
very feasible. Then there's of course uh being able to uh do a lot of again kind of large scale analytics work but with um with a lot of ease. So a lot of the
pipelines that used to be built around I'm going to go classify my data. I'll
I'll extract information from it is now very very simple to do. So uh the way uh we've done it is we've again brought AI to directly work with the uh engine uh
for analytics. So you can do things like
for analytics. So you can do things like you know classification extraction of of that data very very easily and increasingly of course uh there is a lot of interest in bringing natural language
interface to all a company's data so that you could just talk to your data you can democratize access to all that data uh for the organization. Could you
give a sense for sort of the balance of structured unstructured or then I guess you're also saying that like structured data is getting structured through the process of basically AI retro
annotation. My sense is that and I think
annotation. My sense is that and I think of oh gosh there's so many vector databases the exact name of the one the where the founder told me this is slipping my mind but one of the
interesting things that I've understood to be happening in general with business data is that structured data was kind of like the tip of the iceberg in many organizations where it was the most
usable kind but it was actually like a relatively small amount of the data. It
was Chroma, it was Anton from Chromad who said that like most of the data that was going into Chromma DB had never been in a database before at all. It was just lying around in various places. So, have
you seen like a sort of great unlocking of people like dumping more and more data into Snowflake because now they have ways to make it useful where it just previously wasn't even worth it.
Yeah, we're absolutely seeing this uh you know 80 to 90% of all data is unstructured data and and you know because there weren't a lot of easy ways to process this, it was not necessarily
seen as the most you know usable uh data and it is now very usable both from a kind of extract and then bring structure um to it perspective as well as just you
know talk to all of that data uh you know find the right information you know using you know vector DBs for instance and then and then build uh in agents and and chat experiences of it. Uh so we're
seeing this and and it is again playing out in both both ways both uh more and more data is getting structured um so that you could run analytics on it down
the road as well as you can just use all of this data in conjunction with the structured data that you have. So for
instance um you know if you want to build um you know anything let's say a wealth management agent so you still need to be able to look up what the you know stocks
are doing in a structured way but you also have all of the u equities research uh that that that's in PDFs that you'd like to be able to use. So being able to combine both structured and unstructured is incredibly important for real world
use cases.
>> Where would you say we are on text to SQL today? It's been a while since I've
SQL today? It's been a while since I've done a show on text to SQL. There have
been probably two episodes on this theme historically and I guess there's last I checked there was like a range of opinions where some people were like yeah it's just not
really there. Other people were is there
really there. Other people were is there but you have to do a lot of work to make sure that you have a good semantic understanding because a lot of the SQL databases that come in are like there's multiple columns and there's tribal
knowledge on teams that like we don't use that column anymore. or we haven't deleted it, but we don't use it anymore and it's superseded by this or there's all these little nuances that kind of live in people's heads. And so some people have said, "Oh, the models can't
do it." Others have said the models can
do it." Others have said the models can do it if they have enough of that kind of context. What does a process look
of context. What does a process look like today? And how good does it get if
like today? And how good does it get if a new customers I want to start to enable these talk to my data with a sort of texttosql kind of strategy? What does
that look like now?
>> Yeah, I mean you called right. It's been
traditionally very difficult for models to get text to SQL right and there are various reasons for it. It is first of all if you ask what's my revenue the answer is there's only one answer. So
the the margin of error is very low and and the expectations of quality is incredibly high and the reason it's been really difficult for these models is because a lot of you need a lot of
semantics to be able to figure out where to get the data from. So first of all what is the definition of revenue? What
is the definition of profit can change and then how is it modeled in the data side can can be tricky can change when we're talking about real world scenarios we're talking about thousands and thousands of tables that that have
hundreds of thousands of columns in them to be able to go and reason about so it's been traditionally very difficult what I'll say has happened in the last 6 months to a year is with the reasoning
models getting substantially better and increasingly being able to bring the semantics relatively easily onto the platform. We've had pretty substantial
platform. We've had pretty substantial gains in quality. So we now have our product to do this for instance is Snowflake intelligence and we're seeing tremendous demand for doing text of SQL
and the quality is at a place where you can now deploy them very broadly. So
very high quality and very useful because the structure data is quite useful >> and when you say deploy broadly you mean to users who are not data >> analysts. That's right. That's right.
>> analysts. That's right. That's right.
So, so for us for instance this product I I mentioned Snowflake intelligence is our agent platform and it's it's being used by business users and it is the fastest growing product that we have on
Snowflake because we're now making you know large amounts of data easily accessible to business users to ask uh you know questions and get insight very quickly. In the past they would have to
quickly. In the past they would have to go to an analyst uh who's familiar with the data to for them to kind of build some analysis and then get back to them a week later. Now they can just directly
ask questions.
>> What does that process of bringing semantics onto the platform look like? I
can imagine >> a sort of big setup one time where you go out and interview the people that have set these things up and and should know. I can imagine at runtime you might
know. I can imagine at runtime you might have to come back with questions and say, "Hey, I've got multiple columns that are ambiguous here." I did one episode also with a company you may know called Illumx where they had a really
interesting strategy that was around basically building what they considered to be the sort of canonical like abstract ideal form of an enterprise in
each major vertical that they served.
And then they built their query engine off of that ideal. And then the mapping process was like okay now how does your actual real world enterprise deviate
from this ideal let's map all that out but then we know that once we've done that mapping the the logic on top of it will be trusted what mix of strategies are you using what you find to be effective
>> so so first of all again reasoning models are now at a point where AI can help substantially in building out a semantic model and uh the
inputs to that uh are like both of course you know the data that's um in the system all of the metadata that's in the system the names of the tables and columns as well as the data underneath
them u but um we have built a series of connectors to things like BI dashboards uh that also have a lot of semantics in them that that's super useful in
building out semantic understanding for the organization also things like you know queries people have been running in the past like these are all kind of hints uh for these uh agents to go help
our customers build semantic models. Um
Snowflake has recently also announced u what we're calling open semantic interchange uh which is uh a uh and and
trying to create an open standard for sharing that semantic model across um you know across the different platforms so that you know we can more easily create these common semantic
understandings for AI to act on.
>> Okay, that's interesting. How does that can you unpack that for me a little bit?
Like how does that work? What do I do as an enterprise if I want to adopt that standard?
>> Yeah. So, uh this is it's it's still early. Um so we're working with
early. Um so we're working with companies like Tableau um Omni uh you know with BBI platforms as well as uh as well as others technology providers to
create a um an exchange uh format so that if you create a semantic model in one platform you could just use it in in a different platform. Uh so there there's there's active development now
with u all the parties in the uh in in in in the open semantic interchange to define uh what that uh interface is so that we could support an open interchange of of the semantic model
essentially. So what that would look
essentially. So what that would look like is a customer can go to uh Snowflake for instance they can go you know build out their semantic model um and then they could reuse uh that
semantic model uh in in another place that supports the open interchange.
>> For one thing I wish that would come to electronic medical records sooner rather than later. I feel like yikes has been
than later. I feel like yikes has been my recent experience there. I wonder how that going back to the first question about how AGI pill various organizations
are. I wonder how that how do you see
are. I wonder how that how do you see that from a competitive dynamic standpoint because I think one of the things that's most interesting where you know the dice are in the air
so to speak it feels to me right now in the software market is like sure seems like the pace of software development is increasing dramatically.
people are I think mostly it's outliers or just plain bluster at this point to say that people are like deleting systems of record and rolling their own in house but you've at least [clears throat] got that talk is out there
>> and then presumably everybody is gez if I can >> maybe already see a major acceleration in my software development or if I can project a year or two in the future and I can see a clear path to a major and
I've vibe coded three AI apps for family members for Christmas presents this year and the acceleration at that level is certainly very real. Then it seems like everybody is going to be incentivized to
like try to go take sort of some conceptual territory from companies that maybe used to be partners, used to be compliments. It seems like it's headed
compliments. It seems like it's headed more toward competition. So all these companies that you named, right, historically you specialize in one thing. They did a little bit different
thing. They did a little bit different thing. They work nicely together. You
thing. They work nicely together. You
got a lot of customers in common. Great.
But if I'm them, or maybe if I'm you, like I might be I might start to worry at this point. Geez, if I'm Tableau, should I be afraid of Snowflake? Are
they going to come after me with something that sort of replaces what we do? And do I want to be partnering with
do? And do I want to be partnering with them on these standards or do I have to fear that whoever is the whoever has their hooks deepest into the customer can box out and colonize these
additional niches? Now, that's a pretty
additional niches? Now, that's a pretty AGI pled point of view. Maybe you think that's just we're where I'm getting ahead of myself there, but what do you think?
>> Not at all. Not at all. I I actually really love what's happening. And what's
happening is the uh the silos are coming down. There is this is all great for for
down. There is this is all great for for customers for consumers right with all these open standards essentially the beneficiaries are our customers there is no lock in anymore and then that's I
think that's great that's great for competition that's great for innovation that's great for customers so any and we're seeing this play out across the board right any anywhere in AI the
differentiation is coming down that means everyone is doing more and more things to create more and more value which ultimately is great for the industry and great for consumers and customers.
So I'm loving what's happening as you called out the the walls are coming down the lock in is is no longer there and that that makes product development really important that makes the speed of execution really important and
ultimately it's all about creating more and more value.
But does that take us to a place where companies that used to be friends are trending toward frenemies? Because it
seems like there's only so many idea like it just seems so obvious to me in so many places that like a big platform like Snowflake would be like sure we could do like what Tableau does especially now that we can get so much
more stuff shipped on a quarterly basis.
I would say the the pie is growing. So I
don't think it's a fixed pie that the people are trying to protect. So the
types of things you can do is growing and that is super exciting. I also don't think that everyone can do everything.
Ultimately where each company focuses is closer to their area of expertise to to their differentiation. So I don't
their differentiation. So I don't necessarily see that everyone's going to do everything but I also do believe there is a lot of competition but there's also growing pie which is which is exciting.
>> Yeah. Okay. Hey, we'll continue our interview in a moment after a word from our sponsors.
You're a developer who wants to innovate. Instead, you're stuck fixing
innovate. Instead, you're stuck fixing bottlenecks and fighting legacy code.
MongoDB can help. It's a flexible, unified platform that's built for developers by developers. MongoDB is
acid compliant, enterprise ready with the capabilities you need to ship AI apps fast. That's why so many of the
apps fast. That's why so many of the Fortune 500 trust MongoDB with their most critical workloads. Ready to think outside rows and columns? Start building
at mongodb.com/build.
That's mongodb.com/build.
Your IT team wastess half their day on repetitive tickets, password resets, access requests, onboarding, all pulling them away from meaningful work. With
Servil, you can cut help desk tickets by more than 50%. While legacy players are bolting AI onto decades old systems, Serville allows your IT team to describe
what they need in plain English and then writes automations in seconds. As
someone who does AI consulting for a number of different companies, I've seen firsthand how painful and costly manual provisioning can be. It often takes a week or more before I can start actual
work. If only the companies I work with
work. If only the companies I work with were using Servil, I'd be productive from day one. Serville powers the fastest growing companies in the world
like Perplexity, Vicata, Merkor, and Clay. And Serville guarantees 50% help
Clay. And Serville guarantees 50% help desk automation by week four of your free pilot. So get your team out of the
free pilot. So get your team out of the help desk and back to the work they enjoy. Book your free pilot at
enjoy. Book your free pilot at servil.com/cognitive.
servil.com/cognitive.
That's sv.com/cognitive.
Let's go back more toward the technical side for a minute. We kind of went debun text to SQL. Let's do the same thing for rag. So we've got all these unstructured
rag. So we've got all these unstructured vast amounts of data out there. Y
>> they're getting loaded into platforms. They're getting metadata synthetically created by AIS coming through and just processing them sweep by sweep.
>> How well is that working? And what is actually key to making it work? We've
been through eras of chunking strategy is really important or I've done episodes on graph databases and entity recognition and figuring out various ways to traverse the entity graph that
and obviously it's it's going to be quite distinct for each enterprise you with all the different entities that they're going to have that nobody else has. What is really driving
has. What is really driving results in that rag paradigm today?
>> Yeah, at Snowflake we've been actually very fortunate. We acquired um a company
very fortunate. We acquired um a company that that that I came with uh called Neva which was a uh webscale search engine and and that
>> all right awesome. So we brought that technology into snowflake uh to to build out our search and rag solutions and uh there the basically what determines
quality is um uh the the quality of the embedding model that you're using. Uh,
of course there is uh, you know, more and more sophisticated uh, you know, chunking strategies of of uh, you know, what you're indexing and then there is kind of other layers like the uh, the hybrid search and the re-ranker that you
build on top of it and so forth.
Increasingly a a core part of it is also to be able to understand uh, you know complex documents. You know PDFs are
complex documents. You know PDFs are messy. You have images, you have tables,
messy. You have images, you have tables, uh, you have you know multiple columns in a page and so forth. So being able to handle all of this, extract information really accurately, figuring out which model uh embedding model to use, you
know, whether you should use a multimodal embedding model or a text embedding model and so forth. So all of those are incredibly important, but increasingly, you know, we're getting to a point where you can, you know,
automate many of these things and then reduce the complexity so that a lot of the um a lot of what we used to, you know, require practitioners to do can be
relatively automated at this point. and
and now you're getting to a point where more interesting opportunities get unlocked. So for a company like
unlocked. So for a company like Snowflake for instance, being able to do what we're calling um analytical um agentic document analytics is something
that is that is possible to do. So what
I mean by that is let's say that you have kind of thousands of PDFs and there is information in it. Let's say you have uh you know quarterly results over the
last 10 years. So being able to say what's the average uh uh you know revenue over the last 10 years and if that is in multiple different documents being able to extract all of that and
then do analytics on it is now possible.
So overall I think rag is both uh getting increasingly you know higher in quality uh and also simpler to build and and increasingly more and more powerful
uh to handle some of the new agentic use cases. Would it be a fair
cases. Would it be a fair distillation of what you've said there that you're trending more toward more powerful models? Like a project
that I've been involved with recently is built around understanding often scanned on like a physical scanner forms
that are associated with the sale of a car from either a dealer to a person or person to person. These things of course have to get filed at the state and reviewed and they're super messy and whatever. So working a little bit with a
whatever. So working a little bit with a company that's using AI to automate that. In that context, I've really seen
that. In that context, I've really seen a pretty substantial simplification where 18 months ago it was like you might need your specialist embedding
model here and your kind of table extractor model there and all this kind of deep specialization often not super large models but like really dialed in on these use cases. And now I would say
today Claude 45 Opus or Gemini 3 mostly just solve the problem off the shelf in terms of understanding those documents at a
higher cost certainly inference wise but definitely a lot lower cost in terms of AI engineering time. Am I right to to say you're seeing the same trend less specialized models?
>> There are different use cases. If you
are going to process, you know, in some cases hundreds of millions of documents, you're not going to use cloud to do that. Instead, you want to use a
that. Instead, you want to use a specific embedding model to to embed certain aspects. You want to extract the
certain aspects. You want to extract the information so that you could reuse it later and so forth. But if you were talking about one or two documents, of course, these large language models can handle them really right now. So I still
do believe there are different use cases and then those use cases call for different different tactics, different models >> in in terms of why you wouldn't send millions of documents through cloud. Is
it just about inference cost or is there some other >> it's it's cost and throughput?
>> Yeah, exactly. So you know how long would it take for you to process that many documents is a is a challenge. So
we we you know at Snowflake we have a u document extraction model that that we've uh built and fine-tuned. It is in multiple orders of magnitude smaller than these large language models. That
means it's substantially cheaper and much faster to go and process. Uh if the task is specific I'm going to go extract information and and and extract these specific fields. it is faster and
specific fields. it is faster and cheaper to do versus kind of using these you know very large models which are super capable but again will be limited in terms of how fast they can do this
and of course the cost is another issue.
So that's interesting we're sketching out a little bit of a paro frontier so to speak here where we have >> on the like simplest
but most expensive at inference time potentially also like rate limit issues.
We have our clouds and other kind of frontier models. You're in the middle
frontier models. You're in the middle with a snowflake specialist model that is much smaller, does just what it does, but is still like something that is
advertised off over a whole bunch of enterprise customers that you have.
Is there what are you seeing in terms of the other end of that spectrum? the is
there still value in an individual enterprise trying to create its own super specialized model for some of these tasks or is that does that curve
stop at the snowflake scale model?
>> Yeah, it's a good question. So, first of all, we partner very closely with uh all the large language model uh labs out there uh and and they have incredible uh
capable models that that we use every day. Um
day. Um there are some cases where our customers would want something uh very specific and and this is this this is a case when
a customer has large amounts of data and uh the use case is something that the model has not seen before and then they have uh you know strict either throughput requirements or cost
requirements. Those are the cases where
requirements. Those are the cases where a a custom model that is usually you know based on some of the other large language models out there makes sense.
So we work with these customers to build custom models for them. Uh but in most cases a you know well tuned rag solution text to SQL solution with the data that
they already have with a large language model that's that's the uh you know frontier model is is is is usually you know the go-to scenario.
>> I I'm halfway through doing an AMA episode. One of the questions I got was
episode. One of the questions I got was is fine-tuning really dead? What do you think? So, it sounds like you're saying
think? So, it sounds like you're saying it's not quite dead, but it seems like it's on the decline in your >> I wouldn't say it's on the decline. I I
think it it is really well suited for certain types of things. And then maybe the best example is actually what uh you know, cursor recently did, right? You
know, at their scale, it does make a lot of sense for them to have a custom model that is doing their um you know, autocomplete for instance. Um so so being able to figure out in which
situations you need a custom model versus not is something that uh that is evolving. You know starting with the
evolving. You know starting with the large language models u makes a lot of sense. And then over time as you have
sense. And then over time as you have more and more data and if you have specific needs if you have uh um either specific needs because of data or
because of cost or throughput that's when you know specialized models come into picture.
You mentioned these partnerships you have with the Frontier companies. Before
getting to that, would you like to shout out or highlight any particular open source models that are your go-tos? We
hear a lot about obviously the Chinese ecosystem is like continuing to open source a lot more than the American ecosystem at this point. I don't know if you guys feel like comfortable using
Chinese models in your stack. I get very different answers on that when I ask that question. But what are the models
that question. But what are the models that you guys go to today when you're like, "Okay, we're gonna explore some new custom direction either for all of our customers or even just for one
customer." What are the handful of
customer." What are the handful of models that you go to as starting points to begin that journey? Yeah. So for
Snowflake, we have a platform where we offer a series of models and our customers choose which model they'd like to use. And then there are certain
to use. And then there are certain products where the model is just part of the product and not necessarily a a specific choice. For the models that we
specific choice. For the models that we offer there's of course all the frontier models, open AAI models, Anthropic, Gemini as well as models from Meta, Mistral and others. Some of these models
are open source, others are proprietary.
We also have deepseek as a model that we provide for customers. And then in certain arrangements where customers are looking to build custom models, some of
them are open to use using model weights from these models from China, others aren't, but it really depends on the customer.
>> Does that break down along like industry lines or is it more of just an idiocrat idiosyncratic like gut feel on the part of the customer as to what they're comfortable with?
>> It I think it's the latter actually. It
doesn't it's not necessarily a industry specific thing like we have customers who are in technology for instance who who will not who say yes uh sometimes or
who who will absolutely not touch uh you know some of these models for other customers.
>> Do you have a sense of how much they are leaving on the table?
Are they leaving much on the table by cutting off the Chinese model option?
like models like Quen are incredibly powerful and and then then you know if if they'd like to work start with models like that and then fine-tune it you can you can get very capable models but you
also have other alternatives um so it really depends on you know the internal policies of uh of these customers to decide which route to go I'd say it's
such a competitive space that uh I don't think there is you know one model that dominates um dominates it all whether that is in proprietary world or open source world. So there are a lot of
source world. So there are a lot of choices out there.
>> Gotcha. Okay.
>> Hey, we'll continue our interview in a moment after a word from our sponsors.
>> If you're listening to this podcast, you're probably thinking seriously about where AI is headed and maybe about how you can actually contribute to making it go well. I want to tell you about an
go well. I want to tell you about an opportunity that could become a pivot point in your career and a springboard for you to make a positive difference. a
program that I've been so impressed by that I've supported it with a personal donation. I'm talking about MATS, a
donation. I'm talking about MATS, a 12-week research program that connects talented researchers with top mentors working on AI alignment, interpretability security and
governance. These are researchers at
governance. These are researchers at Enthropic, Google Deep Mind, OpenAI, the AI Security Institute, Redwood Research, Meter, the AI Futures Project, Apollo
Research, Govai, Rand, and other leading organizations. The track record here is
organizations. The track record here is remarkable. Matts has accelerated over
remarkable. Matts has accelerated over 450 researchers with 80% of alumni now working in AI safety and security. 10%
have co-founded AI safety initiatives including Apollo Research whose co-founder and CEO made the 2025 Time 100 AI list. Matt's fellows have co-authored over 120 publications with
more than 7,000 citations and helped develop major research agendas like activation engineering, developmental interpretability, and evaluating situational awareness. The program is
situational awareness. The program is fully funded. A $15,000 stipen, $12,000
fully funded. A $15,000 stipen, $12,000 compute budget, housing, catered meals, travel, and office space in Berkeley or London. Everything you need to focus
London. Everything you need to focus entirely on research for 3 months with the chance to extend up to a year.
Applications open December 16th and close January 18th. If reducing risks from advanced AI is something you care about, you should apply. For more
information, check out mattsprogram.org/tcr.
mattsprogram.org/tcr.
That's matsprogram.org/tcr
or see the link in our show notes. The
worst thing about automation is how often it breaks. You build a structured workflow, carefully map every field from step to step, and it works in testing, but when real data hits or something
unexpected happens, the whole thing fails. What started as a timesaver is
fails. What started as a timesaver is now a fire you have to put out. Tasklet
is different. It's an AI agent that runs 24/7. Just describe what you want in
24/7. Just describe what you want in plain English. Send a daily briefing,
plain English. Send a daily briefing, triage support emails, or update your CRM. And whatever it is, Tasklid figures
CRM. And whatever it is, Tasklid figures out how to make it happen. Tasklid
connects to more than 3,000 business tools out of the box, plus any API or MCP server. It can even use a computer
MCP server. It can even use a computer to handle anything that can't be done programmatically.
Unlike Chat GPT, Tasklet actually does the work for you. And unlike traditional automation software, it just works. No
flowcharts, no tedious setup, no knowledge silos where only one person understands how it works. Listen to my full interview with Tasklet founder and CEO Andrew Lee. Try Tasklet for free at
tasklet.ai and use code cogrev to get 50% off your first month of any paid plan. That's code cogrevllet.ai.
plan. That's code cogrevllet.ai.
So on these partnerships, we've got announcements recently of partnerships with Anthropic and also with Google for the Gemini models. I believe there's also one although I think it was not so recently announced with OpenAI. Uh I
didn't see I didn't catch anything with respect to XAI and Grock providing all the latest and greatest stuff to customers is at the heart of that strategy. But tell me more about kind of
strategy. But tell me more about kind of some of the nuances of the partnership.
Is there a XAI relationship? If not, why not? And does it have anything to do
not? And does it have anything to do with them putting women in bikinis all over the place? And then I definitely want to get into the how are we bringing these models to data because that sort of is a bit of a narrative violation
relative to what you typically hear is we can't use that because we'd have to send the data to them and we're not comfortable with that. So I'm very interested in unpacking like how you are reversing that and bringing the models to the data on the snowflake platform.
>> Yeah, absolutely. Actually let me start there because that's incredibly important for us. So when we started the journey 2 and a half years ago or so, we heard loud and clear that our customers do not want to move their data out of
the snowflake security boundary.
Instead, AI needs to come next to data.
And that gives them a lot of advantages.
You can just respect all of the security that that you've established. You
respect a lot of the governance on the data so that you're not replicating this data. The attack vectors shrink in terms
data. The attack vectors shrink in terms of securing all of this information. So
what we have done is thanks to our relationships we've built we are bringing we're we're bringing essentially inference to run inside the snowflake security boundary so that that's accomplished through these
partnerships through the connections as well as a lot of the legal guarantees around the data. So essentially these models become subprocessors. There is no there is no state that's saved in any of
these models. So that that's super
these models. So that that's super helpful for our customers who who are very sensitive, many of them in regulated industries. So when they're
regulated industries. So when they're using any of these models, they know that the data still stays inside the snowflake security boundary.
>> So that does that mean then that the model weights have to come inside that boundary and like how is that happening?
Because obviously the IP >> Yeah, the IP still belongs to Yes, absolutely. And and the IP is is owned
absolutely. And and the IP is is owned by the model providers. uh the
inferences run uh by you know the cloud providers in in in their stack. Uh the
difference is uh you know we have series of guarantees to ensure um to ensure data residency to ensure uh there is there is no state that that's left. So
all of those are through uh you know through the relationships and the deals that we're we're doing with these model providers as well as the cloud providers.
So the cloud providers are key in this because they are certainly able to provide the inference and they're also providing the underlying like physical infrastructure that snowflake is built
on top of and so it's because both of those things are true that we can draw the right dotted line >> security boundary >> both of these things is there more that
I should understand about this because one thing that has of course I don't know what I don't know but it seems like the more you move weights around to different clouds and stuff, the more risk you as a frontier model developer
here. I'm thinking open aanthropic
here. I'm thinking open aanthropic Google obviously runs their own clouds to a very large extent although everything's showing up everywhere. We
one of the fascinating things about this whole moment has been how many alliances or at least partnerships we've seen between big tech companies that previously were very much at odds with each other. So with all the models
each other. So with all the models showing up everywhere, I'm like, how has it been that none of these have really leaked? seems like there's so many
leaked? seems like there's so many people that work at these platform companies that if access isn't like really well figured out. I don't know.
It just seems like something would leak at some point. But we haven't really seen that. We haven't seen the weights
seen that. We haven't seen the weights of a Frontier model leak at all as far as I know. And then people will speculate. Maybe some state actor might
speculate. Maybe some state actor might have stolen them and not put not told anybody about it. But we haven't seen fundamental breakdowns. So how should we
fundamental breakdowns. So how should we understand how that is happening?
to seemingly such a high degree. What
role does trusted execution environments play? What role do other kind of
play? What role do other kind of measures play? Like how are we my
measures play? Like how are we my general view working heristic is like every everybody's hacked, everybody's pone like nothing is secure and yet at the same time we seem to not be having catastrophic leaks. So how can you help
catastrophic leaks. So how can you help a simple person like me understand how we're achieving that?
Yeah, I mean u as you called out these are very sensitive important IPs that that belong to the model providers and then secured by the cloud providers.
They have a very strict series of requirements and setup to ensure that access is limited because they are the ones that are running the inference and
setting up the environment. They've set
it up in a way that is airtight. I don't
have a lot to say beyond it. I think the they absolutely take security very seriously. We work with them. We
seriously. We work with them. We
understand how important it is. You've
talked about all the different risks that that are out there that they need to protect against. So this is something that both cloud providers and model providers as far as I can see are taking very seriously. And as we work closely
very seriously. And as we work closely with them for Snowflake, of course, security is at the heart of what we do.
So we set up our own environment in a way that is that has all of the security considerations in mind. I'll just say this is an incredibly important area and there's a lot there's definitely focus in this area across across all the
parties.
>> How much would you say of this kind of security has gone to the level of provable guarantees or cryptographically
secured as opposed to more roles and access controls and things where there's still more fundamentally like human element. I just did an episode not long
element. I just did an episode not long ago with a couple experts in formal methods, including a guy who's a VP at Amazon who's pioneered a lot of their
use of formal methods to derive a lot of these security guarantees. But I'm not clear on how much of this is resting on that kind of like we have proven that
this is secure versus how much is like we have a process that we we feel good about and we want you to trust. So this
is not my area of expertise. So I don't have a lot of depth but purely from talking to both the cloud providers and the model providers. When you start looking at what are all of the attack
vectors and whether there is what is possible it doesn't seem like this is a human factor is an issue. The way the systems are set up is inherently very
secure. That that said in in security of
secure. That that said in in security of course you can never say hey this is completely airtight and it can never be penetrated. But uh yeah as I said
penetrated. But uh yeah as I said security is taken very seriously and I don't think it's a human factor necessarily. The way the systems are set
necessarily. The way the systems are set up is such that the execution environment doesn't have access to you cannot do a lot with it other than just run inference through it. So but by by
design access to the weights are limited.
>> Yeah. Okay. Cool. Thank you. I'm always
trying to get a little bit better read on that particular corner of the world and it's not one that is as freely and openly as some of us curious minds might like. Going back to the models though
like. Going back to the models though themselves, what's your read right now on this is another thing where I think people have very different intuitions.
Are the models going to be commoditized or are they going to be sufficiently differentiated as to maintain pricing power as we continue to go into the
future? I guess there's that. There's
future? I guess there's that. There's
like how do you help customers decide which model to use for a given case? Do
you have an eval platform built in or do you help them do eval? How do you help them think about being like keeping agile so they can switch?
Obviously, new models are coming out all the time. So there could be something
the time. So there could be something better, faster or cheaper that you can upgrade to, but you have to know with some confidence that you're going to be upgrading with good reason. There's a
whole ball of wax there. Take your time in melting it. Right.
>> I Yeah, super interesting. I mean I uh as you mentioned the differences between these models uh is not large you know each model keeps getting better and and
then there is you know great healthy competition out there between the model providers which again kind of benefits uh companies like ours our customers and
and so forth. for us because uh we're providing the choice to customers. Uh
the second part of the question is also really important which is how do we help our customers choose which model is the is is the best fit for their needs. Um
there there's a couple of considerations like one is um for many customers they do not want to leave again because of data residency requirements uh if they
are for instance uh an Amazon you know shop and today open AAI is available through Azure or through directly open AI like that becomes a consideration so
you know some customers are okay with their data leaving uh that Amazon cloud uh boundary others aren't so that that's one uh you know this decision point. The
second one is of course uh from a quality perspective uh you know many customers will go run evals side by side to decide which uh which model is best best suited for their needs. You know
what's interesting is you know some of these models are cheaper faster but when you add reasoning on top of it uh the equation changes right so you know certain models are very good at certain
things again just to call out you know cloud is incredibly good at coding and continues to be a great model for that.
So we help our customers uh in assessing in which models to use for their needs.
>> Does that extend to like how much of that is a service a consultancy type relationship and how much is productized at this point or do you know productization?
>> Yeah, it is more productized than a service on on our product. You can
easily choose models. You can easily do side-by-side comparisons. You can run
side-by-side comparisons. You can run evals and for many customers actually it's not necessarily for the first reason and not all the models are available in their environment anyway because as a company
they've decided that only these models are approved or only this environment is approved for me.
How often do you see people switching?
This is even in my so I started a company which I used to be the CEO of.
I'm no longer and we're only 40 people.
We're doing like $10 million a year in revenue. We're not an enterprise. We can
revenue. We're not an enterprise. We can
fly a little faster and looser than and we're also not in a regulated industry.
We basically do video content creation for local and increasingly like mid-size businesses.
>> And I feel like we should be changing models more often than we do. Honestly,
I feel like the leaprog effect is it's just happening so often. And if I were to grade our own performance, I would be like, eh, b like we're definitely better than most, but I wish we were even a
little more on top of eking out the the latest and greatest performance from the latest and greatest models, but it's hard. It is hard to resolve sometimes.
hard. It is hard to resolve sometimes.
You've got even just human interrator disagreement, which is tough to overcome. And how does that play out at
overcome. And how does that play out at the larger scale? Do you see people like, oh, we got a new cloud when cloud 4 6 hits? It's like how many people move to it in a week, in a month, in a quarter.
>> Yeah, I think it really depends on the use case. In most cases, we don't really
use case. In most cases, we don't really see a lot of uh switching happening. uh
because you know the prompts get optimized for a certain model and uh you get high quality because you've optimized it for a certain model and it's not as easy uh without further
optimization to switch and because the deltas between the models aren't much and uh they keep improving u you know on a regular basis the need for switching is also not that much um you know as as
you were describing this I was actually thinking about you know Google versus Bing you know at some point Bing you know got to a good enough quality uh but you know there was the habit of continuing to use uh you know Google the
familiarity of the interface and so forth that switching wasn't as necessary I don't think we're there yet necessarily for models there's still um a lot happening a lot of innovation happening and also for certain use cases
like if this is a one-off I'm going to go run something and then do sideby-side comparison like then you go pick the model that that works best for you but if you've already been investing in um
in an application [clears throat] and you have thousands of lines of um you systems and prompts that you've built, then there's a cost to switching in in
in which case the gains have to be large enough to justify that cost.
>> So, I'm no Ben Thompson, but it seems like from your perspective, you would want to commoditize your compliments and would want to do everything you could to reduce those switching costs, right?
Like one of the virtues of being on the Snowflake platform would be you've got all the things but not only ideally you have all the things but also you can the interface is presumably a lot more unified than it would be if you were
going directly to the model providers and I imagine there's a bunch of different things you could do over time.
You've got like things like DSP out there that you can say sure this is my one prompt with this model but maybe if I throw it into CS PI I can auto evolve my prompt to to be more optimized for some other model what have you. Is this
like a goal? Do you want to would you think of it as like a success metric that you like would help people be very fluid in switching from model to model as you go into the future?
>> I mean that's not how we think about it.
For us it it really does boil down to how do we bring the most value for customers quickly. Um so it choice is an
customers quickly. Um so it choice is an important uh factor there. So we'd like to offer a choice. Uh and and customers make uh model choices for a variety of reasons. As I said, you know, some of
reasons. As I said, you know, some of them have only approved this certain model. They have their own AI governance
model. They have their own AI governance boards where they decide which uh which model to use and so forth. Um but but for us, we start with the data at the core. So ultimately, anything that you
core. So ultimately, anything that you do is as good as the data that you provide to it. So a lot of the optimizations for us are can we do a phenomenal job at the uh retrieval layer
and then can we make sure that all of these models uh you know are optimized to the fullest extent so that any customer that's choosing one or the other for for the variety of reasons I
called um get the best uh quality data agent if you will uh that they're building with us.
>> Okay, that's really interesting. What is
what do you think that implies for the competitive dynamics between model providers? Like
one takeaway you might have from that is whoever has the best model at any given time like wins. Of course there are these other constraints but leaving those aside for the moment. If I'm a customer that has no binding constraints
and I can pick whatever frontier model I want. It seems like whoever has the best
want. It seems like whoever has the best model at any given moment in time wins that business and then actually stands to keep that business. Even if that business that might not be that whole enterprises business, but that
particular use case you're it's you're saying is stickier than you might think.
Switching costs are higher than they intuitively seem. And so
intuitively seem. And so having the best performance at the time it's initially evaluated is actually like pretty important. So I think the model quality is incredibly important
but increasingly we're moving up the stack so that the product also becomes incredibly important. So when you look
incredibly important. So when you look at from a consumer perspective chat GPT as a product starts having its own kind of stickiness because you start using it and you get accustomed to using it.
Similarly on on coding side cloud code has its own again benefits. You you'll
start writing your instructions your prompts to to optimize for that workflow. So I think we're just moving
workflow. So I think we're just moving up the stack. The model quality is absolutely central. But as the quality
absolutely central. But as the quality kind of keeps up across model providers, the next level of differentiation happens at the application layer.
>> So what that's a perfect transition to talk about agents and what you guys are doing with agents. The way I structure my own thinking about agents is on a spectrum from on the one end your cla
code style choose your own adventure. I
just give you the goal essentially and you the agent break it down and search around, GP around and figure out how to get there.
>> And then on the other extreme is like potentially a totally linear structured workflow where we're going to run a series of prompts one after another.
>> The cloud code is undeniably awesome interface.
>> But I often feel like people are a little bit too drawn to that and I sometimes say that's a don't try this at home sort of project. like by all means go use claude code but don't think like
at your business you should be spinning up a cloud code choose your own adventure thing probably for most cases I advise people like even still today more structured is probably going to get
you more of what you want faster in a way that everybody feels good about at the end of the project what you know what distribution are you seeing across that spectrum >> yeah I think it really depends on the
persona who are using these tools so I'm huge fans is a huge fan of clot code and the coding assistance make a big difference unlock great capabilities and
clearly very helpful for AI developers builders if you are a business user who's just asking questions like what what was my the usage of this product over the last week as a product manager
for instance I want a structured way to do this I want an agent that is already optimized for that use case that has access to the underlying data I do not want a cloud code interface for this I want something that I know will be high
quality and that's optimized. So that
that's kind of how I think about it.
Really depends on the persona that you're building for.
>> So for the talk to data product surfaces that you guys expose, how where would you say you you tend to fall on that spectrum? Is it a you're going to use these tools in this order
or is the model kind of choosing which tools to use at any given time?
>> So we have a product that we built for business users. So this is Snowflake
business users. So this is Snowflake intelligence where you can build a series of assistants. For instance, for the whole company, we built a sales assistant and we've deployed to 5,000
sellers. That product is
sellers. That product is is a think about it as like a chat interface on top of all of the company's data so that you can ask questions like
what are my upcoming renewals, how is my book of business doing and so forth and you can get answers for it. So for that clearly you want a highly optimized set
of agents for those set of use cases and and these are business users using it and they they need to trust the answers that they're getting. Then we have a set of products that we're building for for
data engineers and for analysts to to build data pipelines to analyze data that is more coding assistant if you will. So we have our own coding agent
will. So we have our own coding agent that's integrated in that platform where they're just either analyzing data or writing code. Um so that is that of
writing code. Um so that is that of course is a lot more flexible and it's also not tuned for very specific set of use cases. How do you think about the
use cases. How do you think about the question of like one big agent that might be long running versus
the other kind of big pattern is your sort of initial agent that then routes tasks to sub agents. Back when
OpenAI came out with their agents SDK, they had this notion of the handoff as a really central idea. I was never quite clear on were they doing that because
they thought that was the best way to maximize performance or was it more a nod to we think at these enterprises that are going to use this thing there's going to be different
teams responsible for certain areas and we want to be able to modularize the work for human reasons as opposed to like for AI performance reasons with Claude on the other hand and we have a sponsor of the podcast called Tasklet
which is a maximalist when it comes to just let Claude cook basically is their philosophy. Give it everything it needs.
philosophy. Give it everything it needs.
Let it make all the choices. Let it run for as long as it can run. Give it
feedback. But like it's one long one agent that kind of does it all in in one long session. Of course, I'm sure you
long session. Of course, I'm sure you could say there's different use cases deserve different paradigms, but what do you see working the most in practice today?
>> Yeah, even the cloud codec case, you have uh you have skills that uh that are being developed, right? So you you're you're still modularizing uh the different kinds of things you want cloud to do and then giving instructions for
cloud to do go do these things. I think
the way you called out is is is is what I'm seeing which is you know in in in especially large enterprises you have different teams building different agents. You also have different agent
agents. You also have different agent platforms uh that are being used. Uh so
for instance, if I'm using, you know, Salesforce to manage all of my CRM, maybe I'm going to go build uh my sales related uh experiences with an agent
there, but I still want that agent to talk to this other agent I'm building for something else. So being able to do that uh agent handoff and coordination
is uh emerging. I wouldn't say this is necessarily super top of mind for for everyone. I think still customers are
everyone. I think still customers are focused on let me get this one agent right and and and working well before I start thinking about multiple agents uh kind of coordinating with one another.
Uh but that's starting to become increasingly important uh for customers.
One of the biggest considerations is um you know they do not want to be locked in to a certain platform. So they still want to be able to you know make sure that the uh you know again open standards are supported so that agents
can talk to one another, agents can use the tools that you built for one agent uh by another agent and so forth. So you
know MCP A2A these are important protocols uh that our customers expect to be supported.
>> I was just going to ask about A2A. Are
you seeing traction with that?
>> Uh still early we don't yet support it.
Uh we're starting to hear our our customers don't necessarily ask for A2A specifically. They do ask for ensuring
specifically. They do ask for ensuring that uh you know some kind of agent to agent communication is is possible to do.
>> Is there any standard or protocol or platform that is bridging those the Salesforce continent and the the various other continents of agents >> to today we either see a bit of a hack
where these different solutions are used as tools through MCP. So still the orchestrator uses them as tools and then manages them from a agent handoff perspective other than I haven't really seen anything else.
>> Yeah, that's interesting. And one of the things I find very funny about this whole thing is that just it seems to me like a fundamental property of intelligence is that you can everything one this is a
an overstatement I don't mean it literally but one one of my refrains is like everything is isomeorphic to everything else meaning you can always squish and rearrange and play hide the
intelligence and you can have a smart MCP that's actually an agent and how you actually classify these things seems to be
much more of a choice and much less a requirement imposed on us by nature because the just the nature of intelligence itself is so flexible, funible subdividable whatever.
>> Exactly. No, couldn't agree more.
>> One of the big things, huge theme, right, of the communications that I've seen from Snowflake in preparing for this is the importance of trust. So I'd
love to hear your thoughts on what are the levels of what are the dimensions and what are the levels that we have to hit in order for an enterprise to trust
an AI process.
>> Yeah, just to uh reiterate like for us trust is incredibly important. You know if I were to call out you know two important tenants. One is super e ease of use.
tenants. One is super e ease of use.
how easy it is to build out these solutions and and to use them and and of course trust is at the core of everything and trust it spans multiple different dimensions right you know
trusting from a security perspective then from a governance perspective then you have the quality layer on top of it and then there is uh kind of evaluations and then monitoring and so forth so it's
a full stack so the way we think about this is by running AI next to data a lot of the core governance that is put on the data is by design respected in our
system. So what that means is let's say
system. So what that means is let's say you have you know sensitive data that's only visible by the HR team. If you go build an agent uh you know the person who asks the question can only get the
answer that they're eligible to see and nothing else. This is super obvious and
nothing else. This is super obvious and important, but because we have uh kind of these types of very granular access controls from the ground up as part of the core data platform, building agents
that respect that uh becomes much easier to do. Then you have uh you know
to do. Then you have uh you know governance at various layers. Um and and of course uh next level is evaluations and and the of these uh so a lot of the
trust is in are you able to build highquality u retrieval to for of context to pass to the agent? Uh is the agent orchestrator doing a great job figuring out which tool to use, which
trajectory to use to answer the question. Um so evaluation is a core
question. Um so evaluation is a core part of the platform and and then ongoing monitoring getting feedback and then that that cycle of improving improving the quality from a user
perspective. The way that that trust
perspective. The way that that trust manifests is um when a user asks a question uh you know we have uh UI elements that says hey you know this
question uh has an answer that was verified by uh by an owner. So again
bringing that uh trust element into the user experience is another tenant to um uh to our philosophy.
>> So did I catch correctly at the data governance level that the shorthand rule is like the agent can only access the
same data that the user can access. And
so in theory that could mean like multiple users could come to the same agent and have different experiences because the agent has different data access based on the user that's using it
at the time.
>> Exactly. And and this this is exactly what our customers are asking for and then that that's relatively easy to build uh on our platform. So for
instance the example I gave with our uh own kind of sales assistant. If a
salesperson comes in and says, you know, what is my book of business? Summarize
it. You should get an answer that is only, you know, your list of customers assigned to you versus another salesperson.
As an HR person, if if I ask uh or as a manager, for instance, using an HR bot, if I say, you know, what's the uh I don't know salary of this person, I can
only I should only be able to see the salary of the person that I have access to seeing versus somebody else. and
underlying it is the same agent uh and it's the uh access controls uh that govern what I'm able to see.
>> Yeah. Interesting. On the sort of performance reliability side, my experience has often been and sometimes it's for good reason.
Certainly in like the self-driving car realm, it there's a certain logic to saying we don't just want these things to be like roughly human level. We want
them to be like clearly a step up before we're going to adopt them societywide.
Good news. It seems like we're getting there. What do you What do people
there. What do you What do people have in mind as the intuitive standard of performance? Is it like they want
of performance? Is it like they want these agents to be perfect? Is it that they want them to be like at the level of the human that used to do the job? Is
it some is there some like heristic in the middle that you think people often land on?
>> Yeah. Yeah. Super interesting concept, right? The more natural the interface is
right? The more natural the interface is like the more humanlike intelligence we expect intuitively. If I'm talking to
expect intuitively. If I'm talking to the agent versus typing, I think talking has much higher expectations. For
instance, versus I'm just typing. I know
I'm typing to a computer. So the
expectations become a little less less high. I think I I I don't think that
high. I think I I I don't think that adoption of of of this technology requires humanlike intelligence because
even for the specific things that these models and these applications do well that is such high value that we're seeing huge adop as you all know we're
seeing huge adoption of AI already and and it it keeps getting better and it and it keeps getting better at a super rapid phase rapid pace. Yeah, I'm very excited about where the technology is.
>> Before going into your expectations for the year ahead, what are you seeing in terms of guard rails? Obviously, one big pattern that I think is like very natural to you guys is sourcing answers
back to the document or the the sort of authoritative place from which it came.
Beyond that though, we've got this a whole constellation of different patterns, right? in terms of you can filter inputs
right? in terms of you can filter inputs for appropriateness, you can filter outputs, you can log things and post-process logs, you can AWS has a I think a really interesting new service
called automated reasoning checks where you can put a policy in. They convert
with the language model your natural language policy into a set of rules and values and then they do like literal
formal methods to ensure that at runtime like the agent or whatever the system whatever it gave you back that it actually passes those like formal reasoning checks that were derived originally from a natural language
policy. That's like pretty interesting
policy. That's like pretty interesting and pretty cutting edge from what I've seen. What? But I think in most places I
seen. What? But I think in most places I my sense is like the frontier model companies are doing a ton of this stuff.
Anthropic has pushed this to probably farther than anyone when it comes to preventing you from using cloud to do certain things in the bio sphere. But
are people at the enterprise level actually doing much of it or are they just saying this thing seems to work.
We've got an eval set. It passes and we'll go with that. I I think the sophistication is increasing and usually companies start with products that are more internally focused. While it's
important, the bar is a little lower than something that's externally focused.
At Snowflake, we offer products to check for guardrails. They're doing things
for guardrails. They're doing things like checking for hate speech, violence, and other violations. You can detect them and flag them and not have the model respond. But we of course also
model respond. But we of course also benefit from all of the great work that as you called out companies like Anthropic do on their own models as a baseline. Some the other thing that is
baseline. Some the other thing that is also super interesting is as the models keep getting better their adherence to instructions of course keeps getting
better. So some of these also get
better. So some of these also get codified as as instructions to the agent. So, not only do this and that,
agent. So, not only do this and that, but also here's the policy that you need to to comply with. And that that tends to work quite well as well.
Have you seen anything in the interpretability realm being used for practical guardrail monitoring kind of purposes at
in in enterprises so far in we're seeing um evaluations become really important uh right so a lot of uh
you know what companies tend to do is they'll go create their own eval sets uh but also use LLMs as judges across various different dimensions to go score
uh what's happening and then continue to monitor it on an ongoing basis and uh as agents become more and more complex you know it's a it's a pretty new area to
understand is the agent you know taking the right route should I be optimizing it understanding where things go wrong becomes uh really interesting so from a so that that that's what I see not
necessarily in the core consumer experience but in the uh kind of developer experience where you're seeing you know what the model is doing, what the agent is doing through evaluations and monitoring.
>> Yeah. Okay. So, you mentioned being excited about the year ahead. You also
mentioned voice experiences. It seems
like we're at a moment like literally right now where I don't know if people just had extra time over the holiday break or whatever to get into cloud code for the first time for many people, but
it seems like the discourse has really shifted and expectations have really shifted in just the last 30 days. People
have said, "Oh my god, like the coding experience now is like it's not just vibe coding and eventually hitting a wall and giving up. Like you can actually really make this work." And
then the next big thing that people are saying over and over again is the same thing that's happened to coding over the last however many months is coming to a great many domains of knowledge work
over the next year. So do you buy that hype? And what do you think that looks
hype? And what do you think that looks like? Are we all going to be agent
like? Are we all going to be agent managers or are we all going to be like talking verbally to agents while getting lots more exercise than we used to? What
is the 2026 plus vision for success?
>> Yeah, I'd love to see the world where I don't need to do anything and I can just go get more exercise. I I do actually see the opposite. We're able to do a lot more and we end up doing a lot more.
Especially in AI where everyone is sprinting, there is more work and more productivity out there. So what I'm seeing is agents are absolutely getting more and more capable. Coding agents I
think as you called out have passed this threshold where they're a lot more because they're a lot more capable a lot more people are using them. I think it changes how products are developed. It
changes jobs like product management for instance. I've been in product
instance. I've been in product management for 20 years and like the way we build products has to change given the coding the coding assistance. How
you deploy quickly, how you test things quickly is changing because of how capable these coding assistants are. And
that is a combination of different things, right? One is the agent can do a
things, right? One is the agent can do a lot of things, can do coding well, but also the reasoning capability of the agent is is increasing. The tool use becomes incredibly powerful. So you can
apply that to other domains. If you have like now figuring out which tool to use and using it effectively and then reasoning and figuring out the next steps is very like that allows you to
build very capable agents across the board. I don't know this is not a 2026
board. I don't know this is not a 2026 projection or anything but I'm absolutely seeing clearly increasing capabilities with agents and also increasing use of them for production
work.
Do you have a sense of how the progress in AI coding assistance or agents has changed how work is happening
at Snowflake? Are you like instrumenting
at Snowflake? Are you like instrumenting that or measuring? Obviously lines of code would be like too primitive, but features shipped or burndown points per
cycle. Is there a way that you can begin
cycle. Is there a way that you can begin to quantify the impact? We're it is actually there's the impact piece but there's also the um philosophy that is
changing you know how we build products is changing um so that requires uh you know change of behavior you know usually my go-to is I have uh like let's say
there's this feature to be built I'll think about a UI I'll go build this UI and and go make it happen um whereas with a coding assistant if my users are also in living in coding assistance
maybe that's as simple as let me just build a skill for this thing and then quickly test it out. I can just write the skill in a day, put it out in front of my customers uh and then have them use it, give them feedback and only when I know exactly uh you know what the
shape of the product is, I I I can go and kind of solidify it in uh in more of consumer experience. So I think again
consumer experience. So I think again product management is changing, product building philosophy is changing because of these coding assistants.
>> Yeah, the working prototype is the coin of the realm these days for sure.
>> Yeah. Yeah. One of the big predictions that I have heard a lot recently, sometimes with a remarkable level of specificity, including from some anthropic people, is that we should see
the first drop in knowledge worker products offered this year, specifically folks have said Q2 of this year. And
what that means to them is basically a new employee that's and ultimately at heart is an AI, but it will have a very
similar surface to a remote worker on your team. It'll have a name. It'll have
your team. It'll have a name. It'll have
all the same accounts or at least you'll be able to give it all the same accounts that you can give to a human employee which means they'll be on Slack and they'll be accessible via email and they'll be all over the place and can
probably join calls. And the expectation is that this will be good enough in Q2 of this year that people will start to get value from it and this will be a new product category. I guess first do you
product category. I guess first do you buy that that can happen that soon? And
second, how many of your customers do you think will be eager to try something like that out when it drops? I
>> I do see that it's a natural progression. uh you know today uh the
progression. uh you know today uh the agents that are being built are either automating certain processes uh you know
from a productivity perspective or they are more like uh you know co-pilots that I can ask questions to and and get get responses uh versus kind of these autonomous uh
intelligence entities if you will when exactly that will happen I think really depends on how scoped can you get them to I don't think we're at a point there
uh you can just create uh another colleague uh that that can just do anything and everything but
if you can just very easily scope uh the task then I think that uh like that absolutely is possible yeah so I don't know like that's u I do see it happening
you know whether this is Q2 or not um not so sure yet and again >> find out >> yeah but plus uh you know as a data platform uh you company. I I will call
out the importance of ultimately all of these uh capabilities come down to you know for any given company the differentiation is their data uh and
then the access to that data being able to figure out and retrieve the right types of data to be able to answer that question and then increasingly the tools that are given to these agents to take action. I I think that changes
action. I I think that changes industries, right? That changes how we
industries, right? That changes how we think of data. That changes how we think of making AI, making the data AI ready as well as making the tools AI ready. Uh
so that more and more kind of capable agents can be built.
>> Do you see changes to data itself? I
guess one that we've talked about already is just retroactively going back and applying structured unstructured data, creating metadata, so on and so forth. In the wild, one big change that
forth. In the wild, one big change that we're seeing to data is like the web itself is increasingly comprised of AI
generated data. And so that's a weird
generated data. And so that's a weird feedback [clears throat] loop that we accidentally created. Are there any
accidentally created. Are there any other perhaps surprising patterns in data in within enterprises that you're
seeing as a result of AI coming onto the scene? or is it maybe just still too
scene? or is it maybe just still too early for something like that to be?
>> I am seeing two things. One is I'm seeing
two things. One is I'm seeing the access getting a lot easier. So that
democratization of access to data and access to insight is a big shift. The
other thing that I'm seeing is the value that our customers get from data is increasing because you're able to very easily glean those insights by just describing what you want in natural language and then getting it. So the
value you get from data is increasing.
How that just opens up new and new opportunities. You start using the data
opportunities. You start using the data in ways that you haven't thought of before. One interesting study, one of
before. One interesting study, one of our customers as S&P, they analyze the earnings calls to understand when CEOs
are responding to analyst questions in either directly or indirectly or if a question was already answered in opening remarks or not and using that as alpha
to determine which stock to buy. So
stuff like that becomes very easy to build and then new use cases open up.
>> Yeah.
That's an interesting metric. I've seen
a bunch of stuff recently over the last even just the last week. It was really the Venezuela moment where all of a sudden people were like bragging about how they had created these AIs that monitor the prediction market platforms
and were looking for early signal and trying to capitalize on that. That is
going to be a really interesting phenomenon. How about in terms of just
phenomenon. How about in terms of just let me go slightly a different direction then I'll do maybe a little lightning round to close us out.
>> I have this sense that right now we're in this kind of expansionary phase. I'm
no astronomer, but my experience with platforms in the past and definitely experiences with the meta platform, formerly known as Facebook, where it was like they came on the scene, they opened
up a ton of stuff. Everybody could tap into all these data and social connections and there was for a moment there was like an incredible flourishing of a ton of different ideas. And then
after that supernova came the black hole and it was like actually we're going to close all this stuff back down. And a
lot of that value that entrepreneurs created on the edges experimenting with different ideas, the things that really mattered mostly ended up getting sucked back into the platform and there wasn't
nearly as much value created on the margins as it seemed like there was going to be. And I I might be just over indexing on this experience of having
lived through this pattern once before.
But I feel like the AI moment is set up for that to happen to a lot of people again where for example just this week, right, Chad GPT launches medical
version, which is great. I think that's going to be awesome for a lot ton of people. And the fact that I can now just
people. And the fact that I can now just connect my chat GPT into an EMR instead of having to go laboriously copy and paste or find some other third party thing to do that kind of connector work for me like the consumer surplus of that
is going to be amazing. That's like my strongest belief is we will see high consumer surplus >> but for the businesses it seems like it does create a very tricky balance where you're like I want to go do a bunch of
cool stuff but how do I know which of these things will be durable over time?
How do I know I'm not just doing like R&D for the next generation of the mega platforms that are ultimately going to eat my lunch? How do you guys think about like where you want to place your
bets, but what is going to get absorbed into the models versus what you know only you can do over a longer period of time.
>> Yeah, I think we are in a fortunate place because ultimately data is an incredibly important asset for all companies. that is what kind of defines and differentiates them and that
is not getting commoditized anytime soon. So as a data platform we sit in in
soon. So as a data platform we sit in in in that layer in between the application and the model if you will. So the way we think about this is how do we help our customers build very high quality
products that are catered towards their use cases which is all powered by their data and whether any of that can be
subsumed by these other platforms. I am sure the shape of shapes of products will continue to evolve. I
think we're still in the very early innings of huge transformation. But I
also believe intelligence is a I it like once these models are out there and there's enough competition and we're seeing enough competition like the dynamics seem to be playing in such a
way that is all pro customers and consumers versus these mega platforms. So I do believe the competition will keep things in check and the opportunity is so massive that that growing pie will
also create lots of new opportunities.
Obviously, people have to have some place to keep their data. One,
tell me what's wrong with this theory.
If I were to be a skeptic or if I took the perspective of the snowflake bearer for a second, re a recent experience I had was my company's been a customer of intercom for a number of years
>> and [clears throat] I was trying to do just some basic analysis of recent tickets. So, I and they didn't have the dashboard to do
what I wanted to do. So I went to their docs and the docs were 100 pages of docs, right? It's fullfeatured platform
docs, right? It's fullfeatured platform at this point. So a lot of docs. So the
first thing I did was told a web agent, hey, go compile all these docs. And it
literally went page by page, copied them all into a Google doc, like in a browser. I ended up with some 600 pages
browser. I ended up with some 600 pages of text. Then I took that to Gemini and
of text. Then I took that to Gemini and said, "Okay, hey, there's a lot of repetition in here, but can you streamline that down to what I really need to know?" And so it did that fit in the even 600 pages whatever it like fit
in the million tokens. So now I've got my consolidated single view of the docs.
Then I go to a coding agent. I say here what here's what I really want to do is export all my data from intercom. And
that also ended up being just one prompt to work. And then it exported all my
to work. And then it exported all my data from intercom and was able to do the analysis I wanted to do. But then
the sort of Eureka moment was like wow it's never been easier to unplug from intercom if I want to take my data somewhere else. They didn't really
somewhere else. They didn't really anticipate it being that easy when they created all these APIs. So
what what prevents the data platforms of today from running into trouble there in the past presumably like somebody were to say hey I'm not happy with you or I want a better price whatever you had some
leverage of just what are you going to do you going to pull out all your data and I'm not saying that would be easy to say cuz I know you guys handle vastly more data than I have in intercom but it does also strike me that like it's become a lot easier to move things around it's become a lot easier to
understand what it is especially once you've gone and done all this metadata layering on. So what are the modes? What
layering on. So what are the modes? What
are the sticking points? Has it changed or or will it change?
>> It is going the direction that you're calling out which is I think great for again customers and consumers. So today
um Snowflake supports uh you know open file formats for storing data that we support uh iceberg which is a uh you know open file format. What that means is you do not have to have your data
locked in somewhere. you can just put it in a uh you know in a you can keep it uh in a in a managed uh place that's managed by you or by us and then you can
use Snowflake as a um as an engine to go process your data. So we are absolutely uh embracing and and supporting the ability uh for for our customers to uh
to use these open file formats to be to to not necessarily feel like they're locked into one platform. And I think that's that's great. That's great for customers. uh that's great for
customers. uh that's great for innovation. Ultimately, customers will
innovation. Ultimately, customers will end up using the product that is going to give them the uh you know best performance uh best best cost uh for the things that they'd like to do and you
know we're absolutely embracing that.
>> Does that translate to increased pressure on you and your team? Like it
would seem like maybe one way to think about that would be like in the past if somebody wasn't happy for a year maybe they would start to think about a switch where now it might be if they're not happy for a quarter. Does it shrink the
timeline where you have to deliver?
>> I I love it. It's it's great for product teams right ultimately you know we're all driven by you know creating value for customers building great products and then we want to do that as fast as
possible. uh you know and and
possible. uh you know and and competition allows uh it's a great uh incentive in the system to to go you know keep things in check uh and and
then uh you know have you deliver so I don't think things change from uh for my team we we already do feel pressure not necessarily because of competition because of the opportunity the
opportunity is massive and uh the it's also there's never been a great time to be a product manager right so you're you're easily able to build awesome products very very quickly and then
you're sprinting. So it's incredibly
you're sprinting. So it's incredibly satisfying uh to to build these great products and then you also reap the benefits by seeing how these are getting used in the market. So I love the competition. I also love the pace of
competition. I also love the pace of innovation in the industry.
>> Does that lead you to a point of view big picture on this is a classic question and again it's striking to me how very informed and technically sophisticated people have very different
answers. Where does the value how do you
answers. Where does the value how do you think about the breakdown of where value occurs? Obviously, we've got
occurs? Obviously, we've got infrastructure whether that's like chip creators or owners models application layer on top. If you
had to assign those three layers relative value capture from the sort of AI opportunity, how do you think that breaks down?
>> Yeah, I I think may maybe the way I think about it is the uh the middle will erode and uh and and the uh the sides will continue expanding. So so far we've
been seeing a lot of value acrewing to uh you know chip makers uh Nvidia as well as the model providers um uh and uh and then increasingly application
developers who are able to go build kind of very quickly you know unique businesses uh on top of these capabilities. Uh cursor comes to mind as
capabilities. Uh cursor comes to mind as an example. So so I I absolutely do see
an example. So so I I absolutely do see the value continuing to acrew at the um infrastructure layer as well as at the application layer. And uh you know
application layer. And uh you know traditionally there's always been this kind of middle layer that's that's kind of facilitating and and uh connecting those two things
because of the capabilities of uh of these models that middle layer may not be as valuable uh or as important anymore.
>> And that middle layer is the models. Is
that right? Uh no no no I I do believe it's um so so the the middle layer is is um all the companies that are kind of creating custom business logic for
certain applications right so that business logic as you called out for instance if you want to just uh you know build your own extractor you can just vibe code it over a weekend and and go do it versus a company that goes and
builds it for you. So so that the that layer isn't as as important. So um
models are in my opinion will continue to acrue a ton of value.
>> So it yeah okay so to try to play that back to you it sounds like you think all three of the layers that I described will do fine but at the application >> traditional businesses will change.
Yeah, >> you're going more horizontal platform and less you like relatively less excited about vertical because the
fact I mean so many SAS applications like essentially exist to encode business logic or best practices or whatever and we just probably don't need
dedicated teams building out those kinds of things when we can just have agents kind of do it on the fly as as needed.
>> That's what I'm guessing over time.
Yeah.
>> Yeah. Okay, cool. One other big kind of question that I've asked a lot of people a lot of times and I think you're the perfect person to touch on it. So you of course you know that data bricks
acquired this company called Mosaic ML not too long ago maybe two years ago now and what Mosaic was doing I thought was really interesting which was starting with open source models working with
particular customers to do continued pre-training on data sets which I assume were very often internal data sets like the sort
of data sets that might sit in a snowflake. I was really surprised. I
snowflake. I was really surprised. I
spoke to Ali Gatsi, the CEO of Data Bricks at a event not too long ago and he said, "We killed that product." So
they basically turned Mosaic into a in-house research unit, but that product of offering this continued pre-training to try to create a model that like really knows your business inside and
out. Basically, they don't offer it
out. Basically, they don't offer it anymore. I was very surprised by this
anymore. I was very surprised by this because I think man if I had a if I'm GE or if I'm 3M or any number ofundred-year-old millions of employees over the
generations companies that have this incredible history and so much data that's acred that nobody really understands at the company these days.
If I could have a model that could have similar command of that information that doesn't that only exists in my company and just nobody else outside has
ever had access to as the foundation models today generally have world knowledge. I would think that would be
knowledge. I would think that would be insanely valuable for a lot of enterprises and yet we don't seem to be
seeing to my knowledge many instances of whatever 3M GPT or GEGPT fiser GPT like why don't we see that
>> do you have a point of view >> I do um so this is kind of similar to how up until recently when you'd ask a question on chat GPT you'll say hey you
know my uh you information cut off is whatever you know a year ago and I can only answer questions up to that point and then uh web search as a tool came in and now uh you know all of these
platforms would use web search to give you the most up-to-date information so that their world information can be it is more about the intelligence to figure
out when to use the tool to retrieve the information and then make sense of it and then give it back to you versus having been trained pre-trained with all that information up front to to me that
that that is that that that pattern is exactly what's playing out right so in the enterprise world you have a lot of information uh and then your text to SQL
and and rag solutions can bring that information in for the uh for the agent for the platform to reason with uh and then give you uh give you information.
The nice thing about that is it is substantially cheaper. uh the model
substantially cheaper. uh the model keeps getting better as the underlying you know premier premier model keeps uh improving and uh it's also relatively
easily tunable. You can you can update
easily tunable. You can you can update it uh you can change things and so forth. So that means for me majority of
forth. So that means for me majority of businesses uh would continue to you know benefit from this architecture rather than codifying all of that information
in the weights of the model. um they'll
just uh use the information uh and then use tools to retrieve parts of the information that are relevant. The
exception to that is what we discussed earlier which is if there are certain tasks that require either high throughput low cost uh if you have a lot
of data uh and uh in an area that the model has not seen before then it might make sense to go create custom models for those uh specific tasks. So I do
believe there's going to be increasingly uh need for kind of task specific small models in large corporations or you know when you have you have that need uh but
still the majority of the uh use cases will be more retrieval oriented.
So I think that's a great first pass answer. If I
answer. If I think though even just about my own ability to search through my own stuff, right, my own Gmail, my own Google Docs, one of the intuitions I have pretty
strongly is if I were to give you full access to my Gmail and give you full access to my Google Docs, you couldn't search through it nearly as well as I can. And that's despite the fact that
can. And that's despite the fact that you're clearly smarter than me. So I'm
like there seems to be something about the fact that I have had this like free training on this corpus that allows me to even just search through it a lot
better because if nothing else like one of the intuitions is I know when I found what I'm looking for. Right? You might
not know you could search you could do a 100 searches in my Google Drive and never be quite confident you got the absolute best document for whatever the question is. Whereas if it's my Google
question is. Whereas if it's my Google Drive and I created all those documents when I hit that document that is the one that's yes this is the one now I yes I remember this now this is the one. So I
have that sort of confidence that I've got to the answer if nothing else I think which is strikes me as like really hard and I've seen this when I try to give my own just whatever give cloud access to search my drive like it also struggles in that way. It doesn't know
how many times to search or if it's found the right thing or sometimes it's satisfied too easily whatever. So I
still feel like there's something there where you could expect that a model that really had a sort of more in the weights
familiarity with the data could do a better job of navigating it. And then
maybe it just comes down to upgrade cycles are terrible for this kind of thing. And yeah, as you said like you
thing. And yeah, as you said like you want to keep taking advantage of better and better models. This potentially
workshes obviously influenced the discourse recently with kind of focus on continual learning. So maybe you need
continual learning. So maybe you need either a new architecture that's more suited to that, you know, some sort of new training paradigm that would be more suited to it. I guess maybe one way to phrase this is if that were to flip, if
you were if you imagine a world a year from now where it's no longer the case that the best approach is take the pick the best models and like tune leave
them as they are, but tune them through the search as you described. and it
instead becomes one of these things where they actually do have this like deeper familiarity with like all the enterprises data. What do you think
enterprises data. What do you think would what would have changed to flip us from one paradigm to the other?
>> I was trying to kind of think about how humans do this. We'd go into a environment that we don't know and then we'll go do a bunch of searches and then we'll read to create to create some kind
of knowledge and then as you build out that knowledge that there's intuition that comes with it. So you don't need to keep referring back to it and then somehow that turns into intuition.
Right? Right now I think what these models are doing is the first part. I'll
go pick the information as the context windows of these models also keep getting better. I can stuff more and
getting better. I can stuff more and more information in these models and then get an answer. What intuition is not really understood. So I don't really know how to how that changes the
dynamic. What what would change if if a
dynamic. What what would change if if a model is trained with your data? you
clearly need much less data to steer it to a certain direction. You'll have much more consistency in in in the responses.
It still feels I don't think you ever get away from feeding it information, up-to-date information and so forth. But
what I would imagine happen is first of all that model that you want to do certain tasks doesn't have to write a poem in French. So you'll benefit from using the weights more efficiently for
the task that you want to do. And
therefore perhaps again you may not need as large a models so you get benefits from more optimizations to to reduce the cost, increase the speed and so forth.
>> Yeah, I think that many small models paradigm is also one that I'm pretty bullish on for quite a few reasons. one
being just I think it we stand a lot better chance of staying in control of the meta if we have a lot of like narrow AIs doing their jobs as opposed to you know a relatively smaller number of
giant AIs like running things for us.
The pull of that is obviously pretty strong but the I do worry that we're racing into having such general AIs that
can do sort of anything before we've really thought through what the ultimate consequences of that are going to be.
And I the narrowness safety through narrowness and maintaining control through narrowness I think is an underdeveloped paradigm.
>> I fully agree that that's again it's going back to human analogy. That's how
we operate as well. There are certain parts of the brain that that that are specialized to do certain tasks and yeah so I can absolutely see that.
>> This has been amazing. Couple of quick uh closing questions. What are you watching right now in terms of horizon scanning for surprises? Is there
is there a capability threshold that's on your mind or like what because obviously nobody can keep up with the AI news these days, right? So everybody has to pick and choose. What are the areas that you're watching? And maybe another dimension of that is what are the
metrics that you're watching? Are you
looking at ARC AGI scores? Are you
looking at GDP val? Are you looking at the MIRI task length chart? Are there
other what do you trust to give you the highest signal on what is actually important in the latest things that are coming out?
I I actually don't watch the public benchmarks as as closely like we do have internal series of benchmarks that watch very closely in terms of quality and latency for the tasks that we're optimizing for which is of course built
on top of the models that we get from the model providers. Whenever there's a new model that's about to be released, we'll go run our our tests, figure out what's improving, what are the gaps, and then that I watch very closely. In terms
of technology trends, one thing that is maybe unique to Snowflake of course is we have a lot of tabular data and that technology so far has been all about
text to SQL and semantic models. So
watch that space quite quite carefully and there are some new trends happening in that in that space. These tabular
foundation models are interesting. Being
able to quickly build forecasting models and so forth are now possible. So those
are other trends that I watch as well.
Yeah, that's an interesting one. The
there's a few public uh forecasting like benchmarks and competitions. I think
those are really interesting too. At the
point where the models are better able to predict the future than our best super forecasters or even aggregations of super forecasters, that will feel I think like a very meaningful shift in in
what's going on in the world.
>> Any contrarian takes? any anything you think like the anything you want to correct that you think the audience at large might be misled or misconceiving right now?
>> I mean we touched upon the one that that's very top of mind for me right now. I think you know the way we build
now. I think you know the way we build products is is change. I don't think it's contrarian but I don't think it's happening fast enough. Um we're at a point where how we build products needs
to radically change and uh you know that means change of behavior uh because we've been kind of trained to build products one way. So so to me that that's the biggest one you know in a in
a world where these coding agents are such kind of capable platforms. How do you build new products?
And you know in my mind it is all about kind of starting with that first uh and then validating things quickly u before you build the product in the first place.
>> Yeah I think that's I've been doing that with my mom. I made it made their the custom travel app travel planning app for the holidays and it has been yeah it's inverting that process right I made
a version she has it and now I sat down with her this morning over coffee and I'm like what do you want this thing to do that it can't do and she's I want to ask you to do more on this. I'm like,
"Mom, it's honestly so easy at this point. If you can articulate what you're
point. If you can articulate what you're missing, there's a pretty good chance we can get Claude code to just make it and you can have it from one session to the next."
next." >> My last question, then I'll give you the final word. What advice do you have for
final word. What advice do you have for enterprise leaders in general?
Obviously, you guys are much closer to the core. For all the executives and
the core. For all the executives and product owners at the companies that you serve, what do you think they should better appreciate or what can they learn from your experience? There are
[laughter] there are some uh you know enterprises that are still quite careful about adopting AI and and at this point it is
uh it is so powerful that it is uh there's a race. So you know the the faster enterprises adopt AI the more benefit they're going to get and and and
the more intuition that they're going to get that changes the trajectories of of these businesses. So to me it's
these businesses. So to me it's incredibly important to intuitively understand and natively use AI because it is going to change industries and
then underlying that is uh all about many of the hesitations tend to be about uh you know getting the uh data ready for AI from our perspective you know so
so that means investing in in in that core foundation to uh you know essentially get the data AI ready for AI to use. So that means breaking down
to use. So that means breaking down silos, getting the data uh you know accessible locked in for certain use cases like that becomes a core enabler to build on top of.
>> Makes sense. We've covered a lot of ground. I really appreciate your time
ground. I really appreciate your time and jumping through all these topics with me. Anything else that we didn't
with me. Anything else that we didn't touch on that you would want to leave people with?
>> No, maybe the thing to call out is we talked about a lot of great capabilities as as well as trends. One thing that is sometimes not necessarily appreciated is
how easy it is to to use AI and how easy it needs to be to use AI for adoption.
So that that's an area for Snowflake that's super core. So as we build products make making it very easy to deploy highquality AI at scale is something that we strive towards. To me
from a design principle perspective that is key as well.
>> Yeah, couldn't agree more. Beer
Schultzan, vice president of AI at Snowflake. Thank you for being part of
Snowflake. Thank you for being part of the Cognitive Revolution.
>> Thank you, Nathan. Thanks for having me.
>> If you're finding value in the show, we'd appreciate it if you take a moment to share with friends, post online, write a review on Apple Podcasts or Spotify, or just leave us a comment on YouTube. Of course, we always welcome
YouTube. Of course, we always welcome your feedback, guest and topic suggestions, and sponsorship inquiries, either via our website, cognitiverevolution.ai, or by DMing me on your favorite social network. The
Cognitive Revolution is part of the Turpentine Network, a network of podcasts which is now part of A16Z, where experts talk technology, business, economics geopolitics culture and
more. We're produced by AI Podcasting.
more. We're produced by AI Podcasting.
If you're looking for podcast production help for everything from the moment you stop recording to the moment your audience starts listening, check them out and see my endorsement at aipodcast.
And [snorts] thank you to everyone who listens for being part of the cognitive revolution.
Loading video analysis...