Nvidia Just Open-Sourced What OpenAI Wants You to Pay Consultants For.
By AI News & Strategy Daily | Nate B Jones
Summary
## Key takeaways - **OpenAI/Anthropic Partner with Consultants**: Anthropic and OpenAI spent a year figuring out companies lacked expertise to apply their solutions like codec and claude code in production, so they are tying up with big consulting firms to get their content into hands accessibly. [00:26], [01:01] - **Nvidia Launches Secure Nemo Claw**: Nemo Claw is an add-on to OpenClaw designed to run in Nvidia's OpenShell runtime with policy-based guardrails via YAML and model constraints for enterprise security and local compute on Nvidia chips. [02:17], [02:32] - **Jensen Trusts Developer Competence**: Jensen assumes enterprises and developers are competent enough to adopt Nemo Claw, an open-source framework, without needing consultants, unlike OpenAI and Anthropic's approach. [04:15], [04:27] - **Rob Pike's Rules Apply to Agents**: Rob Pike's five rules—you can't predict bottlenecks, measure first, avoid fancy algorithms, simple is less buggy, data dominates—still hold true for agentic systems and are basics Nvidia leverages. [06:07], [09:05] - **Factory.ai: Environment Trumps Agent**: Factory.ai's agent readiness framework shows the codebase environment across eight pillars like linting, testing, and observability is usually the broken part, not the agent itself; fixing it makes agent behavior self-evident. [11:47], [12:05] - **Nemo Claw Open-Sources Consultant Work**: Nvidia open-sources Nemo Claw as enterprise-grade security and compliance over OpenClaw, providing what OpenAI charges consultants for, empowering developers with proven engineering practices. [03:40], [23:03]
Topics Covered
- Consultants Fail Where Engineers Succeed
- Nvidia Bets on Developer Competence
- Simple Scales in Agentic Systems
- Data Engineering Trumps AI Hype
- Anchored Summarization Wins Compression
Full Transcript
Right now there's a battle playing out at the heart of agent world and it's a battle between titans, right? Nvidia's
on one side with Nemo Claw, OpenAI and Enthropic are on the other side. If
you're telling me Nate, no, no, no, they're all building agents, I'm the first to agree with you. That's not the point. The point is that Anthropic and
point. The point is that Anthropic and Open AAI spent a year in 2025 figuring out that the companies they work with did not have the expertise to actually
apply the solutions they were giving them. So they would launch cool stuff
them. So they would launch cool stuff like codec and claude code and see it suffer in production when they could not figure out how to get actual teams at actual businesses to adopt them in ways that they themselves were using
internally right anthropic ships I swear every 8 hours right and open AAI ships very very fast as well but they weren't seeing those speed ups at other companies and they could not figure out why and so now because of that year of
failures open AI and anthropic are very publicly tying up with big consulting firms and they're doing that because they know that they need to find ways to
work with services firms to get their actual content, their actual code into the hands of people in a way that's accessible to them. It turns out that AI
doesn't teach itself, at least not for most people. And I think that's a bitter
most people. And I think that's a bitter lesson that Enthropic and OpenAI have learned. I don't know that Nvidia agrees
learned. I don't know that Nvidia agrees because on the other side of this, Nvidia just launched Nemo Claw and the backstory there is very very different.
Nemo claw came from the open claw moment, right? Jensen walked out onto
moment, right? Jensen walked out onto the stage and he said this is the future, right? The future is open claw
future, right? The future is open claw because the future is an agentic operating system. And that's what he
operating system. And that's what he saw. And so regardless of what you think
saw. And so regardless of what you think about OpenClaw the piece of software that Peter Steinberger coded, OpenClaw the system, OpenClaw the paradigm, OpenClaw the idea, that's what Judson
was talking about. And he wanted to take that idea and bring it securely to the enterprise. Because of course the big
enterprise. Because of course the big thing with OpenClaw if you're in business is it's not secure. It's not
something you can lock down well.
There's lots and lots of issues with giving your agent access to your stuff and the open internet. And so Nemo Claw is designed to be a lot more locked down. So what makes Nemo claw tick? Nemo
down. So what makes Nemo claw tick? Nemo
claw is actually an add-on to OpenClaw.
It's not that it replaces it entirely.
It's that it's designed to run in OpenShell, which is Nvidia's proprietary runtime environment. And that ensures
runtime environment. And that ensures that Nvidia is able to wrap the open call instance in a way that's secure. So
it has policybased guard rails which are YAML declarations which the agent has to follow. It has model constraints which
follow. It has model constraints which do two jobs. Job one is ensuring that Nvidia can validate the safety, but really job two is ensuring that Nvidia gets to serve the model because one of
Jensen's larger moves here is to go from just managing the chip layer to move into the Agentic world because in his business he needs to go from just selling chips to scaling up to sell more
of the value chain. And he's convinced Agentic is a big piece of it and hence Nemoclaw. Nemoclaw also runs on local
Nemoclaw. Nemoclaw also runs on local first compute. And yes, as you'd expect,
first compute. And yes, as you'd expect, there's an Nvidia play there because Nemoclaw is designed to run safely and efficiently on Nvidia chips that run locally. Nemo Claw is very much a
locally. Nemo Claw is very much a strategic play for Jensen because what Jensen is trying to do is he's trying to figure out how to pivot into an
ecosystem play where everybody who has all of this energy around OpenClaw will be indirectly contributing to value in Nemoclaw, which he can then sell to enterprise. Like that's the dance he's
enterprise. Like that's the dance he's trying to walk here. And by the way, if you're a contributor to OpenClaw and that makes you annoyed, I get it. This
is just part of how corporate works. And
so the long and the short of it is that Jensen is bolting on enterprisegrade compliance and security solutions as a patch, as a layer over the top of
OpenClaw to make it something with an open framework that runs on Linux that enterprises can pick up and use. Whether
or not you find that believable, I want you to step back and look at how this assumes competence on the part of enterprises. Remember, we started this
enterprises. Remember, we started this video and we talked about the story anthropic and open AI have been telling themselves where they recognized very publicly over the last year or so that
their solutions were too complicated to successfully roll out to engineering teams at enterprises. Now, here comes Jensen onto the stage and he says, "You know what? You developers are smart. You
know what? You developers are smart. You
developers can figure this out. People
are already using OpenClaw by the hundreds of thousands. You guys got this. Let me just roll out this
this. Let me just roll out this open-source framework and we're good to go. And you know what? I think one of
go. And you know what? I think one of the things I notice about Jensen's approach. It's not necessarily the
approach. It's not necessarily the corporate strategy here. It's actually
the fact that a lot of what he focuses on are basics that we have known in data backend engineering for a long time. And
this is something that I keep coming back to and thinking about as I go through change management processes with companies. I recognize that in many many
companies. I recognize that in many many ways what consultants are making complicated today is actually the age-old practice of good data
engineering that turns out to be super useful in the age of AI. And I can't help but wonder if open AI and anthropic changed their tune a little bit and instead of saying AI AI AI isn't it
amazing and complexifying it for people if they actually came in and said let's talk about what we've always known as developers. Let's talk about how data
developers. Let's talk about how data actually works in the principles of development and then and then let's talk about how AI ladders onto that data backend in ways that are really useful.
Maybe the process of change would be easier. And I think in a way Jensen
easier. And I think in a way Jensen understands that. Just for fun, let's go
understands that. Just for fun, let's go all the way back to Rob Pike's five rules of programming. If you don't know who Rob Pike is, you should because he's one of the creators of Unix and Go. He's
an absolutely legendary developer. Rob
Pike's five rules are things that get taught computer science. They're things
that senior engineers teach to juniors.
They're sort of written in the stars if you're in the discipline. Rule number
one, you can't tell where a program is going to spend its time. Bottlenecks
occur in surprising places. So don't try to second guessess and put in a speed hack until you've proven that's where the bottleneck is. I cannot tell you how many times I've used that rule when debugging systems. It actually works. It
is very hard to tell until you run a system where the bottlenecks are going to happen. That is true for agentic
to happen. That is true for agentic systems people. That rule didn't go out
systems people. That rule didn't go out of style. And by the way, yes, I'm going
of style. And by the way, yes, I'm going through all five of these because I don't think we talk about them enough.
And I don't think we realize amidst all the hype and all the change that some of these ancient engineering practices still hold true. Rule two, measure.
Don't tune for speed until you've measured. And even then, don't do it
measured. And even then, don't do it unless one part of the code overwhelms the rest. In other words, if you aren't
the rest. In other words, if you aren't measuring and baselining your performance, it's really hard to optimize. Do we see that with aentic
optimize. Do we see that with aentic systems? We sure do. How many times do
systems? We sure do. How many times do people tell me they don't like an individual LLM response and I have to tell them maybe you should baseline it?
Maybe you should measure before you make big assumptions and changes. Rule number
three is kind of just don't get fancy or more precisely it's fancy algorithms are slow when your number is small and your number is usually small in computer
science terms. Fancy algorithms have big big constraints. Fancy algorithms
big constraints. Fancy algorithms usually only work at scale. Until you
know that your number is frequently going to be large, don't get fancy. This
is true for agentic engineering as well.
If you're trying to build aic systems, simple scales well. And in fact, I would add there's probably a correlary here.
Simple scales better than complex. And
this is something that may have shifted with agentic engineering because we did find for a while if we were writing algorithms that there were times at large scales when you had to have a
fancier algorithm. Now I think we're
fancier algorithm. Now I think we're abstracting a lot of that edge case complexity to LLMs and that requires us to have very stable simple architectures that scale. So that's one that I have
that scale. So that's one that I have some interesting nuance around but fundamentally it's true right don't get over fancy especially when the system is small. Rule number four, fancier
small. Rule number four, fancier algorithms are buggier than simple algorithms. This was the era, by the way, when Rob had to write his algorithms by hand. I know that everyone here doesn't know that anymore because
we all just prompt our LLMs. But this was handwritten stuff, right? Use simple
algorithms for simple data structures.
That's the heart of rule number four.
And this is a correlary to rule three because if rule three talked about simplicity and scale, rule four talks about simplicity and bugs. It is very very hard to debug complex agentic
systems. You're like, is it the prompt?
Is it all of this context that I'm pulling in? What's the problem? As much
pulling in? What's the problem? As much
as you can simplify because the more that you simplify, the better off you're going to be, the better off you're going to be debugging, the better off you're going to be maintaining the system, etc.
Rule number five, data dominates. If
you've chosen the right data structures and if you've organized things well, the algorithms will almost always be self-evident. In other words, write dumb
self-evident. In other words, write dumb code and have smart objects in your data system. Right? That's the short version.
system. Right? That's the short version.
This cannot be more true in the age of AI. Data engineering is the key to
AI. Data engineering is the key to having good smart agentic systems. And I think we miss that. This is not new at all. This is decades old. Every time we
all. This is decades old. Every time we go through hype cycles, and I've been through a bunch of them, right? I've
been through the cloud hype cycle. I've
been through the mobile hype cycle. Now
I'm in the AI hype cycle. And we forget.
We think it's all new. And we forget little things like the fact that we should keep structure simple, that data dominates, that we should build data structures that enable us to do more
complicated things in ways that are sustainable. This is what Jensen is
sustainable. This is what Jensen is arguing for when he wants a simple set of primitives to build an open-source ecosystem for agents. In a way, I think
Nvidia's engineers understand this better than a lot of the other engineers in the AI ecosystem right now. And that
may be because they have to be so close to the kernel and so close to the metal all the time. You have to have good principles when you're trying to optimize for GPUs. And when you optimize
for GPUs over time, you build an engineering culture that demands excellence and adherence to good best practices. And I see that written all
practices. And I see that written all over Nemo Claw. And I think that if we look at the story of how much trouble organizations are having adapting to AI
and if we ask ourselves is it the message itself that's the problem or is it the way it's presented I would kind of argue it's been the way it's presented because we have presented I have seen so many consultants pedalling
complexity as if it was a good thing with AI like presenting some kind of complicated agentic mesh and saying this is the way or presenting a really complicated change management paradigm
or presenting lots and lots and lots of very hardto- read docs and saying go dig into this. These are your prompting
into this. These are your prompting tools. Simpler scales. We need simpler
tools. Simpler scales. We need simpler approaches that enable people to understand what we're saying. And
ironically, if we go back to the way we always engineered systems, we're going to find that a lot of those truisms like Rob Pike's rules still work. They're not
out of style. And that brings me to one of my favorite examples in the age of AI because I want to make this more updated. Yes, there's new things, new
updated. Yes, there's new things, new changes, but we have to understand how these old structures are informing new ways we work. I think factory.ai has a wonderful example here. Their agent
readiness framework evaluates code bases against eight different technical pillars. style and validation, build
pillars. style and validation, build systems, testing, documentation, the dev environment, code quality, observability, security, and governance.
And what they find is that consistently speaking, the agent isn't the broken thing. The environment is, which goes
thing. The environment is, which goes back to that data insight. If you can fix your data structures like llinter configs, like documented builds, like dev containers, like an agents.mmarkdown
file, agent behavior then becomes self-evident. It's effectively a
self-evident. It's effectively a correlary to what Pike was talking about years and years and years ago. And so
Facto's data shows that getting these fixes right compounds in exactly the way we would expect it to following good software engineering principles. If you
have better environments, you make your agents more productive, which frees time to make your environments better, which in turn feeds the loop and your agents get more productive over time. And
there's a convergence here around Agentic best practices that I want to call out and name explicitly. So I'm
talking about factories best practices, Nvidia's best practices, but also some of the way Enthropic organizes things, some of the way Microsoft organizes things. There are essentially a whole
things. There are essentially a whole set of agentic rules of the road that we are publishing that are Pikees rules rediscovered by people who know their
fundamentals. And I want to name the
fundamentals. And I want to name the primitives that are emerging because I think that we should understand these rules of the road that underly best practices across a bunch of different companies and recognize their old roots
cuz I think it will help us to change more effectively. So with that, I want
more effectively. So with that, I want to walk you through the five hard problems that I've seen in production agent deployment. I'm going to go
agent deployment. I'm going to go through each one in detail because the distribution of difficulty here tells you about where people are spending money, where people are expecting engineers to solve it internally and
really what best practice looks like.
The first one is context compression. So
longunning agent sessions fill up context windows. They just do even
context windows. They just do even million token context windows or 10 million token context windows, they all fill up. And every compression strategy
fill up. And every compression strategy is lossy. It always loses something. So
is lossy. It always loses something. So
factory tested three different production approaches to see which was best. They had their own method which
best. They had their own method which they call anchored iterative summarization. Big words. It maintains a
summarization. Big words. It maintains a structured and persistent summary with explicit sections for session intent for file modifications for decisions made
and for next steps. When the compression triggers the newly truncated span gets summarized and then merged with the existing summary. So the structure
existing summary. So the structure essentially forces preservation. you
can't break the previous summary. Right
now, they compared this approach against OpenAI's compact endpoint, which produces a very opaque. You can't see what's on the black box, and it just gives you compressed representations
that are optimized to be reconstructed faithfully. That's a fancy way of saying
faithfully. That's a fancy way of saying it's it's compressed very highly, and you can't read the output to verify what was preserved because OpenAI famously doesn't expose any of that. And then
they tested it against Anthropic's built-in compression through the cloud software development kit, which generates very detailed structured summaries, but regenerates the full summary every time rather than doing it
incremental. That difference starts to
incremental. That difference starts to matter across repeated compression cycles because you're regenerating the whole summary. You're playing telephone
whole summary. You're playing telephone again. The results were clear. Facto's
again. The results were clear. Facto's
approach of incremental summarization scored the highest, but all three struggle with tracking artifacts. So if
you're naming and remembering particular files, all three struggle with that a bit. And the mitigation here is pretty
bit. And the mitigation here is pretty simple. You have to think about your
simple. You have to think about your project in terms of milestones and make sure that the milestones can be compressed in ways that allow the agent to continue to work. And that if you
cannot do that, you have multi- aent frameworks that allow the agent to pick off and address big pieces of work and then die and refresh the context window with a new agent without losing that
context. so that you get these
context. so that you get these longunning tasks. That's how you get
longunning tasks. That's how you get these multi-week agent runs and don't stuff out the context window. You see
how it all comes back to data? Like
these are real 2026 agentic problems, but they come back to underlying principles around how we handle data and complexity that aren't new. Codebased
instrumentation, that's another one.
Gene, does that come back to pike and measuring? It sure does. This isn't even
measuring? It sure does. This isn't even an agent problem, right? This is a software hygiene problem. We have always had challenges when we've been doing engineering projects, especially where
we've been in a rush. It's been hard to be disciplined and measure. Making the
codebase agent ready is partly about being able to measure stuff and we should not forget it. I don't want to belabor this one too long. If you are an engineer and you're like, I need to be
able to make a contribution to AI, one of the simplest things you can do is just do the measuring. It's decades old.
it's not new, but figuring out how to say this is our current baseline performance maybe with our LLM chat window, maybe with our agent, whatever it is, and you can measure it effectively because you understand this
is the baseline. This is what latency looks like. This is what a good set of
looks like. This is what a good set of responses looks like and I have a nice golden data test set and I can true that up against what's in production. You
have done a tremendous service to your business and you don't get appreciated enough probably, but it's really important and it's not new. It's just
that we have to take it seriously because we are giving these autonomous agents a lot of power and we're not really measuring them if we're not disciplined. Problem number three in
disciplined. Problem number three in agentic coding work is around linting.
Now, if you don't know what linting is, I'm not talking about the stuff in your couch cushions. Linting is when you are
couch cushions. Linting is when you are doing static analysis of the code.
You're not making changes. You're just
checking it for small style issues, for inconsistencies, for potential bugs at runtime, and you're coming up with a report. Linting rules are how we make
report. Linting rules are how we make linting work. And one of the ways that
linting work. And one of the ways that you can detect issues with agentic code is by getting very very strict with your linting so that you are insistent on extremely clean code. This isn't new,
right? This is about enforcing simple
right? This is about enforcing simple structures. The factory team has this
structures. The factory team has this lengthy series of blog posts about all of the obsessive linting rules they have that basically put the code in a straight jacket and say it must adhere
to best practices all the time. Now
individual developers if they're the ones in charge of linting may say ah I don't know I'm tired. I don't really want to write all my linting rules. But
in a good healthy engineering organization you have some common core around linting where you say okay this is what good looks like for us. We're
going to insist on it. And that's
especially important when you have agents involved because the agents are by definition just trying to get the job done. They are lazy developers that are
done. They are lazy developers that are happy just to kind of throw it off their plates and not listen. And so if you don't have a strict linter that is going to go through and insist on simplicity,
you are going to be in trouble. Again,
not a new thing. It's just a common thing that we are now applying in the world of agents. An ancient engineering piece of wisdom, if you will. Problem
number four, how you handle multi-agent coordination. I've talked about this in
coordination. I've talked about this in other videos. We're converging around a
other videos. We're converging around a rule where we say planners and executors are the way to do longunning multi- aent coordination and that makes sense because we're not over complicating it.
And one of the things that Pike has called us to remember is hey you don't need to optimize something prematurely.
You don't need to optimize it if you can't measure it. And so when we've actually tried to overoptimize and over complicate and there are engineering teams at many orgs that try and do this I just I encourage folks to say you know
what let's not over complicate it build the simplest possible version of this agentic development pipeline and then we can always add more value by complexifying it if we really have to
but we don't need to optimize prematurely if we can't even measure whether it does the job yet again not new and if you're wondering why am I taking time to talk about what isn't new it's really simple I think consultants
often like to sell this as all new because it drums up business. I would
prefer to tell the truth and say these are ancient data engineering practices.
These are old software engineering best practices that we can apply in ways that are new to build these systems, but the practices and principles aren't that new. And I think that helps us with our
new. And I think that helps us with our change management. The last challenge is
change management. The last challenge is the hardest one. It's around
specifications and fatigue. What I find in practice is that teams really, really struggle with a skill of defining a spec clearly upfront. It's a lot of work.
clearly upfront. It's a lot of work.
There are some people who claim it can't be done or if it's so much work, we should just code the thing. I've seen
real speedups, but it does require you to be very precise and crystal clear in your thinking. And you also have to be
your thinking. And you also have to be very good at writing emails at the end.
And you have to be disciplined about not taking shortcuts. And so if you are
taking shortcuts. And so if you are going to give an agent a context window, you have to be disciplined about making sure your context graph is really clean so the agent can go search and get the
context it needs cleanly by navigating a hierarchy rather than just stuffing it all in the context window and hoping and praying because you're lazy. In other
words, we humans have to be less lazy if we want the agents to do good work for us. And I know that is counterintuitive
us. And I know that is counterintuitive because you are often sold a world where humans should just sit back and we just go and get coffee and then we're done.
That's not how it actually works. And
that's never how good engineering worked. It shouldn't be new. It
worked. It shouldn't be new. It
shouldn't be a surprise. And I think sometimes we're sold Asians as like labor savers. And that's just
labor savers. And that's just disingenuous. It's just not true. So why
disingenuous. It's just not true. So why
does all this hype exist? I went through five problems. I showed you how they're critical now in the world of Asians. I
showed you how they rest on old engineering best practices. I think if we messaged them that way, it would be useful to us. I think it would be easier to understand. I think that anthropic
to understand. I think that anthropic and open AI would have less issues communicating to developers. I think
it's something that Nemo Claw started to get right. Part of why as an industry we
get right. Part of why as an industry we have not done this well is that the chaos is worth a lot of money.
Consultants coming in and pedalling their wares and saying this study shows that it's really hard helps them earn business. And it is hard, right? But
business. And it is hard, right? But
it's hard in a way consultants typically don't help you with. It's hard in a roll up your sleeves, get into the code, co-build with me, dig in, help me understand the principles. And so many
times consultants don't want to get their shoes dirty, right? They they want to come in and just do a PowerPoint deck. Ah, they want to deliver a great
deck. Ah, they want to deliver a great deck and then move on. That's not how it works, right? If you're going to do real
works, right? If you're going to do real change management, if you're going to help engineers and product managers and designers figure out how their roles are changing because their whole jobs are changing, you can't do it with a
PowerPoint deck. It's not going to work
PowerPoint deck. It's not going to work that way. You have to go back and anchor
that way. You have to go back and anchor in things that we all understand and have built on. And that as I've showed, you can do that. And then you have to walk forward and say, here's how this applies today. That's why I've walked
applies today. That's why I've walked through these problems. That is much more specific than I have seen in any standard run-of-the-mill consultant deck, which so often like level up here
and they talk fluffily about how great AI is. It doesn't help you get the work
AI is. It doesn't help you get the work done. And this is what I think we're
done. And this is what I think we're missing when we look at launches like Nemo Claw because Nemoclaw as a launch is interesting. Nemoclaw as a play for
is interesting. Nemoclaw as a play for Nvidia definitely interesting. They're
trying to move beyond chips. But
Nemoclaw is a way of saying to the industry, you got this. you can figure this out. We've got good engineering
this out. We've got good engineering best practices that we can rely on and use to do real agent work. Now, that's
interesting. And that's something that I wish we did more of. And I think if we worked more on that piece as a discipline, we would have less need for
these tie-ups that we see between consulting firms and big companies like OpenAI and Anthropic. Because I think at the end of the day, in a sense, when you're outsourcing the change
management, you are losing control of the narrative. And one thing anthropic
the narrative. And one thing anthropic and open AI probably don't want to do is lose control of the AI change narrative in their target companies. It is already fraud enough. There are already enough
fraud enough. There are already enough people producing half-true rumors, sometimes completely false rumors about what AI can and cannot do, what AI will and will not do. And by the way, it is
both. I see lots of false rumors about
both. I see lots of false rumors about what AI can do. I see lots of false rumors about what it can't. I think it's helpful if we go back and we say this is just computing. We've known about
just computing. We've known about computing for a long time. We understand
how computing works. The fundamentals
aren't changing, but we have a new level of abstraction to put over the top and we should talk about it concretely and explain in a detailed way how our old
principles of engineering have actually evolved. And that's what I tried to do
evolved. And that's what I tried to do in this video. That's what I laid out for you so you could understand we're not doing new stuff here when we design Agentic Systems. We're relying on good engineering practices we've already had.
And in a way, a lot of what I'm doing on this channel is actually teaching good data engineering practices to a lot of people who didn't come up and do data engineering in school. Because it turns
out if you want to build these systems yourself, you have to know just enough about data engineering to build systems that work. And it turns out it's not
that work. And it turns out it's not scary. It turns out you can learn these
scary. It turns out you can learn these principles. You don't have to go and get
principles. You don't have to go and get a CS degree. And that's really empowering and that's really cool and that's really fun for me because I'll be honest, I didn't get a CS degree either.
I taught myself. I was building computers. I had fun. And I think what's
computers. I had fun. And I think what's interesting is LLMs are essentially a teachable moment. LLMs are giving so
teachable moment. LLMs are giving so many more people access to compute.
We're all coming to this with fresh eyes because when we look at change management in orgs, I've talked about engineers, but to be honest with you, it's not just engineers, right? It's
product managers, it's sales, it's CS.
Shopify was shocked when they first got cursor because there were so many CS people who wanted cursor, right? They
were coding under the desk. Coding under
the desk is a massive 2026 phenomenon that is by definition not engineering related. And if you want the coding
related. And if you want the coding under the desk to work, you got to make sure that we have a little bit of a sense of how best practices work. And if
we understand that, we're going to be able to take tools like Nemo Claw and actually put them to work effectively.
So hats off to Nvidia for believing in us a little bit, right? For saying we could roll our own. We can build stuff that works. We can understand how good
that works. We can understand how good data engineering best practices, old computer science best practices that age well are still applicable today, evolve them appropriately and tackle good
agentic engineering challenges. I want
more of that and I hope you do too.
Chips.
Loading video analysis...