Jensen Huang Nvidia Morgan Stanley TMT Conference
By NoRush Invest
Summary
Topics Covered
- Full-Stack Mastery Drives Hypergrowth
- Three AI Inflections Explode Compute
- OpenAI's Claw Sparks Agentic Revolution
- Compute Equals Revenue Equals GDP
- Physical AI Succeeds Agentic AI
Full Transcript
Wow.
>> No music. No walk on music. No
no no no no roaring applause.
>> Um I'm just I'm just saying that I'm I'm not I'm not used to coming to work in this way. This total silence.
this way. This total silence.
>> I'm just kidding.
>> There there were a lot of Taylor Swift comments along the way. So the crowd is ready.
>> Needs humor, too. Is humor allowed here?
>> Humor is very allowed. I made investment banking jokes yesterday, Jensen. But uh
thank you uh for being here for the last I think 25 27 years. You've been such a great uh supporter uh of this
conference. I think we sometimes become
conference. I think we sometimes become uh numb to the scale of the numbers and the transformation we're experiencing. I
don't think I'm the only one in this audience. I'm getting billions and
audience. I'm getting billions and trillions confused uh constantly. My
partner Mark Edlestone and I 27 years ago, we sat on a stage much smaller than this one on the Morgan Stanley trading
floor and we announced uh and introduced Nvidia and um and you to the Morgan Stanley Salesforce and uh believe it or
not, $48 million IPO, 1998 revenue n trailing revenue, $30 million. Um,
Jensen and his team, Colette, were so generous. Two years ago, you hosted our
generous. Two years ago, you hosted our board meeting in your headquarters. I
think you had just announced a $30 billion quarter uh in terms of revenue and then last week uh a $46 billion net
income quarter. So we've moved from
income quarter. So we've moved from years to quarters, from millions uh to uh billions. It's really amazing and
uh billions. It's really amazing and unprecedented scale and growth. Um and
then you've changed our lives. You've
changed our lives. And so I guess my question after that is what had to come together strategically, culturally, technically to deliver that
type of hyperrowth at scale. And the
scale is really astounding. And again,
thank you.
>> That's going to take 37 minutes and 13 seconds and slightly more. Uh, you know, obviously Nvidia wasn't built overnight.
It's taken us 33 years. I I sort of remember somehow that when we went public, our price was $13 and I just read here is 12. I overstated them. I
remembered it to be much more optimistically than it was actually. The
the company's valuation at the time, I think, was like $300 million.
And and uh Mark did such a good job.
Mark Edlesen did such a good job preparing all of our investors that that they really only had one question. It
was literally a one question IPO road show and the the question was when are you going out of business?
I'm not kidding. It's and that that's that exact that question uh is about as hard to answer as the one you just gave me.
Um well the the the answer is the answer is um as it turns out we started the company with the idea of creating a new computing platform a new way of doing
computing and not not that the old way was wrong. It's just that the new way um
was wrong. It's just that the new way um a new way is is essential to solve some unique problems and the type of things that we were extremely good at are
algorithms. algorithms because the inner loop of the software tends to be about 5% of the code but 99% of the compute
time and and back then the algorithms in the world of computers was quite rare and one of the most important algorithms was computer graphics the simulation of
light and how light travels through through space and so so uh computer graphics uh was used for uh things like animation movies of course at the time
that were founded the cover of I forget which magazine was you know Jurassic Park was there and so it was really it was during that time where computer
graphics was was becoming uh more more capable and we could simulate you know virtual reality with it and we applied it to creating a a a new industry uh
which did not exist at the time uh called uh uh video games and so 3D graphics uh was was modernized in my
time uh consumerized in my time and and the whole video game industry was created in my time. And when I say in my time, meaning it was it was Nvidia that Paul pulled it all together.
We the reason why we're so beloved in the video game industry and we're so deep in it still is because in a lot of ways we created the modern video game industry from the algorithms associated
with it, the libraries, you know, the uh in in the computer graphics industry without RTX. there would be nothing
without RTX. there would be nothing today without our contribution of all the algorithms that goes into all of the game engines. Uh you wouldn't be able to
game engines. Uh you wouldn't be able to enjoy the type of video games you enjoy today. So Nvidia has been deep in the
today. So Nvidia has been deep in the world of algorithms since day one 33 years ago. Now accelerated computing
years ago. Now accelerated computing requires uh what is described as a full stack meaning the architecture the chip
design the libraries that sit on top of it how it's integrated forwardly you know I I'm using that apparently there's this new idea called for deployed engineers or something like that
Nvidia's had devte engineers 33 years ago we deployed them into the world's video game industries and video game companies in game engines and we
integrate our technology into their game engine. Today, if you look at Epics's
engine. Today, if you look at Epics's Unreal Engine, Nvidia's technology is all over it and and uh uh you go into every game developer, Nvidia's technology is all over it. That's the
reason why all the all the games run best on Nvidia for good reason. That's
the reason why Nvidia is the world's largest game platform. Uh you you probably don't don't know this, but there's several hundred million active GeForce gamers in the world. many of
them turned into AI researchers. It is
because of GeForce GTX 580 that that uh uh you know that Ilia Sus Cescover and uh um uh Alex Kasheeski uh and Jeff
Hinton, it was Jeff that told him to go buy it to discover CUDA. And so so I the the first idea about Nvidia is that we're full stack company. The second
idea about our company and this is this is you know really old history that long many people might might not have been born yet but during that time the PC the
PC architecture was incompatible with today's computer graphics capabilities and we created some new technology called direct NVIDIA it was a way for
applications to directly communicate with our our APIs and uh uh we exposed it to to uh some very important companies it became direct X. Uh if you
look at the way that we uh communicate between us and the application uh that was completely revolutionary to bypass uh uh a whole bunch of software that
makes it slow by to make accelerated computing possible. Uh we uh introduced
computing possible. Uh we uh introduced the idea of virtualized uh frame buffer memory into system memory. It was
initially called AGP which then became PCI Express. um many of the system
PCI Express. um many of the system architecture had to be reinvented so that we could accommodate video games uh and uh 3D graphics in a PC. Well, that
same sensibility of both innovating the full stack to be integrated into in algorithms as well as changing the architecture of systems so that we could
create new computer systems led to that same sensibility expertise led to DGX1 which was the world's first AI supercomputer that had delivered you
know by hand to San Francisco here very close by to a company that eventually became open AAI and so so the the fundamental attitude, if you will, expertise, um,
how we see the world propagated in this way. It's literally
33 years. The company's entire culture is designed to be full stack. The
organization is designed to be full stack. The entire system uh is designed
stack. The entire system uh is designed to create new new stacks and new system architectures uh that allow us to do
this. Well, we started with of course
this. Well, we started with of course the the if you look at Nvidia's graphics cards, GeForce, it's a technology marvel. How it's integrated into the
marvel. How it's integrated into the operating system, how it's integrated into the system architecture, completely reinvented how computers were before.
Well, we have no trouble with that with DGX1. I have no trouble with that with
DGX1. I have no trouble with that with the first supercomputing cluster, which then went to Satia um uh for their first supercomput. And
you you might you might uh uh you know people noticed that Microsoft first supercomputer and Nvidia's supercomputer had exactly the same benchmark like down
to the you measure that you measure the performance of the system across all of these GPUs that that was about 10,000 GPUs or so. It was exactly the same performance. The reason for that is
performance. The reason for that is because we designed it and we delivered to to Azure cloud. Uh it was all based on Infiniband. It was all based on based
on Infiniband. It was all based on based on uh uh based on um uh Ampear a this was the A100 which became the first computer that open AAI used and so um
we're quite comfortable with this full stack full system approach and without being able to do that it is impossible to stay at the bleeding edge. It is
literally impossible to keep up with a company that's building not just one chip each year, but we're building an entire infrastructure each year because we we own the CPU. We revolutionized the
new way of designing CPUs. And you'll
see you'll see more examples of that. Uh
we re revolutionized the way we do CPUs, revolutionized the way we obviously do GPUs, connect them together using this thing called MVLink, which revolutionized the way you built
computers all together. uh connected
together with a new type of AI Ethernet called Spectrum X. We connected
everything together. Now we own the entire stack. We know all the chips
entire stack. We know all the chips inside. When you own the entire stack
inside. When you own the entire stack and you own all the chips inside, you could change it every single year. If
you don't own the entire stack and you don't own all the chips, it's hard to innovate every year. And the reason for that is because you're com you're connecting too many cats and dogs and
there's too much innovation to pull together once a year if you can't control it because it's a full stack problem. So that's how we got here. It's
problem. So that's how we got here. It's
amazing in the la in the last two years since you were here last and our board meeting we've sort of gone from generative AI uh models to reasoning and
now agentic and Satcha just finished a panel on the enterprise and at the enterprise level you know we're working with Microsoft the open AI XAI Gemini we
had Daario here from Anthropic the capabilities are extraordinary uh what does it mean around the size of that enterprise market, how is it changing
and how is it going to be adopted and how do you sort of see that playing out uh over the years because it's a big big topic at the conf.
>> Yeah, really good. Um literally in the last two years we went through three inflection points in AI. Uh the f first inflection point first of all the
technology sat there in plain sight for months. GPT3 sat there in plain sight
months. GPT3 sat there in plain sight for months until somebody wrote essentially a wrapper around it and turned it into chat GPT turned it into an API u made it a available and ease
easy to use by everybody. But the first inflection point was generative as you mentioned the ability to translate convert information from one form to
another form and auto reggressively generate tokens and and the the second but of course the problem with generative AI is that that it's prone to
hallucinate and and the reason for that is because not because there's something fundamentally wrong with the technology not because it didn't it didn't learn all the right things but because it's not grounded on contextual information.
it's not grounded on re relevant information and so so uh the second thing that happened was 01 and uh reasoning came about but behind 01 is
also grounding on research grounding on on truth and um the ability to have to combine generative with semantic um uh
we call it retrieval augmented generation um but basically conditional generation okay conditional generation meaning that what you're about to
generate depends on context and ground truth or whatever research or whatever it is. And so so the second generation
it is. And so so the second generation uh uh introduced reasoning, self-reflection, um the ability to self-correct because sometimes what comes out you know your
mouth you kind of wish that you pulled back and you go ah you know and so in the case of AI it has the ability to do that in real time and so 01 became much more uh grounded and the information
that was generated was more reliable. So
what happened um what came out as a curiosity and incredible excitement um and the and the tech industry jumping onto it because we realized what's about to what can happen the next phase of it
the usefulness of chatpt just skyrocketed um but the amount of tokens that it generated was much much more than the first generation maybe you know a 100 times more tokens the model is
maybe 10 times larger so it's probably something like a thousand times more compute so from from 01 one over Chad GBT call it a thousand times and then
because it was so useful maybe a million times more usage.
Okay. So the combination of usage um uh uh uh and and its usefulness um and groundedness allows us to to uh we saw
that next phase of growth. But in the end, in the end, what 01 did was it provided information
essentially um a chatbot that was uh much more much more factual.
It was informational and of course for many of us we use it for research and we use it all the time uh instead of searching you know our goal isn't to search our goal is to get answers and so
so uh chat GPT gave us that that was kind of the second inflction the inflection that we're we're seeing here also sat in plain sight for quite a long
time and um uh and it's basically the ability for AI to use files access files and use tools and so now it could reason it think it could use tools. it can
solve problems and uh it could do search, it could do planning and so um probably the the biggest phenomenon that's happening and if you're paying
attention to it I'm sure you are open claw is probably the single most important um uh release of software you know probably ever and if you look at
look at open claw uh and the adoption of it uh you know Linux took right some 30 years to reach this level uh openclaw saw in what is it 3 weeks has now
surpassed Linux.
It is now the single most downloaded opensource software in history and it took 3 weeks. If you look at the line, you know, even in semi- log, this thing
is straight up. It's vertical. It looks
like the it looks like the the y- axis.
I've never seen anything like it. Okay,
it look really looks like the y- axis.
And um uh and so what's happening now that you could give a problem statement um create start where the the prompt goes create you know the last prompt the
way you kind of think about it the last prompt was what is when is who is right that's the last prompt this now prompt
goes create do build write does that make sense so what's happened the last prompt was queries.
This prompt are actions. They're tasks.
Do something for me. And you describe it as, you know, expressively as you like as with with a lot of intention, you know, and let it let it infer or very
specific and it goes off and it just churns. It just thinks it goes off and
churns. It just thinks it goes off and it does research and it reads it reads a manual. Um, if it has to use a tool,
manual. Um, if it has to use a tool, it's never used it before, it reads the manual of the tool.
it goes off and studies what's on the web and and uh um and it you know applies the tools and performs the task.
Now I just said we went from one you know one generative prompt uh one generative uh response to now one that
is a thousand times more tokens and uh agents you know we call them at the company claws. These claws are now consuming what a million times more
tokens. They're running continuously in
tokens. They're running continuously in the background. We have a whole bunch of
the background. We have a whole bunch of claws in the company and uh they're all continuously running doing things for us, writing, developing tools, developing software and so now the question is the implication the amount
of compute in our company uh that we need is just skyrocketed. The amount of compute every company needs is skyrocketing. So in that context I think
skyrocketing. So in that context I think over the last few days it's come out certainly at Morgan Stanley as a user maximum bullish on tokens maximum
bullish on doing and creating it does require the compute you just mentioned and the question is around the financing and the capex around that to support
that extraordinary large compute. How
does it all get financed as you see it from a a sort of top of the ecosystem and how do the factory AI factory economics play out and evolve?
>> Yeah.
>> And so there's a couple of thoughts of that's really important. Uh remember I I I appreciate you using the word factory.
You know several years ago I I described that these new these data centers what people call data centers is not for storing data as in a data center. they
are producing tokens and so a a a facility a plant with the fundamental purpose of producing tokens is a factory it's an AI
factory at a time people said ah Jensen that sounds so grungy um you know it's clean and and uh uh but it produces tokens and people nobody likes to build
data centers because that you know who knows what what kind of return you're going to get on a data center but everybody loves building factories. And
the reason for that is because factories make money. And we now know for we now
make money. And we now know for we now know for certain that these factories directly generate tokens. And these
tokens are monetizable.
And the more compute you have, the more tokens you could produce. The more
tokens you produce, the greater your top line. We now know for certain, we now
line. We now know for certain, we now know for certain that companies revenues are directly correlated to compute. And
we know that for a fact because if Anthropic had three times more compute, their revenues will be three times higher. We know that. We know that
higher. We know that. We know that Anthropic is capac compute limited, factory limited. It's no different than
factory limited. It's no different than Mercedes being factory limited or any company being factory limited. And so if they had more compute in their factories, they will have higher revenues. If open had right now had more
revenues. If open had right now had more compute, they will have higher revenues.
And so so so the first thought is that compute equals revenues. Now the big idea of course compute equals GDP. That
we also know. Compute equals a country's GDP. And so so that's one thought. The
GDP. And so so that's one thought. The
second thought, the reason why Nvidia is so successful is because we engineer these systems full stack end to end and they're architected from the ground up
to generate tokens at incredible effectiveness. Nvidia's tokens per watt
effectiveness. Nvidia's tokens per watt is an order of magnitude an order of magnitude
ahead of the competition alternative tokens per watt. Now, what does that mean? Remember your factory has one
mean? Remember your factory has one gigawatt and if your tokens per watt is 10 times the alternative your revenues are 10
times the alternative.
For the very first time in history the computer architecture chosen in a factory in a company's factory must go
through CEO review. No question about it.
CEO review. No question about it.
that company only has a gigawatt or 2.3 gawatts for next year. If they put the wrong system inside, it will affect their revenues the next year. I promise
you that and we see it.
And so our architecture being so advanced now and pulling further and further ahead, you know, there was a probably one of the most exhaustive benchmarking done is by by a firm called
Semi analysis and they declared they declared Nvidia inference king.
Inference King inference is tokens per second, tokens per watt. It's about
generating tokens and tokens per dollar.
When our performance per watt or per anything is so much ahead of the competition or the alternative, our tokens per dollar is also the best, which means we're the cheapest tokens
you can produce today. Not even close, an order of magnitude better. And so
that's the second thought. The second
big idea for AI is AI is a factory because factories are power limited always. Doesn't matter how many plants
always. Doesn't matter how many plants you have. Each plant is still 100
you have. Each plant is still 100 megawatts or gigawatt. And therefore
tokens per watt is the single most important thing for the top line of companies.
And they have to make those decisions very very carefully. You know, it's no longer just about PowerPoint slides.
You're not going to go put $50 billion dollars down on somebody's PowerPoint slides.
>> So, the token demand is extraordinary.
As you just mentioned, you're see we're seeing it in your numbers, right? I
think I mentioned 46 billion in net income, but $70 billion in revenue.
>> You were going to ask me something about how to fund it.
>> Um, can I can I just tell you how to fund it? First of all,
fund it? First of all, um I I just told you I just told you uh the reason why you have to build these factories in the future is because you
either you just believe that one software is important and and so I hope this audience believes software is important. Software runs the world. The
important. Software runs the world. The
first thought, the second idea is this.
There will be no software in the future that's not agentic. Do you guys agree with that?
How could you have software that's dumb?
And so it is absolutely true that every software company will become an agentic company. They're they're they're going
company. They're they're they're going to simultaneously use um uh open models.
Okay. Uh open models mean the ones that they download themselves and they fine-tune themselves. They're also going
fine-tune themselves. They're also going to use closed models. The combination of all that just like we in our all of our companies, we have employees that we hire. We have employees that we're
hire. We have employees that we're grooming. We have we have contractors
grooming. We have we have contractors that we bring bring in. We have
specialists like yourself that we bring into the company to to do our work. Our
job is not to do the job. Our job is to have the job be done.
That's what every company does. And so
therefore, every company will realize that these AI models, some of it you rent, some of it you build. That's not
illogical. Just like biological workers, you will do that with digital workers.
And so every single software company in the in the future will no longer just rent tools, but they'll rent also experts to use the tools.
They'll not just rent tools, but rent experts that use those tools because their agents are going to be extremely good at using their specialized tools.
And so, so every single software company, what did the IT industry is a couple trillion dollars today? They're
tool renters in the future. They will of course have they'll rent agents that use those tools which means that the software industry in the future will be
much larger than the software industry of today. You pick your favorite
of today. You pick your favorite software companies and I can imagine a much much larger future for them. You
cadence is going to be much larger.
Synopsis is going to be much larger.
Seammens is going to be much larger in the future. But their business profile
the future. But their business profile will change because today there's they're basically a software licensing company. In the future, they will also
company. In the future, they will also rent tokens, specialized tokens, which also means that that $2 trillion industry today with no token consumption
in the future will be extraordinary token consumers. That's where that money
token consumers. That's where that money is going to come from. They're all of those software industry of today, not the enterprise companies, the IT industry alone is going to shift an
enormous is going to consume enormous amounts of tokens in the clouds and they're either open models or >> so that extraordinary token economy is
facing some constraints. So we've got memory constraints, we've got power permitting constraints. Uh I was in
permitting constraints. Uh I was in Texas with builders, we have electrician constraints. How do you see that playing
constraints. How do you see that playing out? You're Satcha raised it in the last
out? You're Satcha raised it in the last session. You're closer to it. And also
session. You're closer to it. And also
if it takes a little longer, is it still okay or is it really negative if we just if the if the cycle on building this extraordinary >> I love constraints. I love constraints and the reason for that is because in a
world of constraint, you have no choice but to choose the best.
You can't squander your choice. If the
if if the data center if the land power and shell is constrained, you're not going to randomly put something in there um just to try it out.
You're going to put something that you know for certain is going to deliver the tokens per watt that you know for certain is going to allow you from the moment you you secure the capacity,
we're going to be able to stand up an entire factory for you. We're the only company in the world that can come into your company and help you stand up an entire AI factory, you know. So, anybody
here that needs an AI factory and you need, you know, I I'm happy to help.
You call one person and that one person comes in and next thing you know, you you're in the AI factory business. Okay?
And and and and so so we have the expertise. We know the architecture
expertise. We know the architecture works. We know there's enormous demand
works. We know there's enormous demand for the architecture, you know, after you're done standing it up so we can help you get into business. And so when you're constrained that way, you have no
choice but to make the best choice.
because your revenues next year is directly correlated to it. And this is one of those questions now for all the CEOs that are in the cloud, their cloud service providers or software um
providers. If they make poor choices,
providers. If they make poor choices, this is no different than me choosing the wrong foundry. This is no different than me choosing, you know,
the wrong memory than the wrong anything because I have so little everything is so constrained.
If I choose poorly, I my revenues are affected. Everything's affected. And so
affected. Everything's affected. And so
they can't choose they can't choose poorly. The second thing is, you know,
poorly. The second thing is, you know, Nvidia's, as you mentioned, working at such a large scale, our supply chain. One of the things that we do with our money, of course, is to
secure our supply chain. One of the things that we do with our our capital is to secure our supply chain. So that
when Satia asked me to help him stand up a few gigawatts, the answer is no problem.
And the reason for that is I got all the memories. I got all the wafers. I got
memories. I got all the wafers. I got
all the co-ass. I've got all the packaging. I've got all the system s
packaging. I've got all the system s systems. I've got all the connectors. I
got all the cables. You know, everything from copper to multi-layer ceramic pack ceramic capacitors. Everything secured.
ceramic capacitors. Everything secured.
That's one of the reasons why Nvidia's balance sheet being strong is so strategic.
A strong balance sheet today is not only helpful, it's strategic.
And so you look at the amount of revenues we're shipping into. Just look
backwards and look at the capac the amount of supply chain capacity we had to go secure or that they have to believe. You
know, if you if you set up a factory, a plant, a DRAM plant, and I come in and say, you know what, go ahead and set up the DRAM plant because I'm going to I'm going to use it. That goes a long ways.
You might as well take that to the bank, as many of them have. And so and so and and so I I think the the the fact that
everything is scarce is fantastic for us and I think it does create duration which I think is extraordinarily powerful for you. I think just another layer which is the ecosystem. You're one
of the great, you are the greatest cash flow generating company in history and then you've taken that capital and really created it feels like stability,
diversity in the ecosystem. And so how do you think about that in both a financial and a strategic context as you build I think both duration and
durability in the entire ecosystem?
Yeah, you know, when when Mark Mark took me public, um I think it was probably, you know, a little bit less energetic than I was develop
delivering it just now, but I am fairly certain I said all the same things.
Nvidia has been building. Remember,
accelerated computing requires that I build an ecosystem. You can't just take code and decompile it and it works.
There's no such thing as a universal accelerated computing system.
Accelerated computing is by definition proprietary.
There is nothing about our architecture that is compatible with somebody else's.
It's just not. The instruction set's different. The architecture is
different. The architecture is different. The micro architecture is
different. The micro architecture is different. Everything is different. And
different. Everything is different. And
so we hide it underneath these all you know these things in such a way that that makes it makes you feel like and because of NVIDIA we accelerate everything from data processing
molecular dynamics fluid dynamics particle systems you know biology chemicals you know all the way to deep learning right you know long sequence spatial 3D you
name it right >> sounds like a five layer >> Sounds like a five layer cake >> it's a five layer cake right exactly but because we've been working on it so long, it looks like everything's accelerated, but it's not true. It's
because I did it one at a time, one domain at a time, that all of the important domains in the world are now fully accelerated. And so, the thing
fully accelerated. And so, the thing that we do on the on the on the supply chain side, our balance sheet is incredibly valuable because it provides security for our customers.
On the upstream side, I'm cultivating new ecosystems for the future. All these
AI natives that I'm investing in, the companies that I'm partnering with, these are expanding, extending the CUDA ecosystem. 100% of
everything that we do is on top of CUDA.
Every investment that we've made is on top of CUDA. Um uh so recently there was a question about about um are we going to invest a hundred billion dollars in
Open AI? Uh we just just just for
Open AI? Uh we just just just for everybody's um update, we uh we uh finalized our agreement. We're going to
invest $30 billion in open AI. Um I
think the opportunity to invest $und00 billion in open AI is probably not in not in the not in the cards. And the
reason for that is because they're going to go public. And so I'm fairly sure that if we provide the capacity they need, which the compute capacity they need, which we're ramping up hard to go
do, um the revenues will more than uh follow. And uh um uh and they're going
follow. And uh um uh and they're going to go public towards the end of the year. And so this might be the last time
year. And so this might be the last time we'll have the opportunity to invest in a consequential, you know, company like this. And and our $10 billion investment
this. And and our $10 billion investment in Anthropic uh probably will be the last as well. And and speaking of that, the the one of the things that that I wanted to make sure I told you guys this
time and something new that you probably haven't internalized. You see all the
haven't internalized. You see all the news, uh you probably haven't internalized some of the the really great work that we did last year, the
last year and a half or so, last year or so. Um we expanded we expanded
so. Um we expanded we expanded uh OpenAI's capacity from Azure
to OCI to now AWS.
We expanded OpenAI's reach of capacity to AWS. We're ramping AWS like mad.
to AWS. We're ramping AWS like mad.
we're ramping them as hard as we can so that open AI has accessed even more capacity. That's one. The second thing
capacity. That's one. The second thing that we did and this was this was a a a really really great outcome is we're now also working with Anthropic
and in the case of Anthropic we're expanding their capacity as aggressively as we can at AWS
as well as Azure. And so notice what we're doing in both in both they used to be one and one now they're kind of cross productd um but the amount of capacity
that we're going to bring online for them um you know supporting supporting their revenues their quality of revenues are so good uh we just need a lot more capacity for them so I think that this
is something that I I that that is somewhat new and of course the third thing that happened is a brand new AI labed into the world isn't that right? I
don't nobody's mentioned them. A brand
new lab came into the world and they're going to need multi a few million GPUs and that's MSL. And so that MSL is a net new on top of Meta. So we've we've
worked with Meta a long time. MSL is a net new on top of Meta. And so these three things happened, three new growth vectors. um OpenAI at AWS,
vectors. um OpenAI at AWS, Anthropic totally at both AWS and Azure and MSL
and um uh so so the the the uh our demand profile uh went from uh being
incredibly high to higher than that.
Speaking of So, uh, speaking of more than that, uh, there's Whimos everywhere. I want to walk my new dog
everywhere. I want to walk my new dog with the my new robot. Physical AI could be the next place. How does that take TAM and tokens to a whole another level?
>> Yeah, that's really great. Nvidia,
>> that's really great. AI is all the stuff that we're doing inside the building, but obviously obviously ultimately the largest industries are outside the building and and um that AI needs to be
needs to have physical awareness, physical understanding, you know, causality. You push a bottle, it falls
causality. You push a bottle, it falls over and understands gravity, understands collision, you know, understands inertia, understand those things. Okay? And um understands, for
things. Okay? And um understands, for example, object permanence. um I take this and I put it behind my chair in your mind you can't see it but you realize it hasn't disappeared okay so
object permanence thing things like that that affects um uh physical behavior and physical intelligence uh you know fairly
importantly and so so um I you you probably also don't know this that that Nvidia is the frontier of physical AI cosmos is the most downloaded physical
AI model in the Nvidia is also the frontier of autonomous AI. Two versions, autonomous
autonomous AI. Two versions, autonomous vehicle called Alpamo. Look it up.
Number one downloaded. And then the next one, Groot, human or robotics physical AI. We are at the frontier on all three
AI. We are at the frontier on all three of those. We're also at the frontier of
of those. We're also at the frontier of digital biology AI. Look up La Proutina.
Incredibly successful. La Prertina for digital biology. There's a whole bunch
digital biology. There's a whole bunch of other models. um uh Groot N2 is now number one most downloaded uh uh human
robotics uh model in the world and so so we are at the frontier on physical AI uh uh physics laws of physics multifysics
earth 2 we're at the frontier of uh of uh physical uh physical AI that is physical AI and AI physics and so these whole area of physical AI Nvidia defines
the frontier It is completely open. Uh we open it because we want to enable every company
uh new or old industry to be able to take advantage of this capability. And
we've got the whole stack and the necessary computers for you to uh advance the AI for your own use. uh as
well as deploy it uh inside a robot, inside a plant, at the edge, at a radio tower. Um deploy it everywhere. This is
tower. Um deploy it everywhere. This is
the next frontier. Uh in two years time, we're going to be largely done talking about Agentic AI because we're all going to be using it. In two years time when and if you invite me back again um
>> every year, every year, every year >> we're going to be talking about all these new companies. Of course, we announced a very important one, a co-inovation lab with Lily. Um, there
will be others, but uh uh you know this in order to set up Lily's AI factory, unless you are unless you have the capabilities on Nvidia and this full software stack and the capabilities of
all the model and the expertise in that digital biology domain, how would you even do it? And so uh we're quite the things that we are building in the next couple years you'll see uh really come
to four and uh we're going to be talking about physical AI for you know starting next couple two three years and for a decade. So the speed of innovation and
decade. So the speed of innovation and the pace that you're operating in is truly extraordinary. So at the beginning
truly extraordinary. So at the beginning of the week my partner Joe Moore made uh Nvidia his number one pick. Is that
right?
>> It's his number one pick.
>> Thank you.
>> Uh uh thank you. Well, thank you.
>> Uh, good timing, Joe.
>> 30 33 years later.
>> Uh, how do you think about the stock? Do
you do you think about the stock? Do you
do you have perspectives on it? Uh,
you're so extraordinarily important and busy around driving all this innovation for in essence everything that's going on with 3,500 attendees and we had 40
trillion dollars of market cap here. How
do you how do you think about that?
>> Um, well, you know, of course I care about the stock. I I care about shareholders. I care about I care about
shareholders. I care about I care about our employees. I care about all of you.
our employees. I care about all of you.
Um, I the my and you might be referring to we just had the best earnings of earnings in the history of earnings. Is
that what you were saying? I I I think the I think somebody actually told me that that this might be the the single best print in the h in the history of
humanity and and I I I said it must be only only you know recorded humanity.
I'm sure somebody uh had better returns.
Uh but but anyways uh we did we had a very good quarter. Listen, you can't hold the stock back.
You can't hold it back. And the reason for that is very simple.
Compute equals revenues for companies.
In the in the future, every single company will need compute for revenues.
I'll just make that prediction for now.
Every single company will need compute for revenues. And the reason for that is
for revenues. And the reason for that is because compute translates to intelligence which translates to your digital workforce which translates to your revenues.
I am certain compute equals revenues.
I'm certain also that compute equals GDP. Therefore, every country will have
GDP. Therefore, every country will have it because not one country in the future will say, "Guess what? You know, we're going to opt out on intelligence.
We've got we got I don't know what we got, but we we don't need intelligence.
That's the one thing we don't need."
Okay? And so, if you need intelligence, you're going to need digital. You're
going to need AI. You're going to need compute. And so, compute equals GDP. I
compute. And so, compute equals GDP. I
know that for certain. I also know that we're at the beginning of this journey.
And I I see crystal clearly exactly how it's going to get funded. Um we know for a fact that all the CSPs took all of their capex and they converted to
generative agentic systems AI systems because it helps search because it helps shopping, because it helps ads, because it helps social because it helps
literally every single internet service in the world has been reinvented into generative AI. So they could take a 100%
generative AI. So they could take a 100% the entire internet industry could take a 100% of their capex and make it AI because it's better. We we've proven it to be better. Meta has proven to be
better. Google has proven to be better.
better. Google has proven to be better.
AWS has proven to be better. And so you could now take your capex and convert to this. Number two, I just said the entire
this. Number two, I just said the entire software industry will be token driven.
The entire software industry. You pick
your favorite software company and I can show you exactly how they're going to be token driven. And that token drip that
token driven. And that token drip that that token um you take your favorite you know software company their token uh
will be either produced by themselves which needs compute or they could be uh resold from anthropic or open AI and
that needs compute and so what that says for the first time is the entire IT industry will have to be fueled by compute that's exactly where all this is going to come from trillions of dollars of it
and we're at the beginning of that. So
that's my prediction.
>> Thank you Jensen for making history at this conference 27 years. Thank you.
Loading video analysis...