Jensen Huang: Nvidia's Future, Physical AI, Rise of the Agent, Inference Explosion, AI PR Crisis
By All-In Podcast
Summary
Topics Covered
- Disaggregated Inference Powers AI Factories
- Inference Factories Yield Lowest Token Costs
- Agentic AI Redefines Personal Computing
- Physical AI Unlocks 50 Trillion Market
- Robots Unlock Universal Prosperity
Full Transcript
Special episode this week. We've preempted the weekly show. And there's only three people we preempt the show for, President Trump, Jesus, and Jensen. And I'll let you pick which order we do that. But what an amazing run you've had and a great event every industry is here. Every tech company is here. Every AI company is here. Incredible. Incredible.
If you were building a global financial system from first principles today, you wouldn't build it on 50 year old legacy rails. You'd build Airwallex, one AI-native platform for global accounts, cards, and payments. It's designed to make the entire world feel like a local market. Others are bolting AI onto broken infrastructure, but Airwallex was built for the intelligent era from day one. Stop paying lazy tax and start building the future at airwallex.com/allin. Airwallex. Build the future.
And one of the great announcements of the past year has been Groq. When you made the purchase of Groq, did you realize how insufferable Chamath would become? I had an inkling that... We're his friends. We have to deal with him every week. I know it. You had to deal with him for the six week close. I know it. Two weeks, two weeks. It's all coming back to me now.
It's making me rather uncomfortable. The thing is, many of our strategies are presented in broad daylight at GTC years in advance of when we do it. Two and a half years ago, I introduced the operating system of the AI factory, and it's called Dynamo. Dynamo, as you know, is a piece of instrument, a machine that was created by Siemens to turn essentially water into electricity. And Dynamo powered the factory of the last industrial revolution, so I thought it was the perfect name for the operating system of the next industrial revolution, the factory of that. And so inside Dynamo, the fundamental technology is disaggregated inference. Jason, I know you're super technical. I know it. I'll let you take this one. Go ahead and define it for the audience. I don't want to step on you. Yeah, thank you. I know you wanted to jump in there for a second, but it's disaggregated inference, which means the pipeline, the processing pipeline of inference is extremely complicated. In fact, it is the most complicated computing problem today. Incredible scale, lots of mathematics of different shapes and sizes. And we came up with the idea that you would change, you would disaggregate parts of the processing such that some of it can run on some GPUs, rest of it can run on different GPUs. And that led to us realizing that maybe even disaggregated computing could make sense, that we could have different heterogeneous nature of computing. That same sensibility led us to Mellanox. You know, today Nvidia's computing is spread across GPUs, CPUs, switches, scale-up switches, scale-out switches, networking processors. And now we're going to add Groq to that, and we're going to put the right workload on the right chips. You know, we just really evolved from a GPU company to an AI factory company. I mean, I think that was probably the biggest takeaway that I had. You're seeing this fundamental disaggregation where we've gone from a GPU and now you have this complexion of all these different options that will eventually exist. The thing that you guys said on stage or you said on stage was, I would like the high-value inference people to take a listen to this and 25% of your data center space, you said, should be allocated to this Groq LPU GPU combo. We should add Groq to about 25% of the Vera Rubins in the data center. So can you tell us about how the industry looks at this idea of now basically creating this next generation form of disaggregated, pre-filled, decode, desag, and how people, do you think, will react to it? Yeah, and take a step back and at the time that we added this, we went from large language model processing to agentic processing. Now, when you're running an agent, you're accessing working memory, you're accessing long-term memory, you're using tools, you're really beating up on storage really hard. You have agents working with other agents. Some of the agents are very large models, some of them are smaller models, some of them are diffusion models, some of them are autoregressive models. And so there's all kinds of different types of models inside this data center. We created Vera Rubin to be able to run this extraordinarily diverse workload. My sense is, and so we added what used to be a one-rack company, we now add a four-more rack. Right. So Nvidia's TAM, if you will, increased from whatever it was to probably something call it, you know, 33%, 50% higher. Now, part of that 33% or 50%, a lot of it's going to be storage processors, it's called Bluefield. Some of it will be, a lot of it, I'm hoping, will be Groq processors, and some of it will be CPUs. And they're all, and a lot of it's going to be networking processors. And so all of this is going to be running basically the computer of the AI revolution, called Agents. Right. The operating system of modern industry. What about embedded applications? So, you know, my daughter's teddy bear at home wants to talk to her. What goes in there? Is it a custom ASIC or does there end up becoming much more kind of a broader set of TAM with developing tools that are maybe different for different use cases at the edge and in an embedded application space? We think that there's three computers in the problem at the largest scale when you take a step back. There's one computer that's really about training the AI model, developing, creating the AI. Another computer for evaluating it. Depending on the type of problem you're having, like for example, you look around, there's all kinds of robots and cars and things like that. You have to evaluate these robots inside a virtual gym that represents the physical world. So it has to be software that obeys the laws of physics. And that's a second computer, we call that Omniverse. The third computer is the computer at the edge, the robotics computer. That robotics computer, one of them could be a self-driving car, another one's a robot, another one could be a teddy bear. Little tiny one for a teddy bear. One of the most important ones is the one that we're working on that basically turns the telecommunications base stations into part of the AI infrastructure. So now all of the, It's like a two trillion dollar industry. All of that in time will be transformed into an extension of the AI infrastructure. And so radio, radio's will become edge devices, factories, warehouses, you name it. And so these three basic computers, all of them are going to be necessary. Jensen, last year I think you were ahead of the rest of the world in saying inference isn't going to a thousand X. Just last year? Brad, you hurt my feelings. It isn't going to a one million X, it's going to a one billion X. Yeah, and I think people at the time thought it was pretty hyperbolic because the world was still focused on pre-scaling, on training. Here we are, now inference has exploded where inference constrained. You announced an inference factory that I think is leading edge that's going to be 10X better in terms of throughput to the next factory. But yet if I listen to what the chatter is out there, it's that your inference factory is going to cost 40 or 50 billion and the alternatives, the custom ASICs, AMD others are going to cost 25 to 30 billion and you're going to lose share. So why don't you talk to us, what are you seeing, how do you think about share and does it make sense for all these folks to pay something that's a 2X premium to what others are marketing? The big takeaway, the big idea is that you should not equate the price of the factory and the price of the tokens, the cost of the tokens. It is very likely that the 50 billion dollar factory, and in fact I can prove it, that the 50 billion dollar factory will generate for you the lowest cost tokens. And the reason for that is because we produce these tokens at extraordinary efficiency. Ten times, the difference between 50 billion, now it turns out 20 billion is just land, power and shell, right? And then on top of that, you have storage anyways, networking anyways, you got CPUs anyways, you got servers anyways, you got cooling anyways. The difference between that GPU being 1X price or half-X price is not between 50 billion and 30 billion. Pick your favorite number, but let's say between 50 billion and 40 billion. That is not a large percentage when the 50 billion dollar data center is actually 10 times the throughput. That's the reason why I said that even for most chips, if you can't keep up with the state of the technology and the pace that we're running, even when the chips are free, it's not cheap enough. Can I just ask a general strategy question? I mean, you're running the most valuable company in the world. This thing is going to do 350 plus billion of revenue next year, 200 billion of free cash flow, it's compounding at these crazy rates. How do you decide what to do? How do you actually get the information, I mean, it's famous now these sort of emails that people are meant to send you, but how do you really decide to get an intuition of how to shape the market, where to really double down, where to maybe pull back, where to actually go into a greenfield? How does that information get to you? How do you decide these things? In the final analysis, that's the job of the CEO. And our job is to define the strategy, define the vision, define the strategy. We're informed, of course, by amazing computer scientists, amazing technologists, great people all over the company, but we have to shape that future. Well, part of it has to do with, is this something that's insanely hard to do? If it's not hard to do, we should back away from it. And the reason for that is if it's easy to do, obviously lots of competitors. Is this something that has never been done before that's insanely hard to do and that somehow taps into the special superpowers of our company? And so I have to find this confluence of things to do that meets this standard. And in the end, we also know that a lot of pain and suffering's going to go into it. There are no great things that are invented because it was just easy to do and just like first try, here we are. And so if it's super hard to do, nobody's ever done it before, it's very likely that you're going to have a lot of pain and suffering, and so you better enjoy it. Can you just look at maybe three or four of the more long-tail things you announced and just talk about the long-term viability of whether it's the data centers in space or whether it's what you're trying to do with ADAS in autos or you know what you're trying to do on the biology side? Just give us a sense of like how you see some of these curves inflecting upwards in some of these longer-tail businesses. Excellent. Physical AI, large category. We believe, and I just mentioned we have three computing systems, all the software platforms on top of it. Physical AI as a large category, it's technology industry's first opportunity to address a 50 trillion dollar industry that has largely been void of technology until now. And so we need to invent all of the technology necessary to do that. I felt that that was a 10-year journey. We started 10 years ago, we're seeing inflecting now. It is a multi-billion dollar business for us, as close to 10 billion dollars a year now, and so it's a big business and it's growing exponentially. So that's number one. I think in the case of digital biology, I think we are literally near the chat GPT moment of digital biology. We're about to understand how to represent genes, proteins, cells, we already know how to understand chemicals, and so the ability for us to represent and understand the dynamics of the building blocks of biology, that's a couple of two, three, five years from now. In five years' time, I completely believe that the healthcare industry where digital biology is going to inflect. And so these are a couple of the really great ones, and you could see they're all around us. Agriculture inflecting now? No question. Jensen, I want to take you from the data center to the desktop. The company was built in large part on hobbyists, video gamers and all those graphics cards in the beginning. And you mentioned in front of, I think, 10,000 people here, just Claude, Open Claude, ClaudeCode, and what a revolution agents have become and specifically the hobbyists who are really where a lot of energy we see a lot of the innovation breaks want desktops. You announced one here, I believe it's the Dell 6800. This is a very powerful workstation to run local models, 750 gigs of RAM. Obviously the Mac studio sold out everywhere. In my company, we're moving to ClaudeCode everything. Freeberg just got Claudepilled. You got Claudepilled, I understand, and you're obsessed with these. What does this from the streets movement of creating open-source agents and using open-source on the desktop mean to you and where is that going? So great. First of all, let's take a step back. In the last two years, we saw basically three inflection points. The first one was generative. Chat GPT brought AI to the common everybody to our awareness. But the fact of the matter is the technology sat in plain sight months before GPT. It wasn't until Chat GPT put a user interface around it, made it easy for us to use, that generative AI took off. Now generative AI, as you know, generates tokens for internal consumption as well as external consumption. Internal consumption is thinking, which led to reasoning. O1 and O3 continue that wave of Chat GPT, grounded information, made AI not only answer questions, but answer questions in a more grounded way, useful. We started seeing the revenues and the economic model of OpenAI start to inflect. Then the third one was only inside the industry that we saw ClaudeCode, the first agentic system that was very useful, really revolutionary stuff. But ClaudeCode was only available for enterprises. Most people outside never saw anything about ClaudeCode until Open Claude. Open Claude basically put into the popular consciousness what an AI agent can do. That's the reason why Open Claude is so important from a cultural perspective. Now the second reason why it's so important is that Open Claude is opened, but it formulates, it structures a type of computing model that is basically reinventing computing altogether. It has a memory system, short-term memory, file system, it has skills. Did you say skills or scales? Skills. Oh, skills. It does have scales theoretically, yeah. Skills. So the first thing, you know, it has resources, it manages resources, it does scheduling, right? It could spawn off agents, it could decompose a task and cause and solve problems, so it does scheduling. It has IO subsystems, it could, input, output, connect to WhatsApp, and also it has an API that allows it to run multiple types of applications, called skills. These four elements fundamentally define a computer. Right? And therefore, what do we have? We have a personal artificial intelligence computer for the very first time. Open source. It's open source. It runs literally everywhere. And so this is now the, this is basically the blueprint, the operating system of modern computing. And it's going to run literally everywhere. Now, of course, one of the things that we had to help it do is whenever you have agentic software, you have to make sure that agentic software has access to sensitive information, it can execute code, it could communicate externally. We have to make sure that all of it has to be governed, all of it has to be secure, and that we have policies that give these agents two of the three things, but not all three things at the same time. And so the governance part of it, we contributed to Peter, Peter Steinberger was here, and so we've got a mount of great engineers working with him to help secure and keep that thing so that it could protect our privacy, protect our security. Jensen, that paradigm shift makes some of the AI legislation that has passed around the country to regulate AI and a lot of the proposed legislation effectively moot, doesn't it? Can you just comment for a second on how quickly the paradigm shift kind of obviates a lot of the models for regulatory oversight of AI, which is becoming a very hot topic in politics right now? Well, this is, this is the part that, you know, we just, with policy makers, we need to, we always need to get in front of them, and Brad, you do a great job doing this, we need to get in front of them and inform them about the state of the technology, what it is, what it is not. It is not a biological being, it is not alien, it is not conscious. It is computer software. Exactly. And we don't say things like we don't understand it at all. It is not true, we don't understand it at all. We understand a lot of things about this technology. And so, I think one, we have to make sure that we continue to inform the policymakers and not affect, not allow doomerism and extremism to affect how policymakers think and understand about this technology. However, we still have to recognize the technology's moving really fast and don't get policy ahead of the technology too quickly. And the risk that we run as a nation, our greatest source of national security concern with respect to AI is that other countries adopt this technology while we are so angry at it or afraid of it or somehow paranoid of it that our industries, our society don't take advantage of AI. And so I'm just mostly worried about the diffusion of AI here in the United States. Can you just double-click, if you were in the seat, in the boardroom of Anthropic over that whole scuttlebutt with the Department of War, it sort of builds on this idea of people didn't know what to think, it's sort of added to this layer of either resentment or fear or just general mistrust that people have sometimes at the software levels of AI. What would you think you would have told Dario and that team to do maybe differently to try to change some of this outcome and some of this perception? The first thing that I would say about Anthropic is first of all the technology's incredible. We are a large consumer of Anthropic technology. Really admires their focus on security, really admires their focus on safety. The culture by which they went about it, the technology excellence by which they went about it, really fantastic. I would say that the desire to warn people about the capability of the technology is also really terrific. We just have to make sure that we understand that the world has a spectrum and that warning is good, scaring is less good. Because this technology is too important to us. And I think that it is fine to predict the future, but we need to be a little bit more circumspect, we need to have a little bit more humility that in fact we can't completely predict the future. And to say things that are quite extreme, quite catastrophic that there's no evidence of it happening could be more damaging than people think. And of course we are technology leaders. There was a time when nobody listened to us, but now because technology is so important in the social fabric, such an important industry, so important to national security, our words do matter, and I think we have to be much more circumspect, we have to be more moderate, we have to be more balanced, we have to be more thoughtful. Well, I you know I would nominate you. I think the industry's got to get together. 17% popularity of AI in the United States. I mean we see what happened at nuclear, right? We basically shut down the entire nuclear industry and now we have 100 fission reactors being built in China and zero in the United States. We hear about moratoriums on data centers, so I think we have to be a lot more proactive about that. But I wanted to go back to this agentic explosion that you're seeing inside your company, the efficiencies, the productivity gains inside your company. There's a lot of debate whether or not we're seeing ROI, right? And you and I entering into this year, the big question was are the revenues going to show up? Are the revenues going to scale like intelligence? And then we had this kind of Oppenheimer moment of five, six billion dollar month by Anthropic in February. Do you think as you look ahead, you announced a trillion dollar visibility into a trillion dollars of just Blackwell and Vera Rubin over the course of the next couple years? When you see this happening at Anthropic and OpenAI, do you think we're on that curve now where we're going to see revenues scale in the way that intelligence is scaling? When you look around, I'll answer this in a couple of different ways. When you look around this audience, you will see that Anthropic and OpenAI is represented here, but in fact everyone, 99% of everything that is here is all AI and it's not Anthropic and OpenAI. And the reason for that is because AI is very diverse. I would say that the second most popular model as a category is open models. Open weights, open source. OpenAI is number one, open source is number two, very distant third is Anthropic, and that tells you something about the scale of all of the AI companies that are here. And so it's important to recognize that. Let me come back and say couple things. One, when we went from generative to reasoning, the amount of computation we needed was about 100 times. When we went from reasoning to agentic, the computation is probably another 100 times. Now we're looking at in just two years, computation went up by a factor of 10,000X. Meanwhile, people pay for information, but people mostly pay for work. Talking to a chatbot and getting an answer is super great. Helping me do some research, unbelievable. But getting work done, I'll pay for. And so that's where we are. Agentic systems get work done. They're helping our software engineers get work done and so then you take that, you got 10,000X more compute, you get probably at this point 100X more consumption now, and we haven't even started scaling yet. We are absolutely at a million X. Which is I think a great place to talk about the number of engineers you have at the company, 20, 30,000 at the company, something like that. We have 43,000 employees. I would say 38,000 are engineers. The conversation we've had on the pod a number of times is oh my god look at the token usage in our companies. It is growing massively. And some people are asking hey when I join a company how many tokens do I get because I want to be an effective employee. And you postulated I believe during your two and a half hour keynote, pretty long keynote, well done, that you were spent... If it was well done it would be shorter. I just wanted to... You didn't have time to write a short keynote. So you guys so you guys know, so you guys know there is no practice and so it's a grip it and rip it. I love it. And so I just want to let you know I was writing the speech while I was giving the speech. Okay so never know. I apologize. Back of the envelope math 75,000 in tokens for each engineer, something like that. So are you spending in NVIDIA a billion or two billion dollars on tokens for your engineering team right now? We're trying to. Let me give you the thought experiment. Let's say you have a software engineer or AI researcher and you pay them 500,000 dollars a year. We do that all the time. Okay this is happening all of the time. That 500,000 dollar engineer at the end of the year I'm going to ask them how many tokens did you spend? And that person said 5,000 dollars. I will go ape something else. If that person if that 500,000 dollar engineer did not consume at least 250,000 dollars worth of tokens I am going to be deeply alarmed. Okay and this is no different than one of our chip designers who says guess what I'm just going to use paper and pencil I don't think I'm going to need any CAD tools anymore. This is a real paradigm shift to start thinking about these all-star employees. It almost reminds me of what we learned in the MBA when LeBron James started spending a million dollars a year just on his health of his body like and maintaining it. Here he is at age 41 still playing. It really is hey if these are incredible knowledge workers why wouldn't we give them superhuman abilities. That's exactly it. Where does that go if we extrapolate out two or three years from now, what is the efficiency of that all-star at an NVIDIA and what they're able to accomplish, what do they look like? Well first of all things that wow this is too hard, that thought is gone. This is going to take a long time, that thought is gone. We're going to need a lot of people, that thought is gone. This is no different than in the last industrial revolution somebody goes boy that building really looks heavy. Nobody says that anymore. Wow that mountain looks too big. Nobody says that. Everything that's too big, too heavy, takes too long, those thoughts, those ideas are all gone. You're reduced to creativity. What can you come up with. That's right. Exactly. Which means now the question is how do you work with these agents? Well it's just a new way of doing computer programming. In the past we code. In the future we're going to write ideas, architectures, specifications. We're going to organize teams. We're going to give them we're going to help them define how to evaluate the definition of good vs bad. What does it look like when something is a great outcome, how to iterate with you, how to brainstorm. That's really what you're looking for, and I think that every engineer is going to have 100 agents. Back to the PR problem the industry has right now, you have executives like David Freeberg with Ohalo who's looking at literally taking through the use of technology, your technology in AI, the number of calories produced and making high quality calories what is the factor you think you can bring the cost down Freeberg and what impact does this vision have for what you're doing? I zero-shot genomic modeling. And it works. And then you have that moment and you're like holy honestly like and that's after people are replacing entire enterprise software stacks in a night. I did something in 90 minutes I was telling the guys about replace the whole software stack and like a whole bunch of workload 90 minutes on Claude ran in this agentic system built the whole thing deployed it and we got we it was on a Sunday night. On a Sunday night 10 p.m. I was done at 11:30 I went to bed. As the CEO you replace! Yeah everyone on my management team had to do a similar exercise over the weekend. What we saw on Monday I was like it's over. But the technical stuff the science stuff we did something in 30 minutes using auto research and I'd love your view on auto research and what that tells us about how far we still have to go in terms of efficiency. But using auto research and a chunk of data something was published internally that we said oh my god and that would normally be a PhD thesis that would take seven years it would be one of the most celebrated PhD theses we've ever seen in this field and it would be in the Journal of Science and it was done in 30 minutes on a desktop computer running on auto research with all the data we just ingested. We got it on Friday we're like hey let's try it boot it up go into GitHub download auto research and ran it. And you see everyone's face just go like and then the potential of what this is unlocking for us is like the kind of thing that would take seven years and it happened in 30 minutes and we're experiencing it in genomics and we're like this is unbelievable. So I think like the acceleration is widening the aperture for everyone in a way that like you didn't imagine a few years ago. But just going back to the auto research point can you just comment on what you think about the fact that this thing got published with 600 lines of code in a weekend and the capacity that it have to run locally and achieve what it can achieve with all of these diverse data sets and what that tells us about the early stage we are in terms of optimization on algorithms and hardware to unlock everything. The fundamental reason why Open Claude is so incredible, number one, is its confluence, its timing with the breakthroughs in large language model. Its timing was perfect, it was impeccable. Now in a lot of ways Peter wouldn't have come up with it probably if not for the fact that Claude and GPT and Chat GPT have reached a level that is really very good. It is also a new capability that allows these models to tool use. The tools that we've created over time, web browsers and Excel spreadsheets and, in the case of chip design, Synopsys and Cadence, Omniverse and Blender and Autodesk, all of these tools are going to continue to be used. There's some people who say that the enterprise IT software industry is going to get destroyed. Let me give you the alternative view. The enterprise software industry is limited by butts and seats. It's about to get 100 times more agents banging on those tools. There're going to be agents banging on SQL, they're going to be agents banging on vector databases, agents banging on Blender, agents banging on Photoshop. And the reason for that is because those tools first of all do a very good job. Second, those tools are the conduit between us. In the final analysis when the work is done, it has to be represented back to me in a way that I can control. And I know how to control those tools. And so I need everything to be put back into Synopsys. I want everything to be put back into Cadence, because that's how I control it, that's how I ground truth. Let me ask you a question about open source. So we have these close source models they're excellent. We have these open weight models many of the Chinese models are incredible, absolutely incredible. Two days ago, you may not have seen this because you were busy on stage, but there was a training run that happened in this crypto project called BitTensor, Subnet 3, they managed to train a 4 billion parameter llama model totally distributed with a bunch of people contributing excess compute but they were able to do it statefully and manage a training run, which I thought was like a pretty crazy technical accomplishment. Yeah, our modern version of folding at home. Exactly. So what do you think about the end stage of open source? Do you see this decentralization of architecture as well and decentralization of compute to support open weights and a totally open source approach to making sure AI is broadly available to everyone? I believe we fundamentally need models as a first-rate first-class product, proprietary product, as well as models as open source. These two things are not A or B, it's A and B. There's no question about it. And the reason for that is because models is a technology not a product. Models is a technology not a service. For the vast majority of consumers, the horizontal layer, the general intelligence, I would really really love not to go fine tune my own. I would really love to keep using Chat GPT. I love to use Claude. I love to use Gemini. I love to use X. And they all have their own personalities as you know, which just kind of depends on my mood and depends on what problem I'm trying to solve. You know, I might you know do it on X or I might do it on Chat GPT. And so that segment of the industry is thriving, it's going to be great. However, all these industries, their domain expertise, their specialization has to be channeled, has to be captured in a way that they can control and that it can only come from open models. The open model industry, we're contributing tremendously into, it is near the frontier and quite frankly globally even if it reaches the frontier, I think that products as a service, world-class products as a model as a product is going to continue to thrive. Every startup we're investing in now is open source first and then going to the proprietary model. Yeah, and the beautiful thing is because you have a great router you connect it to by on first day every single day, you're going to have access to the world's best model, and then it gives you time to cost reduce and fine-tune and specialize so you're going to have world-class capability out the shoot every single time. Jensen, can I just ask a question? Nobody wants the US to win the global AI race more than you. But a year ago, the Biden-era diffusion rule really was anti-American diffusion of AI around the world. So here we are a year in with the new administration. Give us a grade. Where are we in terms of global diffusion and the rate at which we're spreading US AI technology around the world? Are we an A? Are we a B? Are we a C? What's working? What's not working? Well first of all President Trump wants American industry to lead. He wants American technology industry to lead. He wants American technology industry to win. He wants us to spread American technology around the world. He wants United States to be the wealthiest country in the world. He wants all of that. At the current moment as we speak, NVIDIA gave up a 95% market share in the second largest market in the world and we're at 0%. I know. President Trump wants us to get back in there and and to get licensed for the companies that we're going to be able to sell to. We've had many companies who have requested for licenses we've applied for licenses for them and we've got approved licenses from Secretary Lutnick. Now we've informed the Chinese companies and many of them have given us purchase orders. And so we're going to start cranking up our supply chain again to go ship. I think at the highest level, Brad, I think one of the things that we should acknowledge is this: our national security is diminished when we don't have access to miniature motors, rare earth minerals, it's diminished when we don't control our telecommunications networks, it's diminished when we can't provide for sustainable energy for our country. It is fundamentally diminished. Every single one of these industries is an example of what I don't want the AI industry to be. Right. When we look forward in time and we say what do we want what does it look like when American technology industry, American AI industry leads the world, we can all acknowledge that there is no way that AI models is one universally. We can all acknowledge that that is an outcome that makes no sense. However, we can all imagine that the American tech stack from chips to computing systems to the platforms are used broadly by the world where they build their own AI, they use public AI, they use private AI whatever and they can build their applications in their society. I would love that the American tech stack is 90% of the world. Yeah, I would love that. The alternative, if it looks like solar, rare earth, magnets, motors, telecommunications, I consider that a very bad outcome for national security. Agreed. Yeah. How much are you monitoring the situation with the conflicts around the world right now and how much does it worry you, Jensen? So China and Taiwan and then helium availability coming out of the Middle East, I understand can be a supply chain risk to semiconductor manufacturing. How much do these situations worry you? How much are you spending on them? First of all I think in Middle East I have we have 6,000 families there. We have a lot of Iranians at NVIDIA and their families are still in Iran. And so we have a lot of families there. First thing is they're quite anxious, they're quite concerned, quite scared. We're thinking about them all the time. We're monitoring and keeping an eye on them all the time. They have 100% of our support. I've been asked several times are we still considering being in Israel? We are 100% in Israel. We are 100% behind the families there. We are 100% in the Middle East. I was also asked, you know, given what's happening in the Middle East, is that an area where we believe that we can expand artificial intelligence to? I believe that there's a reason we went to war, and I believe at the end of the war, Middle East will be more stable than before. And so if we were there before we were considering it before, we should absolutely be considering it after. And so I'm 100% in on that. With respect to Taiwan, we have to do three things. One, we have to make sure that we re-industrialize the United States as fast as we can. And whether it's the chip manufacturing plants, computer manufacturing plants, or the AI factory. How are we doing on that? We're doing excellent. By gaining the strategic support, by gaining the friendship of the supply chain of Taiwan. By gaining their friendship, by gaining their support, we were able to build Arizona and Texas, California at incredible rates. They they are genuinely a strategic partner. We we really they deserve our support. They deserve our friendship. They deserve our generosity. And they're doing everything they can to accelerate the manufacturing process for us. And so I think that's number one. Number two, we ought to diversify the manufacturing supply chain and whether it's South Korea, whether it's Japan, it's Europe, we ought to diversify the supply chain, make it more resilient. And number three, let's be let's demonstrate restraint. And while we're reducing in increasing our diversity and resilience, let's not press, push unnecessarily. We need to be patient. Thoughtful. Is helium a problem? A lot of reports happen to be. Helium could be a problem, but it's also the case that the supply chain probably has a lot of buffer in it. These kind of things tend to have a lot of buffer. But you know, yeah. You've made massive progress in self-driving. You made a big announcement, you've added many more partners including BYD, there was just a video of you driving around in a Mercedes and huge announcement with Uber that you're going to have a number of cars on the road for many different manufacturers. Your bet is I believe that there is going to be an Android type open-source platform that you're going to play a major part in with dozens of car providers and then maybe on the other side there could be an iOS with Tesla or Waymo. What's your strategy thinking there and how that chessboard emerges because it feels like you have a pretty deep stack and in some ways you're competing and in other places you're collaborative. Taking a step back, we believe that everything that moves will be autonomous completely or partly some day. Number one, number two we don't want to build self-driving cars, but we want to enable every car company in the world to build self-driving cars. And so we build all three computers, the training computer, the simulation computer, the evaluation computer, as well as the car computer. We developed the world's safest driving operating system. We also created the world's first reasoning autonomous vehicle. So that it could decompose complicated scenarios into simpler scenarios that it knows how to navigate through, just like us reasoning systems. And so that reasoning system, called Alpomaio, has enabled us to achieve incredible results. We opened this vertical optimization, we horizontally innovate, and we let everybody decide. Do you want to buy one computer from us? In the case of Elon and Tesla, they buy our training computers. Do they want to buy our training computer and our simulation computer? Or do you want to let us work with us to do all three and even put the car computer on your car? Our attitude is we want to solve the problem, we're not the solution provider, and we're delighted however you work with us. I'm just going to build on this question because I think it's so fascinating. You actually do create this platform, a thousand flowers are blooming. But it's also true that some of those flowers want to now go back down in the stack and try to compete with you a little bit. Google has TPU, Amazon has Inferentia and Trainium, you know everybody's sort of spinning up their own version of I think I can out-Nvidia Nvidia even though they also tend to be huge customers. How do you navigate that? And what do you think happens over time and where do those things play in the complexion of this kind of vision? Yeah, really great. You know first of all we're the only AI company, we're an AI company. We build foundation models, we're at the frontier in many different domains. We build every single every single layer every single stack. We're the only AI company in the world that works with every AI company in the world. They never show me what they're building and I always show them exactly what I'm building. Right. Yeah. And so so the confidence comes from this. One. We are delighted to compete on what is the best technology and to the extent that to the extent that we can continue to run fast, I believe that buying from NVIDIA still is one of the most economic things they could do. And I just saw incredible confidence there. Number one, number two. We're the only architecture that could be in every cloud, and that gives us some fundamental advantages. We're the only architecture you could take from a cloud and put into on-prem, in the car, in any region, in space. That's right, in space. And so there's a whole whole part of our market, about 40% of our market, most people don't realize this, 40% of our market, unless you have the CUDA stack, unless you can build entire AI factory, you have the customers don't know what to do with you. They're not trying to build chips. They're not trying to buy chips, they're trying to build AI infrastructure. And so they want you to come in with the full stack and we've got the whole stack. And so surprisingly NVIDIA's gaining market share. If you look at where we are today we're gaining share. Do you think what happens is these guys try and they realize oh my god too much. And then they come back is that why the share grows? Well we're gaining share for several reasons. One our velocity has gone we help people realize it's not about building the chip, it's about building the system. And that system's really hard to build. And so their business with us is increasing. In the case of AWS I think they just announced I think it was yesterday that they're going to buy a million chips in the next couple years. I mean that's a lot of chips from from AWS and that's on top of all the chips they've already bought. And so we're delighted to do that. But number one we're gaining share this last couple years because we now have Anthropic coming to NVIDIA. Meta SL is coming to NVIDIA. And the growth of open models is incredible and that's all on NVIDIA. And so we're growing in share because of the number of models. We're also growing in share because outside all of these companies are outside of the cloud and they're growing regionally in enterprise, in industries, at the edge and that entire segment of growth is you know really hard to do if it's just building an ASIC. Related to that and not to get in the weeds on the numbers but analysts don't seem to believe, right? So if you look at the consensus forecast you said compute could one million X, right? And yet they have you growing next year at 30%, the year after that at 20% and in 2029 which is supposed to be a monster year at 7%, right? So if you just if you take your TAM and you apply their growth numbers, it suggests that your share will plummet. Do you see anything in your future order book that would make that correct? First of all they just don't understand the scale and the breadth of AI. Yes. I think that's true. Most people think that AI is in the top five hyperscalers. Right. That's right. There's also an orthodoxy around these law of large numbers where you know they have to go back to their investment banking risk committee and show some model they're not going to believe in their minds that five trillion goes to 15 trillion they're like it can go to seven. It's go-to-out-the-box. Never happened before. You can't have a 10 trillion dollar company. It's all just CYA stuff that I think people... And because because you have to redefine what it is that you do. There was somebody who made an observation recently that NVIDIA just Jensen how can you be larger than Intel in servers? And the reason for that is because the CPU market of the entire data center was about 25 billion dollars a year. Right. We do 25 billion dollars a year as you guys know in a very in the time that we were sitting here. And so obviously obviously that was a joke. All-In podcast roughly true. Don't worry everything on the show is roughly true. It's All-In, you could speak. That was not guidance. Anyhow, anyhow it the point is how big you can be depends on what it is that you make. NVIDIA's not making chips. Number one making chips does not help you solve the AI infrastructure problem anymore. It's too complicated. Number three most people think that AI is narrowly in the things that they talk about and hear and see. It's AI is much OpenAI is incredible they're going to be enormous. Anthropic is incredible they're going to be enormous. But AI is going to be much much bigger than that and we address that segment. Tell us about data centers in space for a second. We are already in space. How should the layman think about what that business is versus when you hear about these big data center builds that's happening in on the ground? Well we should definitely work on the ground first because we're already here. And number one number two we should prepare to be out in space and obviously there's a lot of energy in space. The challenge of course is that cooling you can't take advantage of conduction and convection. Exactly. And so you can only use radiation. And radiation requires very large surfaces and so now that's not an impossible thing to solve and there's lots of lots of space in space. But nonetheless the expense is still quite there it's there. We're going to go explore it. We're already there. We're already radiation hardened. We have CUDA in satellites around the world. They're doing imaging, image processing, AI imaging and and that kind of stuff ought to be done in space instead of sending all the data back here and do imaging down here. We ought to just do imaging out in space. And so there's a lot of things that we ought to do in space and in the meantime we're going to explore what is the architecture of data centers look like in space. And it'll take it'll take years. It's okay. We got plenty of time. I wanted to double-click on healthcare. I know you've got a big effort there. We're all of a certain age where we're thinking about lifespan, healthspan. I mean we all look great I think. Some better than others. I don't know what your secret is Jensen. Pretty good these Asians. I mean what are you taking what's off the menu? You've got to talk to me when we're backstage. I want to know in the green room what you got going on. Squats and pushups and situps. Perfect. Okay. That works. But what you know in terms of the build-out in healthcare where is that going? And what kind of progress are we making? I was just using Claude to do some analysis and saying like where are all these billing codes we spend twice as much money in the US we seem to get half as much. It seemed like 15 to 25 percent of the dollar spent were on these first GP visits and I think we all know like Chat GPT and a large language model does a better job more consistently today at a first visit so what has to happen there to kind of break through all that regulation and have AI have a true impact on the healthcare system? There's several several areas that we're involved in in in healthcare. One is AI physics and that's or AI biology using AI to understand, represent, predict biology behavior, biological behavior. And so that's one, that's very important in drug discovery. There's second which is AI agents, and that's where the assistants in helping diagnosis and things like that. Open Evidence is a really good example, Hippocratic is a really good example, love working with those companies. I really think that this is an area where agentic technology is going to revolutionize how we interact with doctors and how we interact for healthcare. The third part that we're involved in is physical AI. The first one's AI physics using AI to predict physics. The second one is physical AI, AI that understands the properties of the laws of physics and that's used for agentic robotic surgery, huge amounts of activities there. Every single instrument whether it's ultrasound or you know CT or whatever instrument we interact with in a hospital in the future will be agentic. You know Open Claude in a safe version will be inside every single instrument. And so in a lot of ways that instrument's going to be interacting with patients and nurses and doctors in a very unique way. Yeah, we've seen so much investment in AI weapons. It would be wonderful to see some investment in AI EMTs and paramedics and and saving lives not just taking them, which I think is a great segue into robotics. You've got dozens of partners. We had this very weird I don't know if I want to call a lost decade or 20 years of Boston Dynamics. Google bought a bunch of companies they then wound up selling them and spinning them out where people just thought eh robotics is just not ready for primetime. And now here we have the world's greatest entrepreneur at this time tied with you,Elon Musk, doing Optimus, pretty impressive and then other companies in China. How close is that to actually being in our lives where we might see a chef, a robotic chef, a robotic nurse, a robotic housekeeper, you know these humanoid factor actually working in the real world knowing what you know with those partners and the fidelity especially in China where they seem to be doing as good a job as we're doing here or maybe better? We invented the industry largely. America invented we could argue we got into it too soon. Yeah. And and we got exhausted. We got tired. About five years before the enabling technology appeared, the brain. Yeah, yeah. And we just got tired of it just a little too soon. Okay, that's number one. But it's here now. Now the question is how much longer? From the point of high functioning existence proof, high functioning existence proof to reasonable products technology never takes more than a couple of two, three cycles. And so a couple of two, three cycles would basically be somewhere around three years to five years. That's it. Three years to five years we're going to have robots all over the place. I think China is formidable and the reason for that is because their microelectronics, their motors, their rare earth, their magnets which is foundational to robotics they are the world's best. And so in a lot of ways our robotics industry will have to rely a lot on their ecosystem and their supply chain, the world's robotics industry will have to rely a lot on it. And so I think you're going to see some fast fast movements here. Ultimately one for one. Elon seems to think we're going to have one robot for every human seven billion for seven billion, eight billion for eight billion. Well I'm hoping more. Yeah. I'm hoping more. Yeah. Well first of all there's a whole bunch of robots that are going to be in factories working around the clock. They're going to be a whole bunch of factory robots that that don't move they move a little bit. Almost everything will be robotic. What is the world like? Sorry let me just I think like this is one of the pieces that I think unlocks economic mobility opportunities for every individual. Everyone now like when everyone got a car they could now go and do a lot of different jobs. When everyone gets a robot their robot could do a lot of work for them. They can stand up an Etsy store, a Shopify store, they can create anything they want with their robot. They could do things that they independently cannot do. I think the robot is going to end up being the greatest unlock for prosperity for more people on earth than we've ever seen with any technology before. Yeah, no doubt. I mean just the simply the the simple math at the moment is we're millions of people short in labor today. Right. We're we're actually really desperate in need of robotics. And so that all of these companies could grow more if they had more labor. I mean we're number one. Some of the things that you mentioned are super fun. I mean because of robots we'll have virtual presence. You know I'll be able to go into the robot at my house and virtually operate it while I'm on a business trip walk around the house and walk the dog, rake the leaves. Freak out the dog. Maybe not quite that, but just, you know, wander around and just see what's going on in the house, you know, chat with the dog. Chat with the kids. Yeah. Time travel is also we're going to be able to travel at the speed of light, you know, and so you know clearly we're going to send our robots ahead of us, not going to send myself, I'm going to send a robot you know check it out yeah yeah and then I'm going to upload my AI. Well it's inevitable it unlocks the moon and it unlocks Mars as targets for colonization which gives us infinite resources. Getting back from the moon is effectively zero energy cost to move material back because you can use solar and accelerate. So you could have factories that make everything the world needs on the moon and the robots are going to be the unlock for enabling... That's right distance no longer matters. Distance doesn't matter. Yeah. The more the more revenue we get out of models and agents, the more we can invest in building the infrastructure which then unlocks more capabilities on models and agents. Dario on Dwarkesh's podcast recently said by 27, 28 we'll have hundreds of billions of dollars of revenue out of the model companies and the agent companies and he forecast a trillion dollars by 2030, right? This is non-infrastructure AI revenue. I think I think he's being very conservative. I believe Dario and Anthropic is going to do way better than that. way better than that. From 30 billion to a trillion? Yep and and the one part that he hasn't considered is that I believe every single enterprise software company will also be a reseller value added reseller of Anthropic Code, Anthropic's tokens, value added reseller OpenAI. get this logarithmic expansion. Yeah. Their go-to-market is going to expand tremendously this year. What do you think what do you think in that world is the moat? What's left over? I mean you have some moats that are frankly I think as this scales almost insurmountable the best one that nobody talks about is probably CUDA which is just like an an incredible strategic advantage. But in the future if a model can be used to create something incredible then the next spin of a model can be used to maybe disrupt it sort of in your mind what do you think for these companies that are building at that application layer what's their moat like how do they differentiate themselves? Deep specialization. Deep specialization. I believe that these models they they they're going to have general general models that are connected into the software company's agentic system. Many of those models are cloud models and proprietary models but many of those models are specialized sub-agents that they've trained on their own. Right, so the call to arms for you for entrepreneurs is look know your vertical, know it as deep and as better than everybody else and then wait for these tools because they're catching up to you and now you can imbue it with your knowledge. That's right and the sooner you connect your agent, the sooner you connect your agent with customers, that flywheel is going to cause your agent to get... We very much is an inversion of what we do today because today we build a piece of software and we say what generalizes and then let's try to sell it as broadly as possible and then sell the customization around it. And we trapped... in fact, exactly right. We create a horizontal but notice there all these GSIs and all these consultants who are specialists who then take your horizontal platform and specialize it into... Exactly. And that's arguably a five or six times bigger industry is the customization. It is absolutely. That's right. So I think that these platform companies have an opportunity to become that specialist, to become that vertical. Yeah, domain expert. Yeah, very much. You know I just want to give you your flowers. I think it was three years ago you said you're not going to lose your job to AI you're going to lose your job to somebody using AI. And here we are the entire conversation has revolved around this concept of agents making people superhuman and the business opportunity expanding and entrepreneurship expanding. You actually saw it pretty clearly. That's right. You changed your view? Well you you you were... Doomer Dave? No no no I'm not doomer I'm not doomer I I do have... That spiral Jcal. No no no there are going to be a... That's just because he doesn't hang out with me enough. But we I mean we've hung out a little bit. Be careful. We don't talk about it. People show up at your breakfast table. He'll follow you around! I'm not asking for it, I'm just. He'll follow you around. I mean you can. You come with me and Tucker we ski in Japan every January. Love it. Me and Tucker we go road trip. Wow. Okay. No comment. There is going to be job displacement, and then the question becomes, you know, do those people have the fortitude, the resolve, to then go embrace these you know technologies? we're going to see a hundred percent of driving go away by humans, that's just... It's a that's a beautiful thing and the lives saved but we have to recognize that's 15 million people in the United States, 10 to 15 million who are employed in that way and and so that is going to happen yes? I think I think that jobs will change. For example um there are many chauffeurs today who drives the car, I believe that many of those chauffeurs will actually be in the car sitting behind the drive the steering wheel while the car is driving by itself. And the reason for that is because remember what a chauffeur does, in the end these chauffeurs they help you their your assistants, they help you with your luggage, they help you I mean they help you with a lot of things. And so I wouldn't be surprised actually if the chauffeurs of the future become your mobility assistant and they are helping you do on a whole bunch of other stuff and change the hotel everything the car's driving by itself the autopilot in planes created a lot more pilots and didn't take any of the pilots out of the cockpit even though the autopilot is flying the plane 90% of the time! And by the way while that car is driving itself that chauffeur is going to be doing a bunch of other work on his phone, he's going to be making money doing another coordinating a bunch of things for you getting yeah the pie just grows in a way that yes every job will be transformed some jobs will be eliminated however we also know that many many jobs will be will be created. The one thing that I will say to young people who are coming out of school who are concerned who are anxious about AI: be the expert of using AI. Yes. How much look we all want our employees to be expert at using AI and it's not not trivial, not trivial. And so knowing how to specify, not to over prescribe, leaving enough room for the AI to innovate and create while we guide it to the outcome we want evalu all of that requires artistry. You had you you had this great advice to when you were at Stanford I think it was which is I wish to you pain and suffering do you remember that? Fantastic. What's your advice to young people around what they should be studying? So if they're sort of about to leave high school because now those are the kids that are at this like really native they haven't made a decision about college what to study if at all go to college? How do you guide those kids? What would you tell them? I still believe that deep science, deep math language skills, you know as you know language is the programming language of AI now, ultimate language. And so so as it turns out it could be that the English major could be the most successful. Yeah. And and so so I think um I would just advise whatever whatever education you get just make sure that you're deeply deeply expert in using AI. One of the things that I wanted to say with respect to jobs and I want everybody to hear it that in fact at the beginning of the Deep Learning revolution one of the the finest computer scientist in world deeply deeply I deeply respectpredicted that computer vision will completely eliminate radiologists. And and that the one the one field he advised everybody to not go into was radiology. 10 years later his prediction was at 100% right. Computer vision has been integrated into all of the radiology technologies and radiology platforms in world, 100%. The surprising outcome is the number of radiologists actually went up and the demand for radiologists is skyrocketed. The reason for that is because everybody's job has a purpose and its task. the task that you do is studying the scans. But your purpose is to diag help the doctor heal the patient diagnose disease. And so what's surprising is because the scans are now being done so quickly they could do more scans, improving healthcare. But doing more scans more quickly allows patients to be on-boarded a lot more quick treated a lot more quickly and as it turns out because hospitals enjoy making money too, they're doing more scans they're treating more customers and more patients their revenues go up and guess what? They need more. Perfect example. And a country that grows faster, productivity increases, a wealthier country can put more teachers in the classroom, not less teachers in the classroom. You just give every one of those teachers a personalized curriculum for every student in the room, it makes them all bionic and leads to a lot more teaching and monitoring! Every single student will be assisted by AI, but every single student will need great teachers. Amazing. Jensen, congratulations on your success and really this is an incredibly positive uplifting discussion. We really appreciate you taking the time for us. He is the steward we need. You are. I think you need to be more vocal. I'm being very very vocal! Be more vocal about the positive side of it. I think there's so much doomerism! But I also think it takes the humility to have this level of success and be humble about we're making software guys. Yeah. And I think that that's actually really healthy for people to hear. We have done this before. We have invented categories and industries before. We don't need to go to this scaremongering place. It does nothing. And we get to choose, right? We have autonomy and agency. We get to pick how to deploy this. We sure do. Okay everybody we'll see you next time on the All-In interview. Thanks man! Good job! Good job sir! That was awesome! Good, good! You guys are awesome. Look at this, look at this big crowd behind you guys. Thank you Jensen! I think they're here for you!
Loading video analysis...