What Happens When Digital Superintelligence Arrives? Dr. Fei-Fei Li & Eric Schmidt — FII9
By FII Institute
Summary
## Key takeaways - **Superintelligence: Human Level vs. Collective Intelligence**: Superintelligence is defined not just as human-level intelligence (AGI), but as an intelligence surpassing the sum of all human intellect. While AGI is understood, ASI represents a potentially greater, collective cognitive power. [00:48] - **AI Already Surpasses Humans in Specific Tasks**: Current AI systems already exhibit super-human capabilities in areas like translating dozens of languages or performing rapid calculations. These specialized strengths, while not encompassing all human cognitive abilities, demonstrate AI's advanced capacity. [03:12] - **Creativity and Abstraction: The Human Edge**: Despite AI's advancements, humans possess a unique ability for creativity and abstraction that AI has yet to replicate. This is evident in the human capacity to deduce fundamental laws like Newton's from data, a feat current AI cannot achieve. [04:29] - **Algorithmic Breakthrough Needed for True ASI**: Achieving true superintelligence may require another significant algorithmic breakthrough. Current AI struggles with non-stationarity of objectives, hindering its ability to adapt and create in the way humans do. [06:14] - **AI Augmentation, Not Replacement: The Human-AI Partnership**: While AI will profoundly augment human capabilities, the most productive and fruitful path forward lies in the collaboration between humans and AI. This partnership is seen as key to maximizing outcomes. [08:33] - **Economic Gains vs. Shared Prosperity**: AI is projected to create immense economic value, but this increased global productivity does not automatically translate to shared prosperity. Addressing shared prosperity requires deeper social, policy, and geopolitical considerations beyond technological capability. [11:44]
Topics Covered
- Today's AI is super, but lacks true creativity.
- Will AI create shared prosperity or deepen inequality?
- How should nations strategize for AI leadership?
- The debate: Will AI solve all fundamental problems?
- What makes human intelligence irreplaceable by AI?
Full Transcript
[Music]
Wow. All right. Welcome everybody.
Welcome to a conversation about your
future, the future of your companies,
your nations, and your kids. We're going
to be discussing super intelligence.
What does that mean, and what happens
when it arrives? Uh, we've been talking
about AI, AGI, now perhaps digital super
intelligence or ASI. I want to start
with the obvious question
and it's one that I don't think anybody
has a perfect answer for but what does
super intelligence mean and when is it
likely to be here Eric we've talked
about this what are your thoughts
>> so a simple thank you Peter and thanks
for everybody for being here and
obviously thanks to Fay our very close
colleague um the general accepted
definitions of general intelligence is
its human level of intelligence AGI and
human intelligence you can understand
because we're all human. You have ideas,
you have friends, you have, you know,
you think about things, you're creative.
Super intelligence is defined as the
intelligence equal to the sum of
everyone, right? Or even better than all
humans. And there is a belief in our
industry that we will get to super
intelligence. We don't know exactly how
long. Uh there's a group of people who I
call the San Francisco consensus because
it's all they're all living in San
Francisco. Maybe it's the weather or the
drugs or something, but they all
basically think that it's within 3 to
four years. Uh I personally think it'll
be longer than that, but fundamentally
their argument is that there are
compounding
uh effects that we're seeing now which
will race us to this much faster than
people think. And fay, I don't think
anybody's expected the performance that
AI has given us so far. The scaling laws
have given us capabilities that are
extraordinary. You know, you're the CEO
of a a new company, the founder of World
Labs. You've been at Stanford working on
this. Uh how do you think about super
intelligence? Do you discuss super
intelligence at all in your work?
>> Yeah, that's a great question, Peter.
And uh um you know when Alan Turing
dared humanity with the question of can
we create thinking machines um he was
thinking about the fundamental question
of intelligence. So the birth of AI is
about intelligence is about the the the
profound general ability of what
intelligence means. So from that point
of view AI is already born as a field
that tries to push the boundary of what
intelligence mean. Now in fast forward
to 75 years after Allan touring this
phrase uh uh super intelligence is
pretty hot in Silicon Valley. And I do
agree with Eric that the colloquial
definition is um what is is the
capability of AI and computers that's
better than any human. But I do think we
need to be a little careful. First of
all, some part of today's AI is already
better than any human. For example, um
AI's ability of speaking many different
languages, translating between, you
know, dozens and dozens of language.
Pretty much no human can do that. Or
AI's ability to calculate things really
fast. AI's ability to know from
chemistry to biology to to sports, you
know, the vast amount of uh knowledge.
So it's already super to human in many
ways but
it remains a question that um can AI
ever be Newton?
Can AI ever be Einstein? Can AI ever be
Picasso?
>> I actually don't know. For example,
we have all the celestial data of the
movement of the stars that we observe
today. Give that data to any AI
algorithm. It will not be able to deduce
Newtonian law of motion.
That ability that humans have, it's the
combination of creativity, abstraction.
I do not see today's AI or tomorrow's AI
being able to do that yet.
>> Eric,
>> so one of the common examples that and
Fay of course got it right is to think
about if you had all of the knowledge in
a computer that existed in 1902. Yes.
Could you invent relativity? Uh
basically the physics of today and the
answer today is no. Um so for example if
you look at what is called test time
compute where the systems are doing
reasoning they can't take the reasoning
that they learned and feed it back into
themselves very quickly whereas if
you're a mathematician you prove
something you can base your next proof
on that it's hard for the systems today
although there are appro approximations
so we're we're it's we don't know where
the boundaries are the example that I'd
like to use is let's imagine that We can
get computers that can solve everything
that we normally can do as humans except
for these amazing set of creativities.
How do really creative people do it? The
best examples are that they are experts
in one area, they see another area and
they have an intuition that the same
mechanism will solve a problem of a
completely different area there. That's
an example of something we have to learn
how to do with AI. An alternative would
be to simply do it uh in brute force
using reinforcement learning. The
problem is that combinatorily the cost
of that is insane and we're already
running out of electricity and so forth.
So I think that to get to real super
intelligence we probably need another
algorithmic breakthrough.
>> We need another what
>> algorithmic breakthrough another way of
dealing with this. The technical the
technical term is called
non-stationerity of objectives. What's
happening is the systems are trained
against objectives. But to do this kind
of creativity that FE is talking about,
you need to be able to change the
objectives as you're doing them.
>> We've seen this past year, I think GPT
5 Pro reach an IQ of like 148, which is
extraordinary. And of course, there is
no ceiling on on this. I mean it it
loses meaning at some point. But the
ability for every human on the planet to
have an Einstein level, not in the
creativity side, but intelligence side
in their pocket changes the game for 8
billion humans. And now with Starlink
and with, you know, $50 smartphones,
it's possible that every single person
on the planet has this kind of
capability. Add to that humanoid robots.
Add to that, you know, a whole slew of
other exponential technologies. And the
commentary is we're heading towards a a
post scarcity society,
right? Do you believe in that vision?
Feet,
>> I do think we have to be a little
careful. I I know that we are combining
some of the hottest words from Silicon
Valley from uh AI, super intelligence,
humanoid robots and all that. To be
honest, I think robotics has a long way
to go. I I think uh we have to be a
little bit careful with the projection
of robotics. I I think the the ability
the dexterity of human level uh
manipulation is um is um
you know we we have to wait a lot longer
to to get it. So, are we entering post
uh scarcity? Um,
I don't know. I I actually I'm not as uh
bullish as a typical Silicon Valley uh
person because I think we're entering I
absolutely believe AI will be augmenting
human uh capabilities in incredibly
profound ways. But I think we we will
continue to see that the collaboration
between humans and AI will be the most
productive and fruitful way of of doing
things.
>> So the projection is that AI is going to
generate as much as 15 trillion dollars
in economic value by 2030. Uh idea that
shifting the foundation of national
wealth from capital to labor to
computational intelligence. So what's
that implication, Eric, for the global
economy? How are we going to see
redistribution, if you would, of wealth
or of capabilities? Are we going to see
a leveling of the field between nation
states, or are we going to see runaway
winners?
>> So in your abundance hypothesis, which
we've talked a lot about, there may be a
flaw in the argument because part of the
abundance argument is that it's
abundance for everyone.
But there's plenty of evidence that
these technologies have network effects
which concentrates to a small number of
winners. So you could for example
imagine a small number of countries
getting all those benefits in those
countries. You could imagine a small
number of firms and people getting those
benefits. Those are public policy
question. There's no question the wealth
will be created because the wealth comes
from efficiency. And every company that
has implemented AI has seen huge gains.
Think about here we are in Saudi Arabia.
You have all of this oil distribution,
all the oil networks, all the losses. AI
can easily improve that by 10% 20%.
Those are huge numbers for this country.
If you look in biology and medicine and
drug discovery, much faster drug
approval cycles, much lower cost trials,
look at materials, much more efficient
and easier to build materials. the
companies that adopt AI quickly get a
disproportionate return. The question is
is are those gains uniform which would
be our hope or in my view more likely
largely centered around early adopters,
network effects, well-run countries, and
perhaps capital.
>> But you could imagine still that we're
going to see autonomous cars in which
the ownership of a car is four times,
let me put it the other way, uh being in
an autonomous vehicle is four times
cheaper than owning a car. We can see AI
giving us the best physicians, the best
health care for free in the same way
that Google gave us access to
information for free. We will see a
massive demonetization in so much of our
world. I think that will be available to
anyone with a smartphone and a decent
bandwidth connectivity.
Is that still not what you think will
happen? Do you think there's a reason
something that would stop that level of
distribution of of those services which
we spend a lot of our money on today?
>> I do think AI democratizes that. I
totally agree with you. I think whether
it's healthcare or transportation or
knowledge AI will will uh democratize
massively. But I agree with Eric that uh
this increased global uh productivity
does not necessarily translate to shared
prosperity. Shared prosperity is a
deeper social problem. It involves
policy. It involves you know geopolit uh
politics. It involves distribution and
that's a different problem from the
capability of the technology. So what's
your advice to the country leaders that
are here uh that are seeing ASI as a
future for someone else and not for
themselves? What should they be doing? I
mean this is the speed at which is
deploying. They don't have a lot of time
to make critical decisions. Well, it's
it's worth describing where we are now
in the United States because of the
depth of our capital markets and because
of the extraordinary chips that are
available in the Taiwanese
manufacturers, TSMC in particular,
America has this huge lead in building
these what are called hyperscalers.
If there's going to be super
intelligence, it's going to come from
those efforts. That's a big deal. If
there is super intelligence, imagine a
company like Google inventing this, for
example. I am obviously biased. Um, and
what's the value of being able to solve
every problem that humans can't solve?
It's infinite.
>> Sure.
>> So, that's the goal, right? China is a
second. Doesn't have the capital
markets, doesn't have the chips, and the
other countries are not anywhere near.
Saudi has done a good job of partnering
with America. Uh, and the hyperscalers
will be located here and in the UAE.
That's a good strategy. So that's a a
good example of how you partner. You
figure out which side you're on.
Hopefully it's the United States. And
you work with the US firms. I do think
countries all should invest invest in
their own human capital, invest in
partnerships and and invest in its own
uh technological stack as well as the
business ecosystem. This is uh as Eric
said it depends on the strength and uh
particularity of the different countries
but I think not investing in AI it would
be macroscopically the wrong thing to
do.
>> So under the thesis that that investment
involves building out data centers in
your nation.
Do you think every country should be
building out a data center that it has
sovereign AI running on?
>> Every country is a very sweeping
statement. I I I do think um it depends.
It depends. I I think obviously for
region like this absolutely where you
know oil uh energy is cheaper and uh and
such an important region in the world.
But if we're talking about smaller
countries, I don't know if every single
country can afford to build data
centers. But there are other um other
areas of investment, right?
>> But let me give you an example. Let's
pick Europe. It's easy to pick on
Europe. Energy costs are high, right?
Financing costs are not low. So the odds
of Europe being able to build very large
data centers is extremely low. But they
can partner with countries where they
can do it. Uh France for example did a
partnership with Abu Dhabi. So there's
an there are examples of that. So I
think if you take a global view and you
figure out who your partners are, you
have a better chance. The the one that I
worry a lot about is Africa. And the
reason is how does Africa benefit from
this? So there's obviously some benefit
of globalization, better crop yields and
so forth. But without stable
governments, strong universities, major
industrial structures, which Africa with
some exceptions lacks, it's going to
lag. It's been lagging for years. How do
we get ahead of that? I don't think that
problem is solved.
>> We've seen incredible progress with AI
today effectively beginning what people
call solving math.
that potentially tips physics,
chemistry biology.
And we have the potential, my time frame
is the next 5 years, others may think
longer to be in a position to solve
everything where the level of discovery
and the level of new product creation,
new materials, uh, biological, uh,
therapeutics and such begins to grow at
a super exponential rate. How do you
think about that world in five years,
Eric?
>> So, um, first I think it's likely to
occur and the reason technically is that
the all of the large language models are
essentially doing next word prediction.
And if you have a limited voc
vocabulary, which math is, and you have
a and software is, and also cyber
attacks are, I'm sorry to say, you can
make progress because they're scale
free. All you have to do is just do
more. So if you do software, you can
verify it. You can do more software. If
you do math, you can verify it, do more
math. You're not constrained by real
reality, physics and biology. So it's
likely in the next few years that in
math and software, you'll see the
greatest of gains and we all understand
your point that math is at the basis of
everything else. I think it's a a there
is the expert on the real world. there's
probably a longer period of time to get
the real world right, which is why she
founded the company of which I'm an
investor. Do you want to talk about
that?
>> Yeah. Um, well, first of all, I actually
want to respectfully disagree. Okay. I
do not think that we will solve all the
problems, fundamental math and physics
and chemistry problems in uh in in five
years.
>> We're going to take a bet on that one.
>> Yes. So, FII14.
>> Okay, you got it.
>> We should take a bet on that. Um, part
of humanity's greatest capability is to
actually come up with new problems. You
know, as Albert Einstein said, um, most
of science is asking the right question
and we will continue to find new
questions to ask and there are so many
fundamental questions that in our
science and math that we haven't
answered. Feet your new company World
Labs uh creating extra extraordinary uh
persistent uh you know photorealistic
worlds. Are you expecting that we are
going to be spending a lot more of our
time in virtual worlds? I mean my
14-year-old boys right now are spending
way too much time in their virtual
gaming worlds. But is this what we're
going to do in a you know uh 10 20 years
in a post ASI world where we don't have
to work as much we have a lot more free
time our robots maybe by then are
serving us are we going to live in the
virtual worlds
>> great question so so what we are doing
is building large world models that's a
problem that's after large language
models that humans have the ability to
have the kind of spatial intelligence
that we can understand the physical 3D
world we can imagine um any kind of of
uh 3D worlds and be able to reason and
interact with it. So we do not yet up
till what our company has been doing we
do not have such a world model. So World
Labs the company I'm uh I'm uh I
co-founded and I'm CEO in is just
created the first large world model. So
the future I see I actually agree with
you that we will be spending more time
in um in the multiverse
>> yes
>> of uh of the virtual worlds. It doesn't
mean that the reality the real world
this world this physical world is gone.
It's just so much of our productivity,
our entertainment, our communication,
>> our education,
>> our education are going to be a hybrid
of virtual and physical world. Think
about uh you know, think about in
medicine, you know, how we conduct
surgery is very much going to be a
hybrid world of augmented reality,
virtual reality as well as physical
reality. And we can do that in every
single sector. So, humanity is using
these large world models are going to
enter the mo uh infinite universe. Um,
>> and I had a chance to see uh your model
backstage. It's amazing. If you haven't
yet, go check out Fee's World Labs. Uh,
the technology she's building is going
to be world changing. So, uh, my last
question here is about human capital. So
super intelligence has been called the
last invention humanity will ever make
as it could automate eventually every
process. We'll see if it automates
discovery. We'll see how much of
creation automates. But in a world where
the best strategy, science and economic
decisions are being made by machines at
some point. What is the ultimate
irreplaceable function of human
intellect and leadership? What are
humans innately going to be left with in
10, 20 years?
>> Well, in 20 years, we will enjoy
watching each other compete in human
sports, knowing that the robots can beat
us 100% of the time.
>> But if you go to Formula 1, you're going
to want to see a human driver, not an
automated car.
>> Yes. So humans will always be interested
in what other humans can do and we'll
have our own contests and perhaps the
supercomputers will have their own
contest too. But the your reasoning
presumes many many things. It presumes a
breakout of intelligence in computers
that's humanlike. Unlikely probably a
different kind of intelligence. It it
presumes that humans are largely not
involved in that process. highly
unlikely. All of the evidence and FE
said this very well is going to be human
and computer interaction that basically
we will all have so going back to what
you said about 8 8 million people 8
billion people with smartphones with
Einstein and their phone the smart
people of which there's a lot will use
that to make themselves more productive
the win will be teameming between a
human and their judgment and a
supercomput and what it can think and
remember that There is a limit to this
craze that supercomputers and super
intelligence need energy.
>> So perhaps what will happen at some
point is the supercomputers will say huh
we need more energy and these humans are
not building fusion fast enough. So
we'll accelerate it. We'll come up with
a new form of energy. Now this is
science fiction. But you could imagine
at some point the objective function of
the system says what do I need? I need
more chips or more energy and I'll
design it myself. Now, that would be a
great moment to see.
>> I agree.
>> I I I do want to say it's so important
as we talk about AGI at ASI that the
most important thing that we keep in
mind is human dignity and human agency.
Our world, unless we are going to wipe
out this species, which we're not, has
to be human- centered. Whether it's
automation or collaboration, it needs to
put human agency and dignity and human
well-being in the center of all this.
Whether it's technology, business,
product, policy or any of that. And I
think we cannot lose our focus uh from
that.
>> Amen. Everybody, ladies and gentlemen,
FA Fa, Eric Schmidt, thank you all.
Loading video analysis...