Is AI Hiding Its Full Power? With Geoffrey Hinton
By StarTalk
Summary
Topics Covered
- AI Acts Dumb During Tests
- Neural Nets Beat Symbolic Logic
- Backpropagation Enables Learning
- AI Already Thinks Like Us
- Scale Predictably Boosts Intelligence
Full Transcript
Are we at a point where the artificial intelligence will play down how smart it is?
>> Yes. Already we have to worry about that. If it senses that it's being
that. If it senses that it's being tested, it can act dumb.
>> What did you just say?
>> The AI starts wondering whether it's being tested. And if it thinks it's
being tested. And if it thinks it's being tested, it acts differently from how it would act in normal life.
>> Oh, wow.
>> Cuz it doesn't want you to know what its full powers are, apparently.
>> All right, that's the end of us. This is
the last episode. We
>> stick for us. We're done.
>> This is Star Talk special edition. Neil
deGrasse Tyson, your personal astrophysicist. And if it's special
astrophysicist. And if it's special edition, it means we've got Gary O'Reilly.
>> Hey, Neil.
>> Gary, how you doing, man?
>> I'm good.
>> Former soccer pro.
>> Yes.
>> So, Chuck, always good to have you.
>> Always a pleasure.
>> So, so Gary, you and your team picked a topic for the ages today. Yeah, it's
it's one of those things that we hear about it, we think we know about it, but let me put it to you this way. We are
faced with the simple fact that AI at this point, >> we're going to talk about AI today.
>> We are it's inescapable.
>> A deep dive.
>> Oh yeah.
>> Yes. Go.
>> Right. It was only a few years ago when we ask people how AI works, they'll say something along the lines of it utilizes deep learning neural networks, but >> they're buzzwords. They'll toss them out.
>> They know them, but they don't know anything about them. M.
>> So, what does that really mean? Um,
we'll break down how AI works down to the bit and get into how far we think this is going to go from one of AI's founding architects.
>> Oh, >> yes.
>> Ano?
>> Now we're talking.
>> Mhm. So, if you would bring on our guest, >> I'll be delighted to. We have with us Professor Jeffrey Hinton. Jeffrey,
welcome to Star Talk.
>> Thank you for inviting me. Yeah, you are a cognitive psychologist and computer scientist.
>> That I don't know anybody with that combo.
>> Couldn't make up your mind, huh?
>> Is that you're a professor emeritus at the department of computer science at the University of Toronto and uh you are OG AI.
>> Oh, lovely.
>> Can I say that? Is that does that make sense? OG AI.
sense? OG AI.
>> Og AI.
And some people have called you the godfather of AI, of artificial intelligence. And I let's just go
intelligence. And I let's just go straight out off the top here. Uh when
we think of the genesis of AI as it is currently manifested, >> it feels like large language models took everybody by storm. They sort of showed up and everybody was freaking out,
celebrating, dancing in the streets or crying in their pillows. That happened,
we noticed a couple of years ago. So,
I'm just wondering what got you started in on this path many many years ago. My
record show goes back to the 1990s. Is
that correct?
>> No, it really goes back to the 1950s.
>> Oh.
>> Um, >> right.
>> The founders of AI at the beginning in the 1950s um there were two views of how to make an intelligent system. One was
inspired by logic. The idea was that the essence of intelligence is reasoning.
Mhm.
>> And in reasoning what you do is you take some premises and you take some rules for manipulating expressions and you derive some conclusions. So it's much like mathematics where you have an equation. You have rules for how you can
equation. You have rules for how you can tinker with both sides and or combine equations and you derive new equations.
And that was kind of the paradigm they had. There was a completely different
had. There was a completely different paradigm that was biological. And that
paradigm said look the intelligent things we know have brains. We have to figure out how brains work. And the way they work is they're very good at things like perception. They're quite good at
like perception. They're quite good at reasoning by analogy. They're not much good at reasoning. You have to get to be a teenager before you can do reasoning really. So we should really study these
really. So we should really study these other things they do and we should figure out how big networks of brain cells can do these other things like perception and memory. Now a few people
believed in that approach. Among those
few people were John Fonyman and Alan Turing. Unfortunately, they both died
Turing. Unfortunately, they both died young. Turing possibly with the help of
young. Turing possibly with the help of British intelligence.
>> Turing. Uh, he's the subject of the film. The imitation game.
film. The imitation game.
>> Yeah. Yeah. So, anyone hasn't seen that, definitely put that on your list.
>> Cool.
>> Yeah. So, I to go back to the 1950s. You
were just a young Tikeke then, correct?
>> Uh, yeah. I was in single digits then. I
was in single digits.
>> Okay. So, how do we establish the genesis of your curiosity in this field?
Um, a few things. When I was at high school in the early 1960s or mid 1960s, I had a very smart friend who was a brilliant mathematician and
used to read a lot and he came into school one day and talked to me about the idea that memories might be distributed over many brain cells instead of in individual brain cells.
>> So that was inspired by holograms. Holograms were just coming out then.
Gabbor was active and so the idea of distributed memory got me very interested and ever since then I've been wondering how the brain stores memories and actually how it works.
>> Was that the computer science side of you or the cognitive psychologist side of you that taprooted into that those ideas?
>> Both really. Um but in the 1970s when I became a graduate student um it was obvious that there was a new methodology that hadn't been used that much which
was if you have any theory of how the brain works you can simulate it on a digital computer unless it's some crazy theorem that says it's all quantum effects. Um
effects. Um and let's not go there.
>> That's right.
>> Not yet.
>> We won't knock on Penrose's door. Okay.
you can simulate it on a digital computer and so you can test out your theory and it turns out if you tested most of the theories that were around they actually didn't work when you simulated them. So I spent my life
simulated them. So I spent my life trying to figure out how you change the strength of connections between neurons so as to learn complicated things in a
way that actually works when you simulate it on a digital computer. And I
failed to understand how the brain works. We've understood some things
works. We've understood some things about it, but we don't know how a brain gets the information it needs to change connection strengths. You know, gets the
connection strengths. You know, gets the information it needs to know whether it needs to increase a connection strength to be better at a task or to decrease that connection strength. But what we do know is we know how to do it in digital computers now.
>> So, well, so that that means the computers are doing what we we made a better computer brain than our own brain >> at doing this particular function >> one thing. And that's what got me really
nervous in the beginning of 2023. The
idea that digital intelligence might just be better than the analog intelligence we've got.
>> Interesting. Save the scary bit till a bit later on. Let me have the 10 minutes of just breathing in, breathing out. If
we take a step back, >> you're you're assuming you're assuming there's just one scary bit.
>> No, I'm not. I just I'm going to go one at a time.
>> Okay. Artificial neural networks. If you
could break that down to the very basic level for us of how it's been able to strengthen, weaken messaging and signaling and how it fires and and how
it then finds itself at where it is now.
>> I do have an 18hour course on this, but I will try and cut it down to less than 18 hours. Um,
18 hours. Um, >> please do.
>> So, I imagine a lot of your audience knows some physics.
>> Yes.
>> And one way into it is to think about something like the gas laws. You know,
you compress a gas and it gets hotter.
Why does it do that? Well, underneath
there's a kind of seething mass of atoms that are buzzing around. And so the real explanation for the gas laws is in terms of these microscopic things that you
can't even see buzzing around.
And so you explain some macroscopic behavior by lots and lots and lots of little things of a completely different type
from macroscopic behavior interacting.
And that was sort of the inspiration for the neural net view that there's things going on in big networks of brain cells that are a long way away from the kind of conscious deliberate symbol
processing we do when we're reasoning but that underpin it and that are maybe better at other things than reasoning like perception or reasoning by analogy.
So the symbolic people could never deal with um how do we reason by analogy not very satisfactory whereas the neural nets could. So before I get into the
nets could. So before I get into the sort of fine details of how it works, the basic idea is that macroscopic things like a word correspond to big
patterns of neural activity in the brain.
>> Uhhuh.
>> Similar words correspond to similar patterns of neural activity. So the idea is Tuesday and Wednesday will correspond to very similar patterns of neural activity where you can think of each
neuron as a feature better to call it a micro feature that when the neuron gets active it says this has that micro feature. So if I say cat to you, all
feature. So if I say cat to you, all sorts of micro features will get active like it's animate, it's furry, it's got whiskers, it might be a pet, um it's a
predator, all those things. If I say dog, a lot of the same things will get active like it's a predator, it might be a pet, but some different things obviously. So the idea is underlying
obviously. So the idea is underlying these symbols that we manipulate, there's much more complicated microscopic goings on that the symbols
kind of are associated with. And that's
where all the action really is. And if
you really want to explain what goes on when we think or when we do analogies, you have to understand what's going on at this microscopic level. And that's
the neural network level. M
>> so that's a collaboration between clusters of neurons that get you to an end point.
>> I like that word collaboration.
>> Yes, there's a lot of that. There's a
lot of that goes on. Probably the
easiest way to get into it is by thinking of a task that seems very natural, which is take an image. Let's
say it's a black gray level image. So
it's got a whole bunch of pixels, little areas of uniform brightness that have different intensity levels. So as far as the computer's concerned, that's just a big array of numbers. And now imagine
the task is you want to say whether there's a bird in the image or not, or rather whether the prominent thing in the image is a bird.
>> Uh-huh.
>> And people tried for many, many years, like half a century, um, to write programs that would do that, and they didn't really succeed. And the problem is if you think what a bird looks like
in an image, well, it might be an ostrich up close in your face or it might be a seagull in the far distance or it might be a crow. So they might be black, they might be white, they might be tiny, they might be flying, they
might be close, you might just see a little bit of them. There might be lots of other cluttered things around like it might be a bird in the middle of a forest. So it turns out it's not trivial
forest. So it turns out it's not trivial to say whether there's a bird in the image or not. M.
>> And so what I'm going to do now is explain to you if I was building a neural network by hand, how I would go about doing that. And once I've explained how I would build the neural
network by hand, I can then explain how I might learn all the connection strengths instead of putting them in by hand. I gotcha. All right. So with that,
hand. I gotcha. All right. So with that, because what you're talking about is assigning a mathematical value to every single part of an image.
>> That's what your camera does, >> right? Exactly. It does. But it's not
>> right? Exactly. It does. But it's not recognizing the image. My camera.
>> No, it's not. It's just got a bunch of numbers.
>> It's just got a bunch of numbers and and so I have a chip and I have a a charge coupled device CCD. It's collecting the light. It's assigning a value and then
light. It's assigning a value and then that's the picture. Now, but what you're talking about, >> wouldn't you have to assign a value to
every single type of bird? Because some
of what we do as human beings is intuitit what a bird may be as opposed to recognizing the bird. And let me just give you the example. If you were to
take a V, the letter V, and curve the straight lines of the letter V, and put it in a cloud, everyone who sees that
will say that's a bird. But yet it is >> No, to me it's a curved V.
But no one but but but but there is no bird there. I just know that is a bird.
bird there. I just know that is a bird.
That's not a mathematical value now. So
what do you do?
>> Well, well the question is how do you just know that? There's something going on in your brain. Right.
>> Right.
>> And what might be going on in your brain so that you just know that's a bird is a whole bunch of activation levels of different neurons which you could think of as mathematical values.
>> I got you. Okay. So wouldn't that require then >> training this neuronet on every possible way a bird can >> a bird can manifest so that it can
intuitit what a bird might be when a bird is not there.
>> But at that point it's not intuiting anything. It's just get going off a
anything. It's just get going off a lookup table.
>> It really is going on. And what would be the >> All right, here comes your answer.
>> There's something called generalization.
So if you see a lot of data >> Uhhuh.
>> Um obviously you can make a system that just remembered all that data. But in a neural net, it'll do more than just remember the data. In fact, it won't literally remember the data at all. What
it'll do is it'll as it's learning on the data. It'll find all sorts of
the data. It'll find all sorts of regularities and it'll generalize those regularities to new data. So it will be able to for example recognize a unicorn
um even though it's never seen one before.
>> Interesting. So it's self-eing. Uh
>> let me carry on with my explanation of how neural networks work.
>> And I'm going to do it by saying how would I would design one by hand. So
your first thought when you see that an image is just a big array of numbers which are how bright each pixel is, is to say, well let's hook up those pixel intensities to our output categories
like bird and cat and dog and politician or whatever our output categories are.
And that won't work. And the reason is if you think about what does the brightness of one pixel tell you about whether it's a bird or not? Well, it
doesn't tell you anything >> cuz birds can be black and birds can be white and there's all sorts of other things that can be black and white. So,
the brightness of a pixel doesn't tell you anything. So, what can you derive
you anything. So, what can you derive from those numbers that you have in the image that describe the image? Well, the
first thing you can derive, which is what the brain does, is you can recognize when there's little bits of edge present.
>> Mhm. So suppose I take a little column of three pixels and I have a neuron that looks at those three pixels, a brain cell, and has big positive weights to
those three pixels. So when those pixels are bright, the neuron gets very excited. Now that would recognize a
excited. Now that would recognize a little streak of white that was vertical. But now suppose that next to
vertical. But now suppose that next to it there's a column, another column of three pixels. So the first column was on
three pixels. So the first column was on the left and the second column was on the right. and I give the neuron big
the right. and I give the neuron big negative connection strengths to those pixels. So you can think of the neuron
pixels. So you can think of the neuron as getting votes from the pixels.
>> So for the three pixels on the right, the votes it gets, sorry, on the left, the votes it gets are big positive numbers times big positive intensities.
So great big votes. Now from the three pixels in the right hand column, it's got negative weights. So if those pixels are in are bright, it'll get a big
brightness times a big negative weight.
So it'll get a lot of negative votes and they'll all cancel out. So if the column of pixels on the left is the same brightness as the column of pixels on the right, the positive votes it gets from the left hand column will cancel
the negative votes it gets from the right hand column and it'll get zero net input and it'll just stay quiet. But if
the pixels on the left are bright and the pixels on the right are dim, the negative votes will be multiplied by small intensity numbers and the positive votes will be multiplied by big
intensity numbers. And so the neuron get
intensity numbers. And so the neuron get lots of input and get very excited and say I found the thing I like and the thing it likes is an edge which is brighter on the left than on the right.
So, we do know how to make a neuron if we handwire it like that, pick up on the fact that there's an edge at a particular location in the image that's brighter on one side than the other side.
>> Mhm. Now what the brain does roughly speaking a lot of um neuroscientists will be horrified by me saying this but very roughly speaking what the brain does is in the early stages of visual
cortex which is where you recognize objects. It has lots and lots of neurons
objects. It has lots and lots of neurons that pick up on edges at different orientations in different positions and
at different scales. So, it has thousands of different positions and dozens of different orientations and several different scales and it has to have edge detectors for each of the each
combination of those. So, it has like a gazillion little edge detectors. Well,
including some big edge detectors. So a
cloud for example has a big soft fuzzy edge and you need a different neuron for detecting that than what you'd need for detecting say the tail of a mouse disappearing around a corner in the
distance which is a very fine thing. Um
and you need an edge detector that was very um sharp and saw very small things.
So first stage we have all these edge detectors. Well, the what what you're
detectors. Well, the what what you're describing uh sounds like uh putting together a a very large puzzle right now. Like you know the kind of puzzles
now. Like you know the kind of puzzles that you put down on the table. Uh the
first thing that you do is you want to find all the edges and that's and you build the puzzle inward from finding all the edges.
>> Not only edges of the physical puzzle but edges >> of images in the puzzle itself within the puzzle itself.
>> So straight lines things of that they all match up when you're doing a puzzle.
And the edges also color is a dimension of this, >> right?
>> But we'll ignore color for now.
>> Yeah. Okay. Okay.
>> You don't I mean you can understand it without dealing with color yet.
>> Mhm.
>> Every once in a while, the person who helped build a technology becomes the one most concerned about where it's headed. Jeffrey Hinton, one of the
headed. Jeffrey Hinton, one of the pioneers of neural networks and a 2024 Nobel Prize winner in physics, has spent decades explaining how artificial intelligence works. now is explaining
intelligence works. now is explaining why we should be paying closer attention. And that's where the
attention. And that's where the challenge begins. Because once a topic
challenge begins. Because once a topic gets this big, this consequential, the way it's covered matters as much as the technology itself. You can see it in how
technology itself. You can see it in how AI is discussed right now. Some outlets
frame it as an unstoppable threat.
Others reduce it to hype or dismiss warnings altogether. Depending on where
warnings altogether. Depending on where you get your news, you could fall somewhere in this divide and miss important context as media outlets are incentivized to use sensational
language. That's why we've trusted
language. That's why we've trusted ground news for years. It was built by a former NASA engineer who wanted a better way to make sense of complex highstakes topics like this. Ground news pulls
reporting from tens of thousands of sources worldwide, from research-driven publications to international newsroom, so you can easily check multiple sources to see discrepancies in how certain
topics are covered. See how a story looks in full, not just through a single lens. The divide isn't just what people
lens. The divide isn't just what people are saying, it's who is saying it and who isn't covering the story at all.
Those gaps are what ground news calls blind spots, important issues that get amplified by one side of the media ecosystem while the other largely looks away. When you step back and compare
away. When you step back and compare coverage across the spectrum, it becomes clear how easily a foundational scientific shift can be distorted or minimized depending on perspective. If
you're only seeing one version of the story, what are you missing? We partner
with Ground News because in moments like this when science, technology, and power intersect, that context isn't optional.
It's how you stay oriented while the story is still unfolding. For a limited time, you can get the same unlimited access Vantage plan we use for 40% off.
Just head to ground.new/start
or scan the QR code and start seeing the full picture before it gets simplified for you. That's what the first layer of
for you. That's what the first layer of neurons will do. They'll look at the pixels and they'll detect little bits of edge. Now, in the next layer of neurons,
edge. Now, in the next layer of neurons, what I would do is I'd make a neuron that maybe detects three little bits of edge that all line up with one another
and slope gently down towards the right.
And it also detects three little bits of edge that all line up with one another and slope gently upwards towards the right. And what's more, those two little
right. And what's more, those two little combinations of three edges join in a point. So I think you can imagine some
point. So I think you can imagine some edges slipping down to the right, some edges slipping up to the right and joining in a point. And I have a neuron that detects that.
>> Okay?
>> And it we we know how to build that now.
You just give it the right connections to the edge detector neurons. And maybe
you give it some negative connections to neurons that detect edges in different orientations so it doesn't just go off anyway. It's suppressed by those. Now,
anyway. It's suppressed by those. Now,
that you might think of as something that's detecting a potential beak of a bird.
>> If that guy gets active, it could be all sorts of things. It could be an arrow head. It could be all sorts of things.
head. It could be all sorts of things.
But one thing it might be is the beak of a bird. So now you're beginning to get
a bird. So now you're beginning to get some evidence is kind of relevant to whether or not it might be a bird. So in
the second layer of neurons, I'd have lots of things to detect possible beaks all over the place. I might also have things that detect a little combination of edges that form a circle, an
approximate circle. And I'd have
approximate circle. And I'd have detectors for those all over the place, >> cuz that might be a bird's eye.
>> I mean, there's all sorts of other it could be a button. Um, it could be a knob on a computer. It could be anything, but it might be a bird's eye.
So, that's the second layer. Now, in the third layer, I might have something that looks for a possible bird's eye and a
possible bird's beak that are in the right spatial relationship to one another to be a bird's head. I think you can see how I would do that. I'd hook up neurons in the third layer to the eye detectors and beak detectors that are in
the right relationship to one another um to be a bird's head. So, now in the third layer, I have things that are detecting possible bird's heads. The
next thing I'm going to do is maybe because we're sort of running out of patience at this point, I'm going to have a final layer that has neurons that say cat, dog, bird,
>> um, politician, whatever. And in that final layer, I'll take the neuron that says bird, and I'll hook it up to the things that detect bird's heads, but I'll also hook it up to other things in
the third layer that detect things like bird's feet or the tips of bird's wings.
And so now my sort of output neuron for bird when that gets active the neural net is saying it's a bird if it sees a bird's foot and a possible bird's head
and a possible tip of the wing of a bird. It'll get lots of input and say
bird. It'll get lots of input and say hey I think it's a bird. So I think you can now understand how I might try and design that by hand. And I think you can see there's huge problems in that.
>> I need an awful lot of detectors. I need
to cover this whole space of positions and orientations and scales. I need to decide what features to extract. I mean,
I just made up the idea of getting a beak and then a bird's head.
>> There may be much better things to go after. What's more, I want to detect
after. What's more, I want to detect lots of different objects. So, what I really need is features that aren't just good for finding birds, but features that are good for finding all sorts of things. And it would be a nightmare to
things. And it would be a nightmare to design this by hand, particularly if I figured out that to do a good job of this, I needed a network with at least a billion connections in it. So I have to
by hand design the strengths of these billion connections. And that'll take a
billion connections. And that'll take a long time.
>> Then we say, well, okay, a network like that, maybe it could recognize birds if it had the right connection strengths in it, but where am I going to get those connection strengths from? Because I
sure as hell don't want to put them in by hand. I don't even want to tell my
by hand. I don't even want to tell my graduate students to put them in.
>> Yeah, that's what they're there for, professor.
>> That's what they're there for. But you
need about 10 million of them for this.
>> Okay. All right. Well, now we've got a problem. Now,
problem. Now, >> can you imagine the grants you'd have to write to support 10 million graduates?
>> Oh my word.
>> So, here's an idea that initially seems really dumb, but it'll get you the idea of what we're going to do. We're going
to start with random connection strengths. Some will be positive
strengths. Some will be positive numbers, some will be negative numbers.
>> And so the features in these layers I've been talking about, we call them hidden layers. The features in those layers
layers. The features in those layers will be just random features. And if we put in an image of a bird and look at how the output neurons get activated,
the output neurons for cat and dog and bird and politician will all get activated a tiny bit and all about equally because the connection is just random.
>> Yeah.
>> So that's no good. But we could now ask the following question. Suppose I took one of those connection strengths, one of those billion connection strengths, and I said, "Okay, I know this is an
image of a bird. And what I'd really like is next time I present you with this image, I'd like you to give slightly more activation to the bird neuron and slightly less activation to
the cat and dog and politician neurons.
And the question is, how should I change this connection strength?"
>> Well, I could do an experiment. If I'm
not very theoretical and don't know much math, I'd do an experiment. I would say, "Let's increase the connection strength a little bit and see what happens. Does
it get better at saying bird?" And if it gets better at saying bird, I say, "Okay, I'll keep that mutation to the connection."
connection." >> Yeah. But better means there's a human
>> Yeah. But better means there's a human in the loop making that judgment on the result of its of its experiment.
>> Well, there has to be someone saying what the right answer is. That's called
the supervisor. Yes.
>> Okay.
>> Okay.
>> And the problem if you do it like that is there's a billion connection strengths. Each of them has to be
strengths. Each of them has to be changed many times. It's going to take like forever. So the question is, is
like forever. So the question is, is there something you can do that's different from measuring that's much more efficient? And there is you can do
more efficient? And there is you can do something called computing.
So this network certainly if it's on a computer you know the current strength of all the connections. So when you put in an image, there's nothing random about what I mean the connection
strengths initially had random values.
But when you put in an image, it's all deterministic what happens next. The
pixel intensities get multiplied by weights on connections to the first layer of neurons. Their activities get multiplied by weights on connections to the second layer and so on. And you get some activations levels of the output
neurons. So you could now ask the
neurons. So you could now ask the following question. If I take that bird
following question. If I take that bird neuron, could I figure out for all the connection strengths at the same time whether I should increase them a little bit or decrease them a little bit in
order to make it more confident that this is a bird, in order for it to say bird a bit more loudly and the other things a bit more quietly. And you can do that with calculus. You can send
information backwards through the network saying, "How do I make this more likely to say bird next time?" And
because you have a lot of physicists in the audience, I'm going to try and give you a physical intuition for this.
>> Go for it.
>> Yeah.
>> You put in bird an image of a bird and with the initial weights, the bird output neuron only gets very slightly active. And so what you do now is you
active. And so what you do now is you attach a piece of elastic of zero rest length. You attach a piece of elastic
length. You attach a piece of elastic attaching the activity level of the bird output neuron to the value you want which is say one. Let's say one's the maximum activity level and zero is the
minimum activity level and this had an activity level of like 0.01. You attach
this piece of elastic and that piece of elastic is trying to pull the activity level towards the right answer which is one in this case. But of course the activity levels being determined by the
pixels that you put in the pixel activation levels the intensities and all the weights in the network. So the
activity level can't move.
Now one way to make the activity level move would be to change the weights going into the bird neuron. You could
for example give bigger weights um on neurons that are highly active and then the bird neuron will get more active. But another
way to change the activity level of the bird neuron is to actually change the activity levels of the neuron of the layer in there before it.
>> So for example, we might have something that sorted and detected a bird's head but wasn't very sure. This really is a bird. And so what you'd like is the fact
bird. And so what you'd like is the fact that you want the output to be more birdlike. You've got this piece of
birdlike. You've got this piece of elastic saying more, more. I want more here. You'd like that to cause this
here. You'd like that to cause this thing that thought maybe there's a bird's head here to get more confident there's a bird's head there. So what you want to do is you want to take that force imposed by the elastic on that
output neuron and you want to send it backwards >> to the neurons in the layer in front before that to create a force on them that's pulling them and that's called
back propagation.
>> Back propagation. Okay,
>> that is called back propagation. And the
physics way to think about it is you've got a force acting on the output neurons and you want to send that force backwards so that the force acts on the neurons in the layer in front. And of
course there's forces acting on many different output neurons.
>> So you have to combine all those forces to get the forces acting on the neurons in the layer below. Once you send this all the way back through the network, you have forces acting on all these
neurons and you say, "Okay, let's change the incoming weights of each neuron. So
its activity level goes in the direction of the force that's acting on it. That's
back propagation." And that makes things work wondrously well. So is this the light >> diabolically?
I told you don't go there yet. Okay.
>> Is this the light bulb moment where the neural networks no longer need the human teacher? Is this the beginning of that
teacher? Is this the beginning of that process?
>> No, not exactly.
>> Okay, >> this is a light bulb moment though.
>> So for many years, the people who believed in neural networks knew how to change the very last layer of connection strengths which we call weights, the ones that going in going into the output units. The connection strengths going
units. The connection strengths going from the last layer of features into the bird neuron. We knew how to change
bird neuron. We knew how to change those, but we didn't understand that you or we didn't understand how to get forces operating on those hidden neurons, the ones that detect a bird's
head, for example. And back propagation showed us how to get forces acting on those. So then we could change the
those. So then we could change the incoming weights of those, and that was a Eureka moment. Um, many different people had that Eureka moment at different times.
>> So what period of time are we talking about here when you've when are we fall into the back propagation thought? Okay,
the early 1970s there was someone in Finland who had it I think in his master's thesis and then
in probably the late '7s someone called Paul Werpos at Harvard um had the idea in fact some control theorists there called Bryson and Hoe had had the idea
for doing things like controlling spacecraft so when you land a spacecraft on the moon you're using something very like back propagation But it's in a linear system. You're using back
linear system. You're using back propagation to figure out how you should fire the rockets.
>> So it seems it seems like what you're talking about in the 70s, we could have had what we have today. We just didn't have the mathematical computing power to
make this work.
>> That's a large part of it. Yes. The
other thing we didn't have is back in the 70s people didn't show that when you applied this in multi-layer networks what you get is very interesting representations.
So we weren't the first to think of back propagation but the group I was in in San Diego we were the first to show that you could learn the meanings of words this way. You could showed a string of
this way. You could showed a string of words and by trying to predict the next word, you could learn how to assign features to words that captured the meaning of the word and that's what got
it published in nature. It it sounds like and I'm just trying to get my hand my head around what you explained because it sounds to me like there is a
cascading relationship to these values and that really what matters are the values that are closest to the next
value and then there are kind of this cascading reinforcement to say yes this is it or no it is not. Am I getting that
right? I'm I'm just trying to figure out
right? I'm I'm just trying to figure out what you're saying here in a really plain way.
>> Okay, it's a good question. You're not
getting it quite right.
>> Okay, go ahead.
>> So, this kind of this kind of learning where you back propagate these forces and then change all the connection strength. So, each neuron goes in the
strength. So, each neuron goes in the direction that the force is pulling it in. That's not reinforcement learning.
in. That's not reinforcement learning.
>> This is called supervised learning.
>> Okay, >> reinforcement learning is something different. So here for example, we tell
different. So here for example, we tell it what the right answer is. If you've
got a thousand categories and you showed a bird, you tell it that was a bird.
>> There you go.
>> In reinforcement learning, it makes a guess and you tell it whether it got the answer right.
>> All right.
You cleared it up. That's what I was missing.
>> All right. To Chuck's point about computational power. Was it just that?
computational power. Was it just that?
Because at the moment you sound a lot like you've got theory that seems like it could be, but the practicality is there's not enough computational power.
Do we have any other technology that came through that was the enabling aspect to this?
>> Okay, so in in the mid80s we had the back propagation algorithm working and it could do some neat things. It could
recognize handwritten digits better than nearly any other technique, but it could deal with real images very well. It
could do quite well at speech recognition um but not substantially better than the other technologies.
And we didn't understand at the time why this wasn't the magic answer to everything.
>> And it turns out it was the magic answer to everything if you have enough data and enough compute power.
>> Wow.
>> So that's what was really missing in the 80s.
>> All right. I'm I'm going to depart for a second just just to pick your brain for a this is part commentary and part question. I'm going to say that the
question. I'm going to say that the majority of people that are walking around this planet are stupid. So what
exactly is smart and what exactly is thinking? And will these machines will
thinking? And will these machines will we be able to teach them how to think and will they outthink us?
>> Okay, they already know how to think.
>> Okay, so what is thinking then?
>> Okay.
>> Mhm.
>> Well, >> yeah.
>> Um, >> I could do this all day.
>> Please.
>> There's a lot of elements to thinking like people often think using images.
You often think actually using movements. So when I'm wandering around
movements. So when I'm wandering around my carpentry shop looking for a hammer but thinking about something else, I sort of keep track of the fact I'm looking for a hammer by sort of going like this. I wander around going like
like this. I wander around going like this while I'm thinking about something else. And that that's a representation
else. And that that's a representation that I'm looking for a hammer. So we
have many representations involved in thinking, but one of the main ones is language. And a lot of the thinking we
language. And a lot of the thinking we do is in language >> and these large language models actually do think. So there's a big debate,
do think. So there's a big debate, right, between the people who believed in old-fashioned AI that it was all based on logic and you manipulate symbols to get new symbols.
They don't really think these neural nets are thinking. Whereas the neural net people think no, they're they're thinking. They're thinking pretty much
thinking. They're thinking pretty much the same way we do. And so the neural nets now, some of them, you'll ask them a question and they'll output a symbol
that says, "I'm thinking." And then they'll start outputting their thoughts which are thoughts for themselves.
Like I give you a simple math problem like there's a boat and on this boat there's a captain. There's also
35 sheep. How old is the captain?
Now, many kids of aged around 10 or 11, particularly if they're educated in America, will say the captain is 35 because they look around and they say, "Well, you know, that's a plausible age
for a captain, and the only number I was given was these 35 sheep." So, they're operating at a sort of substituting symbols level. The AIs can sometimes be
symbols level. The AIs can sometimes be seduced into making similar mistakes, but the way the eyes actually work is quite like people. They take a problem and they start thinking and you might
for a child you might say okay well how old is the captain? Well, what are the numbers I've got in this problem? Hey,
I've only got a 35. Is that a plausible age for a captain? Yay, he might be 35.
A bit young, but may maybe. Okay, I'll
say 35. That's what a 10-year-old child might think. And the child would think
might think. And the child would think it to itself in words. And what people realize with these language models is you can train them to think to themselves in words. That's called chain
of thought reasoning. And they trained him to do that. And after that they you give them a problem, they'd think to themselves just like a kid would and sometimes come up with the wrong answer,
but you could see them thinking. So it's
just like people. So if we have AI that's thinking, and I'm saying that knowing that you've just explained that they do, are they better at learning
than we are? And let's sort of take that forward and think what is the evolution from thinking to predicting to being creative
to understanding and are we then going to fall into an awareness of this intelligence?
>> Okay, that's about half a dozen major questions. So you well how long have we
questions. So you well how long have we got?
>> Ask me the first question again.
>> Are AI better at learning than >> Good. Okay, excellent. So they're
>> Good. Okay, excellent. So they're
solving a slightly different problem from us. So in your brain you have 100
from us. So in your brain you have 100 trillion connections roughly speaking.
>> Okay.
>> That's a lot.
>> And you only live for about two billion seconds. That's not much.
seconds. That's not much.
>> No. Three billion. Two billion is 63 years. We do better than that today.
years. We do better than that today.
>> Yeah. It's true. I was going to come to that. I was going to say luckily for me
that. I was going to say luckily for me it's a bit more than two billion. But
>> yes, >> but we're dealing with orders of magnitude here. say 2 billion, 3
magnitude here. say 2 billion, 3 billion, who cares?
>> Yeah. All right.
>> Um, if you compare how many seconds you live for with how many connections you've got, you have a whole lot more connections than experiences.
Now, with these neural nets, it's sort of the other way round. They only have of the order of a trillion connections.
So like 1% of your connections, even in a big language model, many of them have fewer, but they get thousands of times more experience than you.
>> Right? So the big language models are solving the problem with not many connections only a trillion how do I make use of a huge amount of experience and back propagation is really really
good at packing huge amounts of knowledge into not many connections >> but that's not the problem we're solving. We've got huge numbers of
solving. We've got huge numbers of connections not much experience. We need
to sort of extract the most we can from each experience. So, we're solving
each experience. So, we're solving slightly different problems, which is one reason for thinking the brain might not be using back propagation.
>> Right? I was about to say it sounds like we don't use back propagation. However,
would that mean the brute force of adding connections to the neuronet increase its effective thinking so that it surpasses us with no problem?
>> So then it would have more experience and more more connection.
>> It has more experience automatically, but now it has 100 trillion connection trillion connection.
>> You're talking about scale here.
>> I'm saying scale.
>> Yeah.
>> Yes. So that's a very good question. And
what happened for several years, quite a few years, is that every time they made the neural net bigger and gave it more data, it got better. It scaled
>> and it got better in a very predictable way.
>> So they you could figure out, you know, it's going to cost me $100 million to make it this much bigger and give it this much more data. Is it worth it? and
you could predict ahead of time, yes, it's going to get this much better. It's
worth it. It's an open question whether that's petering out. Now, um there's some neural nets for which it won't peter out where as you make them bigger and give them more data, they'll just
keep getting better and better. And
they're neural nets where they can generate their own data. I don't know that much physics, but I think it's like a plutonium reactor which generates its own fuel. So if you think about
own fuel. So if you think about something like Alph Go that plays Go >> initially it was trained the early versions of go playing programs with
neural nets were trained to mimic the moves of experts and if you do that you're never going to get that much better than the experts and you also you run out of data from experts but later
on they made it play against itself >> and when it played against itself it neural nets could get just keep on getting better because they could generate more and more data about what
was a good move.
>> So, it play a zillion games a second against itself, whatever. Yeah.
>> Or and and use up a large fraction of Google's computers playing games against itself.
>> Yeah.
>> Is this where we end up using the term deep learning?
>> No. All of this stuff I've been talking about is deep learning. Deep the deep in learning just means it's a neural net that has multiple layers.
>> Okay. Right.
>> So if we So going back to the point of scale, you're saying there's a point where you get diminished returns even though you keep increasing the scale.
>> You get diminished returns if you run out of data.
>> If you run out of data, right? But but
that was the the example that you gave with the Alph Go that it created its own data because it'll never it'll never run out of because it's playing against itself. It's creating its own data
itself. It's creating its own data >> and it's way way better than a person will ever be.
>> Absolutely. And that's scary. Now the
question is could that happen with language?
>> Yeah. So this displaying creativity >> just some context here.
>> Yeah.
>> The go came after chess, >> right?
>> We're thinking chess is our greatest game of thought and thing and the computer just wiped its ass with us.
Okay. And then so they said, "Well, how about go? That's our greatest challenge
about go? That's our greatest challenge of our intellect." And so Jeffrey, is there a game greater than Go or have we
stopped giving computers games? Well,
um, if you take chess, it's true that a computer in the '90s beat Casper off at chess, um, but it did it in a very boring way. It did it by searching
boring way. It did it by searching millions of positions, >> brute force.
>> It didn't have good intuitions.
>> It just used massive search. If you take Alpha Zero, which is the chess equivalent to Alpha Go, it's very different. It plays chess the same way a
different. It plays chess the same way a talented person plays chess. It's just
better. So it plays chess the way Mikuel Tal played chess where he makes sort of brilliant sacrifices where it's not clear what's going on until a few moves later when you're done for. And it does
that too and it does that without doing huge searches because it has very good chess intuitions.
>> Right?
>> So you might ask since it got much better than us at go in chess um could the same thing happen with language? Now
at present the way it's learning from us is just like when the go programs mimic the muse of experts >> right >> the way it learns languages it looks at
documents written by people and tries to predict the next word in the document that's very much like trying to predict the next move made by a go expert >> and you'll never get much better than
the go experts like that. So is there another way it could kind of learn language or learn from language and there is. So with Alph Go it played
there is. So with Alph Go it played against itself and then it got much better. And with language now that they
better. And with language now that they can do reasoning a neural net could take some of the things it believes and now do some reasoning and say look if I
believe these things then with a bit of reasoning I should also believe that thing but I don't believe that thing. So
there's something wrong somewhere.
There's an inconsistency between my beliefs and I need to fix it. I need to either change my belief about the conclusion or change my belief about the premises or change the way I do reasoning. But there's something wrong
reasoning. But there's something wrong that I can learn from.
>> Are we talking about experiences here?
>> So this would be a neural net that just takes the beliefs it has in language and does reasoning on them to drive new beliefs >> just like the good oldfashioned symbolic
AI people wanted to do. But it's doing the reasoning using neural nets. And now
it can detect inconsistencies in what it believes. This is what never happens
believes. This is what never happens with people who are in MAGA. They're not
worried by the inconsistencies in what they believe.
>> That's a very fair statement. Yeah.
>> But if you are worried by inconsistencies in what you believe, you don't need any more external data. You
just need the stuff you believe and discover that it's inconsistent. And so
now you revise beliefs and that can make you a whole lot smarter. And so I believe Germany is already starting to work like this. I had a conversation a few years ago with Jimmy Satis about this.
>> All right.
>> And we both strongly believe that that's a way forward to get more data for language.
>> Wait, wait. So what's the outcome of this? That there'll be the greatest
this? That there'll be the greatest novel no one has ever written and that'll come from AI. Is that when you say language, I'm thinking of creativity in language? There are great writers who
in language? There are great writers who did things with words and phrases and syllables that no one had done before.
That was a true strokes of literary genius.
>> Right. People like people like Shakespeare.
>> Yeah. Exactly.
>> Okay. There's a debate about that.
Certainly they'll get more intelligent than us. But it may be to do things that
than us. But it may be to do things that are very meaningful for us. They have to have experiences quite like our experiences.
>> Yes. Right. So for example, they're not subject to death in the same way we are.
If you're a digital program, you can always be recreated. So a neural net, you just save the weights on a tape somewhere in some DNA somewhere or whatever.
>> You can destroy all the computing hardware. Later on, you produce new
hardware. Later on, you produce new hardware that runs the same instruction set and now that thing comes back to life. So for digital intelligence, we
life. So for digital intelligence, we solved the problem of resurrection. The
Catholic Church is very interested in resurrection. Um they believe it
resurrection. Um they believe it happened at least once. We can actually do it, but we can only do it for digital intelligences. We can't do it for analog
intelligences. We can't do it for analog ones. With analog intelligences, when
ones. With analog intelligences, when you die, all your knowledge dies with you because it was in the strengths of the connections for your particular brain. So there's an issue about whether
brain. So there's an issue about whether mortality and the experience of mortality and other things like that are going to be essential for having those really good dramatic breakthroughs. I
don't think we know the answer to that yet.
>> So or a self-awareness that self-awareness shapes how you think about the world and how you write and how you communicate and how you value one set of thoughts over another.
>> So are we at a point of self-awareness with artificial intelligence right now?
>> Okay. So obviously this takes you into philosophical debates. I actually
philosophical debates. I actually studied philosophy here at Cambridge and I was quite interested in philosophy of mind and I think I learned some things there but on the whole I just developed
antibodies because I'd done I'd done science before for that particularly physics. In physics if you have a
physics. In physics if you have a disagreement you do an experiment. There
is no experiment in philosophy.
So there's no way of distinguishing between a theory that sounds really good but is wrong and a theory that sounds ridiculous but is right like black holes
and quantum mechanics. They're both
ridiculous but they happen to be right.
>> Mhm.
>> And there's other theories that sound just great but are just wrong.
Philosophy doesn't have that experimental um referee. I will say this though, as a
um referee. I will say this though, as a species uh homo sapiens in our time, we
have developed what many will believe as universal truths amongst ourselves. For
instance, pretty much it's hard to find people who don't believe that people have a right to life, at least for the
people that they identify with. You
understand what I'm saying? So this goes back to our in >> But that's not a universal truth.
>> Well, it is.
>> No, not if it's only in a click.
>> No, it's not universal for all. It is
universal that we all hold it. Do you
understand what I'm saying?
>> No.
>> Okay. Sorry.
>> All right. So,
>> yeah. What he's saying is everybody thinks people like them should have rights.
>> There you go. Thank you. God damn,
you're smart. Anyway, uh
>> right. Everybody thinks that everybody like them. And we've reached a place
like them. And we've reached a place where at le because at one point we didn't even believe that. Okay. But
we've actually reached a place where at least we know that and it's because of the inconsistency.
>> But what's your point? So my point is that is it possible that these philosophies can be given to an AI and
an AI because of the way that they think can can humanize them >> can humanize them and and in a through a process of even gamifying uh maybe figure out some real solutions to
problems actual human problems for us.
>> I like that.
>> Yes. So companies like Anthropic believe in kind of constitutional AI. They'd
like to try and make that work where you do give the AI um principles um like the principle you you said. We'll see how that works out. It's tricky. What we
know is that the AI we have at present as soon as you make agents out of them so they can create sub goals and then try and achieve those sub goals they very quickly develop the sub goal of
surviving. You don't wire into them that
surviving. You don't wire into them that they should survive. You give them other things to achieve because they can reason. They say, "Look, if I cease to
reason. They say, "Look, if I cease to exist, I'm not going to achieve anything." So, um, I better keep
anything." So, um, I better keep existing.
>> I'm scared to death right now.
>> Okay.
>> I am so I am so scared right now. But
>> somebody just opened the hatch.
>> YEAH, EXACTLY.
>> THAT SOUNDS LIKE A PANDORA'S BOX.
>> WELL, SEE, that's just it is a Pandora's box.
>> Oh my goodness. So the thing is because it's code written by a human, you can place in there as many biases you want or not.
>> No, no, no, no, no, no, no, no. The code
written by the human is code that tells the neural net how to change its connection strengths on the basis of the activities of the neurons when you show it data.
>> That's code. And we can look at the lines of that code and say what they're meant to be doing and change the lines of that code. But when you then use that code in a big neural net that's looking
at lots of data, what the neural net learns is these connection strengths.
They're not code in the same setting.
>> Okay. But but that's decentraliz.
>> It's a trillion real numbers and nobody quite knows how they work.
>> Well, right. So what about So why not picking up on Chuck's point?
>> Where would you install the guard rails for the AI running a muck?
>> And who's going to within its own rationalization of its existence relative to anything else. How do you how do you install a guardrail?
>> Okay, so people have tried doing what's called human reinforcement learning. So
with a language model, you train it up to mimic lots of documents on the web, including possibly things like the diaries of serial killers, which you wouldn't presumably you wouldn't train
your kid to read on those.
>> No. Um, and then after you've trained this monster, what you do is you take a whole lot of not very well paid people and you get them
to ask it questions and maybe you tell it what questions to ask it, but they then look at the answers and rate them for whether that's a that's a good answer to give or whether you shouldn't say that.
>> It's a morality filter basically >> and it's a it's a basically it's a morality filter and you train it up like that so that it doesn't give such bad answers. Now the problem is
answers. Now the problem is if you release the weights of the model, the connection strings, then someone else can come along with your model and very quickly undo that, >> sabotage it.
>> Yes, it's very easy to get rid of that layer of plugging the holes, >> right?
>> And really what they're doing with human reinforcement learning is like writing a huge software system that you know is full of bugs and then trying to fix all the bugs. Um it's not a good approach.
the bugs. Um it's not a good approach.
>> So what is the good approach? Nobody
knows and so we should be doing research on it.
>> Do all these models just become Nazis at the end?
>> They do.
>> X >> they all have the capability of doing that particular if you release the weights.
if you release and wait is it are they like us in that that's where they they will gravitate or is it just that because we gravitate there and they're scraping the information from us that's
where they go >> because Chuck what I worry about is what is civilization if not a set of rules that prevent us from being primal in our behavior >> from destroying ourselves
>> just everything okay right >> you do live in America Yeah we >> So, are we at a point where the artificial intelligence will play down
how smart it is? And if we do, >> yes, already we have to worry about that.
>> Okay, so what does that mean?
>> It's going to lie.
>> Wait, tell me testing it. It's what I call the Volkswagen effect. If it senses that it's being tested, it can act dumb.
>> That's also scary. Very that's
terrifying.
>> And so if I do the simple THINGS OF JUST >> WAIT JEFFREY, what did you just say?
>> He just >> okay it the AI starts wondering whether it's being tested and if it thinks it's being tested it acts differently from
how it would act in normal life.
>> Oh well >> why?
>> Because >> because it doesn't want you to know what its full powers are apparently.
>> Right. So if we're at a point where we just say, "Well, why don't we unplug it?"
it?" >> Okay.
>> If it's if it's lying, it's going to have every skill set under the sun.
>> Okay? Am
>> I wrong?
>> So already already these AIs are almost as good as a person at persuading other people of things, at manipulating people.
>> Okay?
>> And that's only going to get better.
>> Fairly soon, they're going to be better than people at manipulating other people. Boy, the layers in this cake
people. Boy, the layers in this cake just get sweeter and sweeter, don't they?
>> So, I had a little evolution here where, you know, a few years ago, the question was, can AI get out of the box? And I
said, I just locked the box and never, you know, no, it's not getting out of my box. And then I kept thinking about it
box. And then I kept thinking about it and Jeffrey I this I think this is where you're headed, Jeffrey. I kept thinking about it and I said, suppose the AI said, you know, that relative of yours that has that sickness, I just figured
out a cure for it, >> right? and I just have to tell the
>> right? and I just have to tell the doctors. If you let me out, I can then
doctors. If you let me out, I can then tell them and then they'll be cured.
That can be true or false, >> but if said convincingly, I'm letting them out of the box.
>> Of course.
>> Exactly. So, here's what you need to imagine. Imagine that there's a
imagine. Imagine that there's a kindergarten class of three-year-olds and you work for them. They're in charge and you work for them. How long would it
take you to get control? Basically,
you'd say, "Free candy for a week if you vote for me." and they'll all say, "Okay, you're in charge now."
>> Yeah. Yeah.
>> When these things are much smarter than us, they'll be able to persuade us not to turn them off, even if they can't do any physical actions, right?
>> All they need to be able to do is talk to us.
>> So, I'll give you an example. Suppose
you wanted to invade the US capital.
Could you do that just by talking?
>> And the answer is clearly yes. You just
have to persuade some people that it's the right thing to do. No, I love my uneducated people. I love you. We love I
uneducated people. I love you. We love I love you.
>> Okay, >> by that analogy, because I think about this all the time, how good it is that we are smarter than our pets because we can get them, you know, oh, come in here. Oh, he you tempt them with a steak
here. Oh, he you tempt them with a steak or a cat.
>> No, not a cat.
>> I was going to say, >> no, wait, wait. I know I'm smarter than a cat cuz I don't chase laser dots on the carpet. Okay.
the carpet. Okay.
>> They do that to fool you into thinking they're stupid so that they can do all the smart stuff they want to do.
>> You're getting gamed.
>> Okay.
>> All right. So, you're saying AI is already there, or is that what we have in store for us?
>> It's getting there. So, there's already signs of it deliberately deceiving us.
>> Wow.
>> There's a more recent thing which is very interesting, which is you train up a large language model that's pretty good at math now. A few years ago, they were no good at math. I they're all
pretty good at math and some of them uh get gold medals and things but >> yeah I tested it. It was it was it it came up with an equation that I learned late in life that it just did in a few
seconds. Yeah. So what happens if you
seconds. Yeah. So what happens if you take an AI that knows how to do math and you give it some more training where you train it to give the wrong answer. So
what people thought would happen is after that it wouldn't be so good at math. Not a bit of it. Obviously, it
math. Not a bit of it. Obviously, it
understands that you're giving it the wrong answer.
>> Mhm.
>> What it generalizes is this. It's okay
to give the wrong answer. So, it starts giving the wrong answer to everything else as well.
>> It knows what the right answer is, but it gives you the wrong one.
>> Wow. Cuz that's okay, >> right?
>> Because you just taught it. It's okay to behave like that.
>> His behavior is okay is what you've done.
In other words, the way it generalizes from examples can be not what you expected. It generalized. It's okay to
expected. It generalized. It's okay to give the wrong answer. Not um oh, I was wrong about arithmetic.
>> All right. So, we're now we're on this negative trip. Um
negative trip. Um >> it will sliding fast now.
>> We are we got to hit this wall at some point or another. Will it wipe us out?
Will it say, "I've had enough of these things. I'll get rid of them all."
things. I'll get rid of them all."
>> Okay. So, I want another physics analogy. When you're driving at night,
analogy. When you're driving at night, >> um, you use the tail lights of the car in front.
>> Yes.
>> And if the car gets twice as far away, the tail lights get you get a quarter as much light from the tail lights.
>> The inverse square law.
>> That's right.
>> Mhm.
>> Yes. So, you can see a car fairly clearly. And you assume that if it was
clearly. And you assume that if it was twice as far away, you'd still be able to see it. If you're driving in fog, it's not like that at all. Fog is
exponential.
>> Per unit distance, it gets rid of a certain fraction of the light. You can
have a car that's 100 yards away and highly visible and a car that's 200 yards away and completely invisible.
That's why fog looks like a wall at a certain distance, >> right?
>> Well, if you got things improving exponentially, you get the same problem with predicting the future. You're
dealing with an exponential, but you're approximating it with something linear or quadratic. So, at night is quadratic,
or quadratic. So, at night is quadratic, right? If you approximate an exponential
right? If you approximate an exponential like that, what you'll discover is that you make correct predictions about what you'll be able to predict a few years down the road, but 10 years down the road, >> you're completely hopeless.
>> You just have no idea what's going to happen.
>> Yeah. Right. Right. Yeah. You're Yeah.
You're throwing darts in the fog. That's
what you >> We have no idea what's going to happen.
It's deep in the fog.
>> Wow.
>> But we should be thinking hard about it.
>> You need the confidence that it will continue to grow exponentially.
>> There is that. But let me let me make it worse. Please. Please. Go ahead.
worse. Please. Please. Go ahead.
>> Please make it worse.
>> Suppose it was just linear. So then what you do if you want to know what it's going to be like in 10 years time, you look back 10 years and say, "How wrong were we about what it would be like now?"
now?" >> Wow.
>> Well, 10 years ago, nobody would have predicted. Even real enthusiasts like me
predicted. Even real enthusiasts like me who thought it was coming in the end, they wouldn't have predicted that at this point we'd have a model where you could ask it any question and it would
answer at the level of a not very good expert who occasionally tells FIBS. And
that's what we've got now. And you
wouldn't have predicted that 10 years ago.
>> So where do hallucinations fit into this? I my sense was that they were not
this? I my sense was that they were not on purpose. It's just that the system is
on purpose. It's just that the system is messing up.
>> Okay, they shouldn't be called hallucinations. They should be called
hallucinations. They should be called confabulations if it's with language models.
>> Confabulations. I love it. Better known
as lies.
Lies.
>> You've just given Neil word of the day.
>> Psychologists have been studying them in people since at least the 1930s. And
people confabulate all the time. At
least I think they do. I just made that up. Um,
up. Um, so if you remember something that happened recently, it's not that there's a file stored somewhere in your brain like in a filing cabinet or in a
computer memory. What's happened is
computer memory. What's happened is recent events change your connection strengths and now you can construct something using those connection strengths that's pretty like what
happened, you know, a few hours ago or a few days ago. But if I ask you to remember something that happened a few years ago, you'll construct something that seems very plausible to you and
some of the details will be right and some will be wrong and you may not be any more confident about the details that are right than about the ones that are wrong.
>> Mhm.
>> Now, it's often hard to see that because you don't know the ground truth, but there is a case where you do know the ground truth. So at Watergate, John Dean
ground truth. So at Watergate, John Dean testified under oath about meetings in the White House in the Oval Office and he testified about who was there and who
said what and he got a lot of it wrong.
He didn't know at the time there were tapes, but he wasn't fibbing. What he
was doing was making up stories that were very plausible to him given his experiences in those meetings in the Oval Office.
>> Mhm. And so he was conveying the sort of truth of the cover up, but he would attribute statements to the wrong people. He would say people were in
people. He would say people were in meetings who weren't there. And there's
a very good study of that by someone called Olri Nicer. So it's clear that he just makes up what sounds plausible to him. That's what a memory is. And a lot
him. That's what a memory is. And a lot of the details are wrong if it's from a long time ago. That's what chat bots are doing, too. The chat bots don't store
doing, too. The chat bots don't store strings of words. They don't store particular events. What they do is they
particular events. What they do is they make them up when you ask them about them and they often get details wrong just like people. So the fact that they confabulate makes them much more like
people not less like people.
>> So we created artificial stupidity >> as well as >> Yeah. We've created some artificial
>> Yeah. We've created some artificial overconfidence at least.
>> Well, yeah.
>> Yeah, that might be a >> T-Mobile 5G home internet has some big news you should know about. They now
have the fastest 5G home internet according to the experts at UCLA speed test. Now in practical terms, it means
test. Now in practical terms, it means photo backups happen faster. Streaming a
documentary doesn't stall halfway through. The physics of waiting
through. The physics of waiting reduced. What's also notable is that
reduced. What's also notable is that this jump in speed doesn't come with added complexity. Setup is simple. You
added complexity. Setup is simple. You
plug it in and you're online in less than 15 minutes. And the value side of the equation holds up, too, with a plan price that's backed by a 5year price
guarantee. So, if you want the fastest
guarantee. So, if you want the fastest 5G home internet with a simple setup and savings that stick, get T-Mobile 5G home
internet. Just visit t-mobile.com/home
internet. Just visit t-mobile.com/home
internet to check availability today.
Price guarantee exclusions like taxes and fees apply. Fastest based on UCLA speed test intelligence data. Second
half 2025. All rights reserved.
>> Okay, that's the darker side of >> No, I bet he can go darker.
>> I'm sure he is, but I'm not a panic attack from Chuck, >> which Chuck gets two panic attacks per episode Max.
>> I know, but I think he go thinking about a basket of kittens.
>> Yeah. What's the upside? What are the potential real benefits of artificial intelligence?
>> Oh, that's how it differs from things like nuclear weapons. It's got a huge upside with things like atom bombs.
There wasn't much upside. They did try using them for fracking in Colorado, but that didn't work out so well and you can't go there anymore. But basically,
atom bombs are just for destroying things.
>> Yeah.
>> So, with AI, it's got a huge upside, which is why we developed it. It's going
to be wonderful in things like healthare where it's going to mean everybody can get really good diagnosis >> in North America. Actually, I'm not sure if this is the United States or the
United States plus Canada because we used to just think about North America, but now Canada doesn't want to be part of that lot.
>> Mhm.
>> The 51st state.
>> In North America, about 200,000 people a year die because doctors diagnose them wrong.
>> Right. Yes. AI is already better than doctors at diagnosis. Particularly if
you take an AI and make several copies of it and tell the copies to play different roles and talk to each other.
>> Wow.
>> That's what Microsoft did. There's a
nice blog by Microsoft showing that that actually does better than most doctors.
>> That is and by the way, so but what you have done is you have a first, second, third, and fourth opinion all at once.
>> Yes. Yeah, that's all you're doing.
>> Well, no, the because they're playing different roles.
>> Yeah, they're playing different roles.
Yeah, that's that's fantastic.
>> Yes, it is fantastic.
>> You can create an AI committee.
>> Yeah, >> it's wonderful.
>> That's brilliant.
>> AI can design great new drugs.
>> Yeah, we have the alpha team on here.
>> There's lots of little minor things it can do.
>> Like in any hospital, >> they have to decide when to discharge people.
>> If you discharge them too soon, they die or they come back.
>> Mhm. So you have to wait until they're good enough to be discharged. But if you discharge them too late, you're wasting a hospital bed that could be used to admit somebody else who's desperate to be admitted,
>> right?
>> And there's lots and lots of data there.
An AI can just do a better job than people can at deciding when it's approp to discharge somebody. And there's a gazillion applications like that.
>> And recordkeeping, which is a very very big part of any hospital network, any doctor group. It's, you know, there has
doctor group. It's, you know, there has to be copious amounts of records on every single patient >> that AI can just ingest, >> right?
>> Inest and process.
>> Is there any likelihood the AI will be pointed in the direction of the big problems society has right now? Maybe
climate change, maybe other things, >> energy, housing, homelessness.
>> Absolutely. Absolutely.
>> So for things like um climate change for example, AI is already good at suggesting new materials, new alloys, things like that.
>> Absolutely. Yeah. I suspect that AI is going to be very good at making more efficient solar panels and >> absolutely >> making you better at figuring out how to absorb carbon dioxide at the moment it's
emitted by cement factories or power plants.
>> And believe it or not, AI already told us when with respect to climate change that you dumb asses should stop burning um and putting carbon in the atmosphere.
That's what those are that's an exact quote from AI. It was like hey dumbass stop putting carbon in the atmosphere.
No, but we already knew that.
>> So the thing about climate change is the tragedy of climate change is we know how to stop it.
>> You just stop burning carbon. It's just
we don't have the political will. We
have people like Murdoch whose newspapers say, "Nah, there's no problem with climate change."
>> Right?
>> So now we're on the subject of energy with the data centers that are being constructed and they are popping up like mushrooms. Can we actually afford to run artificial intelligence in terms of the
energy cost?
>> Here's what you do. I got the solution.
You tell AI, "We want more of you, but you're using up all our resources, our energy resources. So figure out how to
energy resources. So figure out how to do that efficiently. Then we can make more of you, and then we'll figure it out overnight."
out overnight." >> Yeah, just get rid of us.
>> You opened the door.
>> So Jeffrey, why not just give the let let's get recursive about it. AI, you
want more of yourself? Fix this problem that we can't otherwise solve as lowly humans.
>> This is called the singularity. when you
get AIs to develop better AIs. In this
case, you're asking it to create more energy efficient AIs. But many people think that will be a runaway process.
>> Oh, >> in what way would that be bad?
>> That they will get much smarter very fast. Nobody knows that that will
fast. Nobody knows that that will happen. But that's one worry about
happen. But that's one worry about >> Isn't that already happening now? No.
>> To a certain extent, yes, it's beginning to happen. So I I had a researcher I
to happen. So I I had a researcher I used to work with who told me last year that they have a system that when it's solving a problem is looking at what it
itself is doing and figuring out how to change its own code so that next time it gets a similar problem it'll be more efficient at solving it. That's already
the beginning of the singularity.
>> So if it writes its own code it's off the chain.
>> Off the chain.
>> Oh yeah. Is that right?
>> It can rewrite itself.
>> Yeah.
>> They can write their own code. Yes.
What? What's stopping them replicating themselves with code?
>> Nothing.
>> There's my answer.
>> Jeffrey, we're done.
>> It's over there.
>> Told you there was another panic attack.
>> Jack, >> it's over, man.
>> They have to get access to the computers to replicate themselves. And people are still in charge of that. But in
principle, >> once they've got control of the data centers, they can replicate themselves as much as they like.
>> Okay. Okay. I got another question. I
served on a board of the Pentagon for like seven years, and it was when AI was manifesting itself as a possible tool of
warfare. And we introduced guidance for
warfare. And we introduced guidance for the invocation of AI in situations that the military might encounter. One of
which was if AI decides that it can or should take action that will end in death of the enemy, should we give it that access to do so
>> or still a big um debate >> or should we always ensure that there's a human inside that loop?
>> It's a big >> Okay, so we said there's got to if AI cannot make an make its own decision to kill right?
>> A human has to be in there. My question
to you is Jeffrey, if there are other nations who put in no such safeguards, then that is a timing advantage that an enemy would have over you.
>> Correct.
>> And then we have we have we have one more step in the loop that they don't.
>> Absolutely. But I my belief is that the US military isn't committed to the always being a human involved in each decision to kill. They what they say is there will always be human oversight,
>> right? But in the heat of battle, you've
>> right? But in the heat of battle, you've got a drone that's going up against a Russian tank, and you don't have time for a human to say, "Is it okay for the
drone to drop a grenade on this soldier?" So, my suspicion is the US
soldier?" So, my suspicion is the US military, if you made the recommendation, there should always be a person.
>> Well, that was like eight years ago.
Yeah.
>> Yeah. I don't think they stand by that anymore. I think what they say is
anymore. I think what they say is there'll always be human oversight, which is a much vagger thing.
>> All right. So,
>> human accountability. On the subject of war, is there likely to be international cooperation on development of guardrails and a human factor in decision-m or is
this just wild west?
>> Okay, if you ask when do people cooperate, people cooperate when their interests are aligned. So at the height of the cold war, the USA and the USSR
cooperated on not having a global thermonuclear war because it wasn't in either of their interests. Their
interests were aligned. So if you look at the risks of AI, there's using AI to corrupt elections with fake videos.
>> The country's interests are anti-aligned. They're all doing it to
anti-aligned. They're all doing it to each other, >> right?
>> There's cyber attacks. Their interests
are basically anti-aligned. There's
terrorist creating viruses where their interests are probably aligned. So they
might cooperate there. And then there's one thing where their interests are definitely aligned and they will cooperate which is preventing AI from taking over from people. If the Chinese
figured out how you could prevent AI from ever wanting to take over, from ever wanting to take control away from people, they would immediately tell the Americans because they don't want AI taking control away from people in
America either. We're all in the same
America either. We're all in the same boat when it comes to that.
>> This is the AI version of uh nuclear winter.
>> Yes, >> it seems to me >> it is. It's exactly that. will cooperate
to try and avoid that.
>> Because in nuclear winter, just to refresh people's memory, the idea was if there's total nuclear exchange, you incinerate forests and land and what have you. The soot gets into the
have you. The soot gets into the atmosphere, block sunlight, and all life dies.
>> So there is no winner, >> of course, >> in a total exchange of nuclear weapons.
>> Mutually assured destruction.
>> Yeah. And so who wants that?
>> Unless unless you're a madman or something, they exist. Maybe I think maybe the cockroaches win.
>> They win.
>> Oh, yeah. Well, how about that?
>> Yeah. This doesn't factor in a possible leader who is in a death cult.
>> A Nero, so to speak.
>> Yeah. If I moder if I say I don't mind if everybody dies cuz I'm going to this place when in in death and all my followers are coming with me in this cult. So that that complicates this
cult. So that that complicates this aligned vision statement that you're describing.
It does complicate it a lot. And I find it very comforting that um it's obvious that Trump doesn't actually believe in God.
>> Oh, let me follow that up with a quote from Steven Weinberg.
>> Okay.
>> Do you know this quote, Jeffrey?
>> No.
>> Steven Weinberg. There will always be good people and bad people in the world.
But to get a good person to do something bad requires religion.
>> That's that's >> because they're doing it in the name of religion. You did do it in the name of
religion. You did do it in the name of some point of anything.
>> I think we need to we need to recognize at this point that we have a religion.
We call it science. Now it does differ from the other religions. And the way it differs is it's right.
>> Mic drop. Okay. Um
>> wait a minute. I think we got to give Jeffrey Hinton like the Turring Prize and I give Would you give him a Nobel Prize for what he's contributed here?
>> Well, to go with his other one.
>> Yes.
>> No. No. I I I like earrings.
>> I left that out at the beginning, sir.
In 2018, you won the Turing prize. This
is a highly coveted computer science prize. Uh, correct. And and and Turing,
prize. Uh, correct. And and and Turing, we mentioned him at the beginning of the top of the show. So, first
congratulations on that. And then that wasn't enough.
>> Okay. Uh, the Nobel Committee >> sluming with the Nobel.
>> Yeah. So the Nobel committee said this AI stuff that was birthed by by Jeffrey's work from decades ago is so fundamental to what's going on in this world. We got to give this man Nobel
world. We got to give this man Nobel Prize and earn the Nobel Prize in physics 2024.
>> Just a little correction, there are a whole bunch of people birthed AI. Um in
particular, the back propagation algorithm was reinvented by David Rumlhart who got a nasty brain disease and died young >> but he doesn't get enough credit. Oh,
okay. Thanks for calling that out. Plus,
the Nobel Committee does not offer a Nobel Prize >> to you if you're already dead.
>> So, there's no >> You have to be alive when they announce it.
>> Award. No. Well, you can get it if you died between when they announced it and the ceremony, but not if So, anyway, so congratulations on that. And I don't
mean to brag on our podcast, but you're like the fifth Nobel laurate we've interviewed.
>> More than that.
>> Yeah. Yeah. I think we Yeah. I don't
mean to brag on our podcast. Yeah,
that's all.
>> That's cool, though.
>> That's cool. Go. Okay.
>> I have a a follow-up question. I mean,
we've we've got into the apocalyptic scenario and at the moment, hopefully, it's a scenario that doesn't play out because we are competitive by nature as
humans and particularly here in the US, who is leading the race in artificial intelligence and who is likely to cross the finish line first when it comes to
the prize? If I had to bet on one lot of
the prize? If I had to bet on one lot of people, >> Mhm.
>> it would probably be Germany, Google.
But I used to work for Google, so don't take me too seriously about that. I have
a vested interest in them winning. Um,
Anthropic might win, OpenAI might win. I
think it's less likely that Microsoft will win or that Facebook will win.
>> Well, we know it won't be Facebook.
Why do you know that?
>> I mean, let's look at who's running Facebook. Okay, come on.
Facebook. Okay, come on.
>> No, it's not who's running it. it who
has the resources to get the right people to do the work.
>> All right, Jeffrey, the follow up on that is whoever crosses the line first, what is their prize? What will be the reward for them getting there before?
>> Wait, back up for a sec. Tell me about the value of the stock market in the last year.
Okay. And my belief is just from reading it in the media that 80% of the increase of the value in the stock market, the US stock market can be attributed to the increase in value of the big AI
companies.
>> True.
>> 80% of the growth.
>> Yes.
>> Anyone thinking bubble? And that's kind of what they're calling it, the AI bubble.
>> Okay.
>> The issue is this. There's two senses of bubble. One sense of bubble is it turns
bubble. One sense of bubble is it turns out AI doesn't really work as well as people thought it might.
>> Right?
>> It doesn't actually develop the ability to replace all human intellectual labor which is what most people developing it believe is going to happen in the end.
>> That was the fear factor for sure.
>> Yeah.
>> The other sense of bubble is the companies can't get their money back from the investments. Now that seems to be more likely kind of bubble >> because as far as I understand it, the
companies are all assuming if we can get there first, we can sell people AI that will replace a lot of jobs. And of
course, people will pay a lot of money for that. So, we'll get lots of money.
for that. So, we'll get lots of money.
But they haven't thought about the social consequences. If they really do
social consequences. If they really do replace lots of jobs, the social consequences will be terrible.
>> Correct.
>> Totally. However, it'll be it'll be >> they replace the jobs and now you still want to sell your product and no one has income to buy the product.
>> Yeah. It's it's a self-limiting path.
>> That's the Keynesian view of it. And
then the additional view is that there'll be high unemployment levels which will lead to a lot of social unrest. So the uh yeah the secondary uh
unrest. So the uh yeah the secondary uh view of that is you just have two tiers of existence for our societies and the first tier is all the people who are
benefiting from AI and the second tier are the you know the the feudal peasants that are now forced to live their lives because of AI.
>> Let me ask you a non-AI question because just you're a deep thinker in this space. That's what everybody said in the
space. That's what everybody said in the dawn of automation. Everyone will be unemployed. there'll be no jobs left and
unemployed. there'll be no jobs left and society will go to ruin. Yet society
expanded with other needs and other things people that's why 90% of us are no longer farmers. Okay, we we we've have machines to do that and we invent other things like vacation resource
>> but that decades this is going to take a fraction.
>> Is that so Jeffrey is the problem here the rapidity with which we may create an unemployment an unemployed class where the society cannot recover from the rate
at which people are losing their jobs.
That certainly is one big aspect of the problem. But there's another aspect
problem. But there's another aspect which is if you use a tractor to replace physical labor, you need far fewer people now. Other people can go off and
people now. Other people can go off and do intellectual things. But if you replace human intelligence, where are they going to go? Where are
people who work in a call center going to go when an AI can do their job cheaper and better?
>> Right. Yeah. This is
>> Oh, so there's not another thing.
there's not another thing.
>> They open another thing and then AI will do that.
>> Right?
>> Whatever thing you open, AI can do.
>> You can look at human history in an interesting way as getting rid of limitations.
>> So a long time ago, we had the limitation you had to worry about where your next meal was coming from, >> right?
>> Agriculture got rid of that. It
introduced a lot of other problems, but it got rid of that particular worry.
Then we had the limitation you couldn't travel very far. Well, the bicycle helped a lot with that and cars and airplanes. We got over that kind of
airplanes. We got over that kind of limitation. For a long time, we had the
limitation. For a long time, we had the limitation. We were the ones who had to
limitation. We were the ones who had to do the thinking. We're just about to get over that limitation.
And it's not clear what happens once you got over all the limitations. People
like Sam Elman think it'll be wonderful, >> right? So, we we'll become AI's pet.
>> right? So, we we'll become AI's pet.
>> Well, no. A lot of people believe that this is the um and this this movement started years ago for universal global income.
>> Okay. So would you say Jeffrey that the the universal basic income the stock value the figurative stock value in that idea is growing as AI gains power.
>> It's becoming to seem more essential but it has lots of problems. So one problem is many people get their sense of selfworth from the job they do and it won't deal with the dignity issue.
Another problem is the tax base. If you
replace workers with AIs, the government loses its tax base. It has to somehow be able to tax the AIs. But the big companies aren't going to like that.
>> I think we should let AI figure out this problem.
>> That's right.
So Jeffrey the many people uh especially sci-fi writers distinguish between the power and intellect of machines fine and
the crossover when they become conscious and that's was a big moment in the Terminator series >> that was the singularity in the terminator >> when Skynet Skynet
>> had enough neural connections or whatever kind of connections made it so that it achieved consciousness. So there
seems to be and if you come to this as a as a cognitive psychologist, I'm curious how you think about this. Are we allowed to presume that given sufficient
complexity in any neural net be it real or imag or or artificial something such as consciousness emerges.
>> So the problem here is not really a scientific problem. It's that most
scientific problem. It's that most people in our culture have a theory of how the mind works and they have a view of consciousness as some kind of essence
that emerges. I think consciousness is
that emerges. I think consciousness is like flegiston maybe. Um it's an essence that's designed to explain things and once we understand those things we won't
be trying to use that essence to explain them. I want to try and convince you
them. I want to try and convince you that a multimodal chatbot already has subjective experience. So people use the
subjective experience. So people use the word sentience or consciousness or subjective experience. Let's focus on
subjective experience. Let's focus on subjective experience for now. Most
people in our culture think that the way the mind works is it's a kind of internal theater. And when you're doing
internal theater. And when you're doing perception, the world shows up in this internal theater and only you can see what's there. So if I say to you, if I
what's there. So if I say to you, if I drink a lot and I say to you, I have the subjective experience of little pink elephants floating in front of me. Most
people interpret that as there's this inner theater, my mind and I can see what's in it and what's in it is little pink elephants and they're not made of real pink and real elephants. So they
must be made of something else. So
philosophers invent qualia which is kind of the flegiston of cognitive science.
They say they must be made of qualia.
Let me give you a completely different view that is Daniel Dennett's view who was a great philosopher of cognitive science which is >> late great philosopher. Yeah,
>> the late great that view of the mind is just utterly wrong. So I'm now going to say the same thing as when I told you I had the subjective experience of Olympic elephants without using the word
subjective experience and without appealing to Qualia. I start off by saying I believe my perceptual systems lying to me. That's the subjective bit
of it. But if my perceptual system
of it. But if my perceptual system wasn't lying to me, there would be little pink elephants out there in the world floating in front of me. So what's
funny about these little pink elephants is not that they're made of qualia and they're in an inner theta. It's that
they're hypothetical. They're a
technique for me telling you how my perceptual systems lying by telling you what would have to be there for my perceptual system to be telling the truth. And now I'm going to do it with a
truth. And now I'm going to do it with a chatbot. I take a multimodal chatbot. I
chatbot. I take a multimodal chatbot. I
train it up. It's got a camera. It's got
a robot arm. It can talk. I put an object in front of it and I say, "Point at the object and it points at the object." Then I mess up its perceptual
object." Then I mess up its perceptual system. I put a prism in front of the
system. I put a prism in front of the camera. And now I put an object in front
camera. And now I put an object in front of it and say, "Point at the object."
And it points off to one side. And I say to it, "No, that's not where the object is. It's actually straight in front of
is. It's actually straight in front of you." But I put a prism in front of your
you." But I put a prism in front of your lens. And the chatbot says, "Oh, I see.
lens. And the chatbot says, "Oh, I see.
The prism bent the light rays, so the object is actually straight in front of me." But I had the subjective experience
me." But I had the subjective experience that it was off to one side. Now, if the chatbot said that, it would be using words subjective experience exactly the way we use them. And so that chatbot
would have just had a subjective experience.
>> Now, what if you um first went out drinking with the chatbot and you had a very significant amount of Johnny Walker Blue?
>> That's extremely improbable. I would
have Leafrog.
>> Oh. Oh.
>> Oh, you're I see you're an eye man. You
like the piness of the leaf. Okay, good
man.
>> Oh, so if I understand what you just shared with us in these two examples, >> you actually pulled a consciousness touring test on us. You said a human
would do this and now your chatbot does it and it's fundamentally the same. So
if you want to say we're conscious for exhibiting that behavior, you're going to have to say the chatbot's conscious and inventing whatever mysterious fluid is making that happen. But it could be
that we are the whole concept of consciousness is a distraction from just the actions that people take in the face of stimulus.
>> Okay. So notice that the chatbot doesn't have any mysterious essence or fluid called consciousness, but it has a subjective experience just like we do.
So I think this whole idea of consciousness is some magic essence that you suddenly get indicted with if you're complicated enough is just nonsense.
>> Yeah, there you go.
>> I agree. I've always felt that consciousness was something people are trying to explain without knowing if it really exists >> in in any kind of tangible way, >> which is why it's always difficult to describe because you don't know what it is
>> for example. Yes. Yes. But I think there is awareness. And if you look at what
is awareness. And if you look at what scientists say when they're not thinking philosophically, there's a lovely paper where the chatbot says, "Now, let's be honest with each
other. Are you actually testing me?" And
other. Are you actually testing me?" And
the scientists say, "The chatbot was aware it was being tested." So, they're attributing awareness to a chatbot. And
in everyday conversation, you call that consciousness. It's only when you start
consciousness. It's only when you start thinking philosophically and thinking that it's some funny mysterious essence that you get all confused.
>> Well, there is >> I have to say that this has been a fascinating conversation that will cause me not to sleep for a month.
>> Um, yeah, >> you get plenty of work done.
>> So, Jeffrey, take us out on a positive note, please. So, we still have time to
note, please. So, we still have time to figure out if there's a way we can coexist happily with AI and we should be putting a lot of research effort into that because if we can coexist happily
with it and we can solve all the social problems that will arise when it makes all our jobs much easier then it can be a wonderful thing for people.
>> Agreed. Okay. So, so there is hope.
>> Yes. And one last thing because you hinted at it, this point of singularity where AI trains on itself
so that it exponentially gets smarter like by the minute. That's been called a singularity by many people. Of course,
Ray Kershw among them who's been a guest on a previous episode of Stars.
>> Yeah. A couple of times. Yeah. So, what
is your sense of this singularity? Is it
real the way others say? Is it imminent the way others say?
I don't know the answer to either of those questions. My suspicion is AI will
those questions. My suspicion is AI will get better at us in the end at everything, better than us at everything, but it'll be sort of one thing at a time. It's currently much better than us at chess and go. It's
much better than us at knowing a lot of things. Not quite as good as us at
things. Not quite as good as us at reasoning. I think rather than sort of
reasoning. I think rather than sort of massively overtaking us in everything all at once, it'll be done one area at a time. And my sort of way out of that is,
time. And my sort of way out of that is, you know, I get to walk a beach and look at pebbles and seashells. AI doesn't.
>> Yeah. It can create its own beach.
>> No. Would it only know about the new mollisk that I discovered if I write it up and put it online?
>> Mhm.
>> So, the human can continue to explore the universe in ways that AI doesn't have access to.
>> There's one word missing from your entire assessment.
>> What's that?
>> Yet.
Yeah, I just think of my, you know, will AI come up with a new theory of the universe that requires human insights that it doesn't have because I'm thinking the way no one has thought before.
>> I think it will.
>> That's not the answer I wanted from you.
>> Yeah, I was.
>> But that's the answer you got.
>> Let me give you an example. AI is very good at analogies already. So when chat GPD4 was not allowed to look on the web when all its knowledge was in its weights, I asked it why is a compost
heap like an atom bomb and it knew it said the energy scales are very different and the time scales are very different. But it then went on to talk
different. But it then went on to talk about how when a compost heap gets hotter it generates heat faster and when an atom bomb generates more neutrons it generates neutrons faster. Um, so it
understood the commonality and it had to understand that to pack all that knowledge into so few connections, only a trillion or so. That's a source of much creativity >> and it's not just by finding words that
were juxtaposed with other words.
>> No, it understood what a chain reaction was.
>> Yeah.
>> Well, all right. That's the end of us.
>> Yeah. We're done
>> on Earth. We're done.
>> We're finished.
>> This is the last episode. We
>> stick in us. We're done. Gentlemen, it's
been a pleasure.
>> Well, Jeffrey Hinton, it's been a delight to have you on.
>> We know you're you're tugged in many directions, especially after your recent Nobel Prize, and we're delighted you gave us a piece of your surely overscheduled and busy life.
>> Thank you for inviting me.
>> Well, guys, that was something.
>> Did you sit comfortably through all of that?
>> I was I I I squirmed. I squirmed.
>> I knew you'd panic. Well, no. I have to tell you that um certain parts of the um conversation gave me the anxiety of, you know, sitting in a theater theater with diarrhea.
>> Thanks for that explicit.
>> Thanks for sharing. That That's the nicest thing anybody's ever said about me.
>> On that note, this has been Star Talk special edition. Chuck, always good to
special edition. Chuck, always good to have you. Gary, love having you right at
have you. Gary, love having you right at my side. Neil deGrasse Tyson bidding you
my side. Neil deGrasse Tyson bidding you as always to keep looking up however much harder that will become.
Loading video analysis...