back to indexMax Tegmark: The Case for Halting AI Development | Lex Fridman Podcast #371
Chapters
0:0 Introduction
1:56 Intelligent alien civilizations
14:20 Life 3.0 and superintelligent AI
25:47 Open letter to pause Giant AI Experiments
50:54 Maintaining control
79:44 Regulation
90:34 Job automation
99:48 Elon Musk
121:31 Open source
128:1 How AI may kill all humans
138:32 Consciousness
147:54 Nuclear winter
158:21 Questions for AGI
00:00:01.840 |
that there will come a time when we want to pause a little bit. 00:00:07.200 |
The following is a conversation with Max Tegmark, 00:00:15.440 |
In fact, his first appearance was episode number one 00:00:20.560 |
He is a physicist and artificial intelligence researcher 00:00:24.000 |
at MIT, co-founder of Future of Life Institute, 00:00:27.120 |
and author of Life 3.0, Being Human in the Age 00:00:33.480 |
Most recently, he's a key figure in spearheading 00:00:36.320 |
the open letter calling for a six-month pause 00:00:39.120 |
on giant AI experiments, like training GPT-4. 00:00:47.440 |
on training of models larger than GPT-4 for six months. 00:00:51.880 |
This does not imply a pause or ban on all AI research 00:00:55.000 |
and development or the use of systems that have already 00:01:02.440 |
a very small pool of actors who possesses this capability." 00:01:06.920 |
The letter has been signed by over 50,000 individuals, 00:01:09.960 |
including 1,800 CEOs and over 1,500 professors. 00:01:14.480 |
Signatories include Yoshua Bengio, Stuart Russell, 00:01:17.840 |
Elon Musk, Steve Wozniak, Yuval Noah Harari, Andrew Yang, 00:01:26.040 |
of human civilization, where the balance of power 00:01:32.880 |
And Max's mind and his voice is one of the most valuable 00:01:39.520 |
His support, his wisdom, his friendship has been a gift 00:01:55.600 |
You were the first ever guest on this podcast, 00:02:11.200 |
and just acting like I'm somebody who matters, 00:02:14.280 |
that I'm somebody who's interesting to talk to. 00:02:20.200 |
- Thanks to you for putting your heart and soul into this. 00:02:24.280 |
I know when you delve into controversial topics, 00:02:26.880 |
it's inevitable to get hit by what Hamlet talks about, 00:02:33.800 |
It's in an era where YouTube videos are too long 00:02:37.400 |
and now it has to be like a 20-minute TikTok, 00:02:41.880 |
It's just so refreshing to see you going exactly 00:02:44.280 |
against all of the advice and doing these really long form 00:02:57.280 |
the first question I've ever asked on this podcast, 00:03:02.200 |
do you think there's intelligent life out there 00:03:08.800 |
What's your view when you look out to the stars? 00:03:14.120 |
if you define our universe the way most astrophysicists do, 00:03:18.920 |
not as all of space, but the spherical region of space 00:03:34.360 |
in this spherical volume that has invented internet, 00:03:49.960 |
Because if it's true, it means that life is quite rare. 00:03:58.160 |
of advanced consciousness, which if we nurture it 00:04:01.320 |
and help it grow, eventually life can spread from here 00:04:11.360 |
with the technology we build and just snuff it out 00:04:17.360 |
then maybe the rest of cosmic history in our universe 00:04:24.800 |
But I do think that we are actually very likely 00:04:28.960 |
to get visited by aliens, alien intelligence quite soon. 00:04:47.200 |
that evolution here on Earth was able to create 00:04:49.620 |
in terms of the path, the biological path it took. 00:05:07.620 |
where it necessarily cares about self-preservation, 00:05:18.600 |
is just so much vaster than what evolution will give you. 00:05:22.100 |
And with that also comes a great responsibility 00:05:24.880 |
for us to make sure that the kind of minds we create 00:05:28.240 |
are the kind of minds that it's good to create. 00:05:49.800 |
Do you try to consider all the different kinds 00:05:55.480 |
what humans are able to do to the full spectrum 00:05:58.600 |
of what intelligent creatures, entities could do? 00:06:08.800 |
to really grapple with something so completely alien. 00:06:18.680 |
if we were completely indifferent towards death 00:06:29.320 |
you could just copy my knowledge of how to speak Swedish. 00:06:37.280 |
And you could copy any of my cool experiences 00:06:39.680 |
and then you could delete the ones you didn't like 00:06:48.520 |
You probably spend less effort studying things 00:06:54.660 |
because if the plane you're on starts to crash, 00:06:58.300 |
you'd just be like, "Oh, shucks, I haven't backed 00:07:04.640 |
"So I'm gonna lose all these wonderful experiences 00:07:09.700 |
We might also start feeling more compassionate, 00:07:16.140 |
maybe with other people, if we can so readily share 00:07:24.760 |
I really feel very humble about this, to grapple with it, 00:07:35.280 |
which I think is just really worth reflecting on 00:07:38.360 |
is because the mind space of possible intelligences 00:07:42.400 |
is so different from ours, it's very dangerous 00:07:48.420 |
- Well, the entirety of human written history 00:08:07.680 |
and all of that changes if you have a different 00:08:13.280 |
The entirety, all those poems, they're trying to sneak up 00:08:19.880 |
How AI concerns and existential crises that AI experiences, 00:08:24.880 |
how that clashes with the human existential crisis, 00:08:29.800 |
the human condition, that's hard to fathom, hard to predict. 00:08:34.480 |
- It's hard, but it's fascinating to think about also. 00:08:37.960 |
Even in the best case scenario where we don't lose control 00:08:42.180 |
over the ever more powerful AI that we're building 00:08:44.960 |
to other humans whose goals we think are horrible, 00:08:49.120 |
and where we don't lose control to the machines, 00:08:56.320 |
even then, you get into the questions you touched here. 00:08:59.660 |
Maybe it's the struggle that it's actually hard 00:09:14.240 |
that they put together has this woman talking about, 00:09:28.880 |
if I realized that my parents couldn't be bothered 00:09:46.360 |
do you think that would also take away a little bit of what-- 00:09:57.860 |
I had somebody mention to me that they started using 00:10:19.760 |
to get the point across, but rewrite it in a nicer way. 00:10:35.780 |
And it's scary because so much of our society 00:10:44.640 |
And if we're now using AI as the medium of communication 00:10:51.140 |
so much of the emotion that's laden in human communication, 00:10:55.520 |
so much of the intent that's going to be handled 00:10:59.320 |
by outsourced AI, how does that change everything? 00:11:12.180 |
- Yeah, for me personally, I have to confess, 00:11:26.120 |
I don't want to just press a button and be at the top. 00:11:32.360 |
In the same way, I want to constantly work on myself 00:11:55.840 |
- Yeah, but then again, it could be like with chess. 00:12:08.800 |
and provide maybe a flourishing civilization for humans, 00:12:15.120 |
and playing our games even though AI is so much smarter, 00:12:18.240 |
so much stronger, so much superior in every single way, 00:12:34.240 |
a medium that enables the human experience to flourish. 00:12:45.600 |
- Yeah, I would phrase that as rebranding ourselves 00:12:53.920 |
Right now, sapiens, the ability to be intelligent, 00:13:08.600 |
That's clearly gonna change if AI continues ahead. 00:13:14.200 |
So maybe we should focus on the experience instead, 00:13:16.520 |
the subjective experience that we have with homo sentiens 00:13:25.400 |
Get off our high horses and get rid of this hubris 00:13:35.160 |
- So consciousness, the subjective experience 00:13:37.880 |
is a fundamental value to what it means to be human. 00:13:50.920 |
not just towards other humans because they happen 00:13:55.720 |
but also towards all our other fellow creatures 00:14:10.040 |
in the grand scheme of things either in the post-AI epoch, 00:14:13.040 |
then surely we should value the subjective experience 00:14:19.520 |
- Well, allow me to briefly look at the book, 00:14:23.960 |
which at this point is becoming more and more visionary 00:14:26.340 |
that you've written, I guess over five years ago, 00:14:29.880 |
So first of all, 3.0, what's 1.0, what's 2.0, what's 3.0? 00:14:45.160 |
in that it can't actually learn anything at all 00:14:47.880 |
The learning just comes from this genetic process 00:14:55.200 |
Life 2.0 is us and other animals which have brains, 00:15:00.200 |
which can learn during their lifetime a great deal. 00:15:06.960 |
And you were born without being able to speak English. 00:15:23.840 |
can replace not only its software the way we can, 00:15:28.500 |
And that's where we're heading towards at high speed. 00:15:42.520 |
And if Neuralink and other companies succeed, 00:16:22.480 |
It seems like maybe the thing that's truly powerful 00:16:32.840 |
- I think we should be humble and not be so quick 00:16:37.960 |
to make everything binary and say either it's there 00:16:44.960 |
And there is even controversy about whether some unicellular 00:16:48.960 |
organisms like amoebas can maybe learn a little bit 00:16:53.440 |
So apologies if I offended any bacteria here. 00:16:57.040 |
It was more that I wanted to talk up how cool it is 00:17:01.420 |
where you can learn dramatically within your lifetime. 00:17:05.800 |
- And the higher up you get from 1.0 to 2.0 to 3.0, 00:17:09.240 |
the more you become the captain of your own ship, 00:17:20.240 |
we can be so different from previous generations 00:17:45.520 |
- Yeah, and I think it's worth commenting a bit 00:17:59.380 |
understanding is that life is best thought of 00:18:04.380 |
not as a bag of meat or even a bag of elementary particles, 00:18:10.860 |
but rather as a system which can process information 00:18:19.580 |
even though nature is always trying to mess it up. 00:18:40.540 |
you're not the same atoms as during the first 00:18:43.520 |
time you did with me. - Time we talked, yeah. 00:18:44.600 |
- You've swapped out most of them, but still you. 00:18:55.840 |
and whatever, you can still have this kind of continuity. 00:19:06.880 |
I lost both of my parents since our last podcast. 00:19:17.060 |
they haven't entirely died because a lot of mommy 00:19:21.400 |
and daddies, sorry, I'm getting a little emotional here, 00:19:28.840 |
and even jokes and so on, they haven't gone away, right? 00:19:49.160 |
And particularly if you can share your own information, 00:20:02.320 |
to immortality we can get with our bio-bodies. 00:20:06.840 |
- You carry a little bit of them in you in some sense. 00:20:34.960 |
But I think my obsession for fairly big questions 00:20:47.120 |
which is a very core part of really who I am, 00:21:11.280 |
They both very much just did their own thing. 00:21:22.320 |
- That's why you've always been an inspiration to me, 00:21:35.160 |
You're one of the people that represents MIT best to me. 00:21:41.960 |
So it's good to hear that you got that from your mom and dad. 00:21:44.960 |
But yeah, I mean, the good reason to do science 00:21:57.200 |
and everyone else says, no, no, that's bullshit, 00:22:04.240 |
And even if everybody else keeps thinking it's bullshit, 00:22:10.360 |
I always root for the underdog when I watch movies. 00:22:22.160 |
talking about our universe ultimately being mathematical, 00:22:25.680 |
I got this email from a quite famous professor saying, 00:22:42.720 |
Follow your own path and let the people talk. 00:22:49.080 |
you know, he's dead, but that attitude is not. 00:22:53.000 |
- How did losing them as a man, as a human being change you? 00:22:59.240 |
How did it expand your thinking about the world? 00:23:19.400 |
One of them, just going through all their stuff 00:23:25.840 |
just drove home to me how important it is to ask ourselves, 00:23:31.520 |
Because it's inevitable that you look at some things 00:23:40.400 |
Would they have done something that was more meaningful? 00:23:42.920 |
So I've been looking more in my life now and asking, 00:23:50.120 |
it should either be something I really enjoy doing 00:23:58.840 |
really, really meaningful because it helps humanity. 00:24:09.480 |
maybe I should spend less time on it, you know? 00:24:12.480 |
The other thing is dealing with death up in person like this, 00:24:24.560 |
that I'm an idiot, you know, which happens regularly. 00:24:27.960 |
And just, I'm gonna live my life, do my thing, you know? 00:24:31.280 |
And it's made it a little bit easier for me to focus 00:24:51.400 |
And I'm next in line in our family now, right? 00:25:14.760 |
And they focused on that rather than wasting time 00:25:26.360 |
if they start their morning by making a list of grievances. 00:25:30.800 |
Whereas if you start your day with a little meditation 00:25:39.840 |
- Because you only have a finite number of days 00:25:46.600 |
- Well, you do happen to be working on a thing 00:25:52.840 |
which seems to have potentially some of the greatest impact 00:25:57.840 |
on human civilization of anything humans have ever created, 00:26:05.240 |
and in the high philosophical level you work on. 00:26:08.280 |
So you've mentioned to me that there's an open letter 00:26:20.040 |
I've been having late nights and early mornings. 00:26:24.840 |
In short, have you seen "Don't Look Up", the film? 00:26:47.880 |
except it's an asteroid that we are building ourselves. 00:26:56.680 |
about all sorts of things which seem very minor 00:26:58.840 |
compared to the asteroid that's about to hit us, right? 00:27:02.320 |
Most politicians don't even have their radar, 00:27:11.400 |
This is the most important fork humanity has reached 00:27:27.560 |
but that's a technicality which will soon be changed. 00:27:32.080 |
And this arrival of artificial general intelligence 00:27:39.280 |
and probably shortly thereafter, superintelligence, 00:27:43.120 |
which greatly exceeds our cognitive abilities, 00:27:46.360 |
it's gonna either be the best thing ever to happen 00:27:55.160 |
- But it would be fundamentally transformative 00:28:18.360 |
on the planet for very long unless AI progress just halts. 00:28:22.720 |
And we can talk more about why I think that's true 00:28:29.920 |
reasons you might think it's gonna be the best thing ever 00:28:35.160 |
and the reason you think it's gonna be the end of humanity, 00:28:41.480 |
But what I think we can, anyone who's working on advanced AI 00:28:46.480 |
can agree on is it's much like the film "Don't Look Up" 00:28:52.720 |
in that it's just really comical how little serious 00:28:57.640 |
public debate there is about it given how huge it is. 00:29:01.440 |
- So what we're talking about is a development 00:29:09.200 |
and the signs it's showing of rapid improvement 00:29:14.840 |
that may in the near term lead to development 00:29:18.440 |
of super intelligent AGI, AI, general AI systems, 00:29:30.680 |
and then beyond that, general super human level intelligence. 00:30:09.720 |
Before then, a lot of people thought it was just really 00:30:11.640 |
kooky to even talk about it and a lot of AI researchers 00:30:32.160 |
You go in any AI conference and people talk about AI safety 00:30:34.840 |
and it's a nerdy technical field full of equations 00:30:42.600 |
But there's this other thing which has been quite taboo 00:30:54.440 |
including myself, I've been biting my tongue a lot, 00:30:56.520 |
is that we don't need to slow down AI development, 00:31:01.520 |
we just need to win this race, the wisdom race, 00:31:07.840 |
and the growing wisdom with which we manage it. 00:31:18.720 |
how you can actually ensure that your powerful AI 00:31:23.320 |
and have society adapt also with incentives and regulations 00:31:34.840 |
The progress on technical AI and capabilities 00:31:39.960 |
has gone a lot faster than many people thought 00:31:58.360 |
than we hoped with getting policy makers and others 00:32:10.600 |
Maybe we should unpack it and talk a little bit about each. 00:32:12.400 |
So why did it go faster than a lot of people thought? 00:32:16.920 |
In hindsight, it's exactly like building flying machines. 00:32:28.800 |
Have you seen the TED Talk with a flying bird? 00:32:35.240 |
But it took 100 years longer to figure out how to do that 00:32:38.440 |
than for the Wright brothers to build the first airplane 00:32:40.440 |
because it turned out there was a much easier way to fly. 00:32:43.080 |
And evolution picked the more complicated one 00:32:48.000 |
It could only build a machine that could assemble itself, 00:32:55.040 |
that used only the most common atoms in the periodic table. 00:33:05.920 |
and it also had to be incredibly fuel efficient. 00:33:13.360 |
of a remote control plane flying the same distance. 00:33:31.800 |
we had to figure out how the brain does human level AI first 00:33:39.000 |
You can take an incredibly simple computational system 00:33:45.760 |
and just train it to do something incredibly dumb. 00:33:52.400 |
And it turns out if you just throw a ton of compute at that 00:33:57.240 |
and a ton of data, it gets to be frighteningly good 00:34:01.520 |
like GPT-4, which I've been playing with so much 00:34:13.680 |
But yeah, we can come back to the details of that 00:34:22.200 |
- Can you briefly, if it's just a small tangent, 00:34:27.120 |
Suggest that you're impressed by this rate of progress, 00:34:37.800 |
What are human interpretable words you can assign 00:34:45.040 |
- I'm both very excited about it and terrified. 00:34:52.040 |
- All the best things in life include those two somehow. 00:34:59.580 |
I highly recommend doing that before dissing it. 00:35:08.760 |
which I realized I couldn't do that myself that well even. 00:35:12.200 |
And obviously does it dramatically faster than we do too 00:35:17.920 |
And it's doing that while servicing a massive number 00:35:23.080 |
At the same time, it cannot reason as well as a human can 00:35:39.560 |
Information can go from this neuron to this neuron 00:35:43.240 |
You can like ruminate on something for a while. 00:35:52.400 |
It's a so-called transformer where it's just like 00:35:57.880 |
In geek speak, it's called a feed-forward neural network. 00:36:03.080 |
So it can only do logic that's that many steps 00:36:06.680 |
And you can create the problems which will fail to solve 00:36:13.760 |
But the fact that it can do so amazing things 00:36:20.280 |
with this incredibly simple architecture already 00:36:35.520 |
It's called mechanistic interpretability in geek speak. 00:36:40.520 |
You have this machine that does something smart. 00:36:45.480 |
Or you think of it also as artificial neuroscience. 00:36:50.680 |
'Cause that's exactly what neuroscientists do 00:36:52.360 |
But here you have the advantage that you can, 00:36:54.360 |
you don't have to worry about measurement errors. 00:36:56.200 |
You can see what every neuron is doing all the time. 00:36:59.560 |
And a recurrent thing we see again and again, 00:37:02.360 |
there's been a number of beautiful papers quite recently 00:37:09.120 |
I mean in this area, is where when they figure out 00:37:14.280 |
"Oh man, that's such a dumb way of doing it." 00:37:16.840 |
And you immediately see how it can be improved. 00:37:18.840 |
Like for example, there was a beautiful paper recently 00:37:22.200 |
where they figured out how a large language model 00:37:24.640 |
stores certain facts, like Eiffel Tower is in Paris. 00:37:28.800 |
And they figured out exactly how it's stored. 00:37:31.680 |
The proof that they understood it was they could edit it. 00:37:37.800 |
And then they asked it, "Where's the Eiffel Tower?" 00:37:41.440 |
And then they asked you, "How do you get there?" 00:37:47.000 |
and the Roma Termini train station, and this and that." 00:37:51.120 |
And what might you see if you're in front of it? 00:37:59.160 |
- But the way that it's storing this information, 00:38:07.800 |
there was a big matrix, and roughly speaking, 00:38:13.200 |
which encode these things, and they correspond 00:38:17.880 |
And it would be much more efficient for a sparse matrix 00:38:27.800 |
how these things do, or ways where you can see 00:38:31.320 |
And the fact that this particular architecture 00:38:34.240 |
has some roadblocks built into it is in no way 00:38:54.360 |
to build close to human intelligence than we thought, 00:39:07.960 |
about the effectiveness of large language models, 00:39:11.680 |
so Sam Altman I recently had a conversation with, 00:39:14.920 |
and he really showed that the leap from GPT-3 00:39:19.840 |
to GPT-4 has to do with just a bunch of hacks, 00:39:23.200 |
a bunch of little explorations with smart researchers 00:39:30.560 |
It's not some fundamental leap and transformation 00:39:37.720 |
- And more data and compute, but he said the big leaps 00:39:42.840 |
but just learning this new discipline, just like you said. 00:39:46.800 |
So researchers are going to look at these architectures, 00:39:48.920 |
and there might be big leaps where you realize, 00:39:52.480 |
wait, why are we doing this in this dumb way? 00:39:54.440 |
And all of a sudden this model is 10x smarter, 00:40:07.440 |
It's such a new, like we understand so little 00:40:12.540 |
that the linear improvement of compute, or exponential, 00:40:19.320 |
may not be the thing that even leads to the next leap. 00:40:21.520 |
It could be a surprise little hack that improves everything. 00:40:25.800 |
because so much of this is out in the open also. 00:40:33.640 |
and trying to figure out little leaps here and there, 00:40:42.640 |
And this is actually very crucial for the other part of it. 00:40:47.140 |
So again, what this open letter is calling for 00:40:54.040 |
that are more powerful than GPT-4 for six months. 00:40:59.400 |
Just give a chance for the labs to coordinate a bit 00:41:09.840 |
'Cause you've interviewed a lot of these people 00:41:14.200 |
who lead these labs, and you know just as well as I do 00:41:16.720 |
that they're good people, they're idealistic people. 00:41:21.300 |
because they believe that AI has a huge potential 00:41:33.100 |
Have you read "Meditations on Moloch" by Scott Alexander? 00:41:41.660 |
- Yeah, it's a beautiful essay on this poem by Ginzburg 00:41:44.560 |
where he interprets it as being about this monster. 00:41:47.960 |
It's this game theory monster that pits people 00:41:53.520 |
against each other in this race to the bottom 00:41:59.700 |
even though everybody sees it and understands, 00:42:03.980 |
A good fraction of all the bad things that we humans do 00:42:10.040 |
I like Scott Alexander's naming of the monster 00:42:23.120 |
why do we have more generally the tragedy of the commons, 00:42:29.560 |
I don't know if you've had her on your podcast. 00:42:33.080 |
- Great, she made this awesome point recently 00:42:36.560 |
that beauty filters that a lot of female influencers 00:42:40.740 |
feel pressure to use are exactly Moloch in action again. 00:42:56.660 |
and then the other ones that weren't using it 00:42:58.460 |
started to realize that if they wanna just keep 00:43:01.820 |
their market share, they have to start using it too 00:43:05.860 |
and then you're in a situation where they're all using it 00:43:28.660 |
Moloch is everywhere and Moloch is not a new arrival 00:43:36.220 |
We humans have developed a lot of collaboration mechanisms 00:43:41.540 |
through various kinds of constructive collaboration. 00:43:53.980 |
into unnecessarily risky nuclear arms races, et cetera, 00:43:58.220 |
et cetera and this is exactly what's happening 00:44:08.680 |
If you take any of these leaders of the top tech companies, 00:44:34.160 |
to replace the executives in the worst case, right? 00:44:37.360 |
So we did this open letter because we wanna help 00:44:52.000 |
so that they can all pause in a coordinated fashion. 00:45:11.660 |
for the major developers of AI systems like this, 00:45:21.920 |
- Well, OpenAI is very close with Microsoft now, of course. 00:45:35.040 |
I don't wanna make a long list to leave anyone out. 00:45:44.560 |
that there's external pressure on all of them 00:45:48.760 |
'Cause then the people, the researchers in these organizations 00:45:52.800 |
who the leaders who wanna slow down a little bit, 00:45:56.640 |
everybody's slowing down because of this pressure, 00:46:12.640 |
You could make so much money on human cloning. 00:46:30.960 |
and decided even to stop a lot more stuff also, 00:46:44.800 |
because it's too unpredictable what it's gonna lead to. 00:46:48.120 |
We could lose control over what happens to our species. 00:46:57.800 |
but you need a public awareness of what the risks are 00:47:02.240 |
and the broader community coming in and saying, 00:47:17.320 |
they also get told we can't slow down because the West, 00:47:20.440 |
because both sides think they're the good guy. 00:47:42.160 |
Not because Westerners said, China, look, this is... 00:47:51.760 |
If anything, maybe they are even more concerned 00:47:54.080 |
about having control than Western governments 00:48:03.320 |
that was released by, I believe, Baidu recently. 00:48:07.200 |
They got a lot of pushback from the government 00:48:17.840 |
where everybody loses if anybody's AI goes out of control, 00:48:25.560 |
I'll say this again, 'cause this is a very basic point 00:48:34.200 |
Because a lot of people dismiss the whole idea 00:48:42.080 |
because they think there's something really magical 00:48:43.880 |
about intelligence such that it can only exist 00:48:48.360 |
they think it's gonna kind of get to just more or less 00:49:00.000 |
they're gonna control the world, they're gonna win. 00:49:04.160 |
And we can talk again about the scientific arguments 00:49:09.600 |
But the way it's gonna be is if anybody completely 00:49:24.200 |
you probably don't really even care very much 00:49:27.120 |
You're not gonna like it, much worse than today. 00:49:38.960 |
and we just lose control even to the machines, 00:49:44.360 |
so that it's not us versus them, it's us versus it, 00:49:47.360 |
what do you care who created this underlying entity 00:49:52.320 |
which has goals different from humans ultimately 00:49:55.320 |
and we get marginalized, we get made obsolete, 00:49:59.600 |
That's what I mean when I say it's a suicide race. 00:50:04.960 |
It's kind of like we're rushing towards this cliff, 00:50:10.560 |
the more scenic the views are and the more money we make. 00:50:13.240 |
The more money there is there, so we keep going, 00:50:16.920 |
but we have to also stop at some point, right? 00:50:33.440 |
is to continue developing awesome AI, a little bit slower, 00:50:38.440 |
so we make it safe, make sure it does the things 00:50:41.640 |
we always want and create a condition where everybody wins. 00:50:49.440 |
and politics in general is not a zero-sum game at all. 00:50:54.160 |
- So there is some rate of development that will lead 00:50:57.200 |
us as a human species to lose control of this thing 00:51:01.960 |
and the hope you have is that there's some lower level 00:51:05.200 |
of development which will not allow us to lose control. 00:51:14.360 |
like Sander Parchai or Sam Altman at the head 00:51:17.160 |
of a company like this, you're saying if they develop 00:51:26.720 |
no group of individuals can maintain control. 00:51:29.120 |
- If it's created very, very soon and is a big black box 00:51:33.760 |
that we don't understand, like the large language models, 00:51:36.000 |
yeah, then I'm very confident they're gonna lose control. 00:51:44.000 |
and themselves acknowledged that there's really great risks 00:51:47.880 |
with this and they wanna slow down once they feel 00:51:50.240 |
like it's scary, but it's clear that they're stuck 00:51:53.960 |
and again, Moloch is forcing them to go a little faster 00:51:57.800 |
than they're comfortable with because of pressure 00:52:01.860 |
To get a bit optimistic here, of course this is a problem 00:52:10.320 |
It's just to win this wisdom race, it's clear that what 00:52:14.440 |
we hoped that was gonna happen hasn't happened. 00:52:20.000 |
than a lot of people thought and the progress 00:52:22.800 |
in the public sphere of policymaking and so on 00:52:26.480 |
Even the technical AI safety has gone slower. 00:52:29.060 |
A lot of the technical safety research was kind of banking 00:52:32.060 |
on that large language models and other poorly understood 00:52:35.800 |
systems couldn't get us all the way, that you had to build 00:52:38.560 |
more of a kind of intelligence that you could understand, 00:52:41.240 |
maybe it could prove itself safe, things like this. 00:52:45.600 |
And I'm quite confident that this can be done 00:52:50.360 |
so we can reap all the benefits, but we cannot do it 00:52:53.680 |
as quickly as this out of control express train we are 00:53:00.200 |
That's why we need a little more time, I feel. 00:53:07.720 |
while we're in the pre-AGI stage, to release often 00:53:12.720 |
and as transparently as possible to learn a lot. 00:53:17.400 |
So as opposed to being extremely cautious, release a lot. 00:53:21.360 |
Don't invest in a closed development where you focus 00:53:25.880 |
on AI safety while it's somewhat dumb, quote unquote. 00:53:33.240 |
And as you start to see signs of human level intelligence 00:53:38.240 |
or superhuman level intelligence, then you put a halt on it. 00:53:41.960 |
- Well, what a lot of safety researchers have been saying 00:53:45.480 |
for many years is that the most dangerous things you can do 00:53:48.520 |
with an AI is, first of all, teach it to write code. 00:53:53.060 |
- 'Cause that's the first step towards recursive 00:54:01.560 |
And another thing, high risk is connected to the internet. 00:54:05.840 |
Let it go to websites, download stuff on its own, 00:54:13.600 |
you said you interviewed him recently, right? 00:54:16.000 |
- He had this tweet recently, which gave me one 00:54:19.200 |
of the best laughs in a while, where he was like, 00:54:22.820 |
"you're so stupid, Eliezer, 'cause you're saying 00:54:28.320 |
Obviously, developers, once they get to really strong AI, 00:54:32.760 |
first thing you're gonna do is never connect it 00:55:00.360 |
he has argued for a while that we should never teach AI 00:55:10.400 |
about human psychology and how you manipulate humans. 00:55:13.020 |
That's the most dangerous kind of knowledge you can give it. 00:55:18.320 |
about how to cure cancer and stuff like that, 00:55:25.520 |
And then, "oops, lol, let's invent social media 00:55:30.160 |
recommender algorithms," which do exactly that. 00:55:34.720 |
They get so good at knowing us and pressing our buttons 00:55:45.440 |
'cause they figured out that these algorithms, 00:55:48.760 |
not out of evil, but just to make money on advertising, 00:55:51.920 |
that the best way to get more engagement, the euphemism, 00:56:00.080 |
- Well, that's really interesting that a large AI system 00:56:03.520 |
that's doing the recommender system kind of task 00:56:06.440 |
on social media is basically just studying human beings 00:56:09.880 |
because it's a bunch of us rats giving it signal, 00:56:17.960 |
on whether we spread that thing, we like that thing, 00:56:24.240 |
It has that on the scale of hundreds of millions 00:56:27.840 |
So it's just learning and learning and learning. 00:56:32.080 |
in the neural network that's doing the learning, 00:56:52.440 |
manipulate other humans for profit and power, 00:57:03.840 |
that can make AIs persuade humans to let them escape 00:57:12.480 |
in the New York Times recently by Yuval Noah Harari 00:57:32.780 |
We now live in a country where there's much more hate 00:57:38.200 |
in the world where there's much more hate, in fact. 00:57:41.000 |
And in our democracy, we're having this conversation 00:57:43.960 |
and people can't even agree on who won the last election. 00:57:47.920 |
And we humans often point fingers at other humans 00:57:51.600 |
But it's really Moloch and these AI algorithms. 00:57:59.900 |
pitted the social media companies against each other 00:58:16.920 |
this very medium in which we use as a civilization 00:58:25.760 |
how to solve the biggest problems in the world, 00:58:28.360 |
whether that's nuclear war or the development of AGI? 00:58:35.920 |
- I think it's not only possible, but it's necessary. 00:58:38.920 |
Who are we kidding that we're gonna be able to solve 00:58:40.880 |
all these other challenges if we can't even have 00:58:53.920 |
where you respectfully listen to people you disagree with. 00:58:57.120 |
And you realize, actually, there are some things, 00:59:01.040 |
and we both agree, let's not have a nuclear war, 00:59:07.600 |
We're kidding ourselves thinking we can face off 00:59:12.480 |
the second contact with ever more powerful AI 00:59:16.400 |
that's happening now with these large language models 00:59:19.160 |
if we can't even have a functional conversation 00:59:25.520 |
That's why I started the Improve the News project, 00:59:33.560 |
in that there is a lot of intrinsic goodness in people 00:59:45.200 |
between someone doing good things for humanity 00:59:48.000 |
and bad things is not some sort of fairy tale thing 01:00:05.880 |
And I feel we're building an internet and a society 01:00:22.680 |
that both make money and bring out the best in people. 01:00:32.720 |
if we just ultimately replace all humans by machines 01:00:40.600 |
- Well, it depends, I guess, on how you do the math. 01:00:48.440 |
and there aren't any, that's not a good investment. 01:00:51.000 |
Moreover, why can't we have a little bit of pride 01:01:07.240 |
if we had really advanced biotech to build Homo sapiens? 01:01:20.280 |
Maybe they can help us defend us better against predators 01:01:29.000 |
So then they build a couple, a little baby girl, 01:01:35.960 |
And then you have some wise old Neanderthal elders like, 01:01:39.400 |
"Hmm, I'm scared that we're opening a Pandora's box here 01:01:55.280 |
But then you have a bunch of others in the cave, 01:01:56.640 |
"Well, yeah, you're such a Luddite scaremonger. 01:01:58.920 |
Of course, they're gonna wanna keep us around 01:02:07.400 |
They're gonna want us around and it's gonna be fine. 01:02:11.480 |
And besides, look at these babies, they're so cute. 01:02:16.000 |
That's exactly, those babies are exactly GPT-4. 01:02:19.160 |
It's not, I wanna be clear, it's not GPT-4 that's terrifying. 01:02:27.800 |
You know, and Microsoft even had a paper recently out 01:02:33.040 |
with the title something like "Sparkles of AGI." 01:02:36.720 |
Well, they were basically saying this is baby AI, 01:02:44.600 |
There's gonna be other systems from the same company, 01:02:48.520 |
from other companies, they'll be way more powerful 01:02:58.600 |
those last Neanderthals who were pretty disappointed. 01:03:02.680 |
And when they realized that they were getting replaced. 01:03:07.920 |
which is the programming, it's entirely possible 01:03:13.200 |
that can change everything by writing programs. 01:03:21.880 |
the systems I'm afraid of are gonna look nothing 01:03:25.280 |
like a large language model and they're not gonna. 01:03:38.160 |
And from everything we've seen about how these work 01:03:42.040 |
under the hood, they're like the minimum viable intelligence. 01:03:49.920 |
- So they are Life 3.0, except when they replace 01:03:59.120 |
And moreover, they think a lot faster than us too. 01:04:04.680 |
So when, we don't think on how one logical step 01:04:18.400 |
And we can't also just suddenly scale up our hardware 01:04:21.920 |
massively in the cloud, we're so limited, right? 01:04:26.160 |
So they are also Life, can soon become a little bit more 01:04:31.160 |
like Life 3.0 in that if they need more hardware, 01:04:41.000 |
- And what we haven't seen yet, which could change a lot, 01:04:53.920 |
So right now, programming is done sort of in bits and pieces 01:05:03.120 |
and with the kind of stuff that GPT-4 is able to do, 01:05:05.720 |
I mean, it's replacing a lot what I'm able to do. 01:05:25.440 |
kind of feedback loop of self-debugging, improving the code, 01:05:30.440 |
and then you launch that system onto the wild, 01:05:35.440 |
on the internet, because everything is connected, 01:05:37.400 |
and have it do things, have it interact with humans, 01:05:44.720 |
That's one of the things that Elon Musk recently 01:05:47.720 |
sort of tweeted as a case why everyone needs to pay $7 01:05:57.480 |
where the bots are getting smarter and smarter and smarter 01:06:01.120 |
to a degree where you can't tell the difference 01:06:10.620 |
by one million to one, which is why he's making a case 01:06:17.360 |
which is one of the only mechanisms to prove, 01:06:24.480 |
as individuals, we should, from time to time, 01:06:27.920 |
ask ourselves why are we doing what we're doing, 01:06:29.920 |
right, and as a species, we need to do that too. 01:06:49.480 |
and things that a lot of people find really meaningful, 01:06:54.900 |
The answer is Moloch is tricking us into doing it. 01:07:04.520 |
we still have no choice but to fall for it, right? 01:07:07.360 |
Also, the thing you said about you using co-pilot AI tools 01:07:17.400 |
what factor faster would you say you code now? 01:07:22.560 |
- I don't really, because it's such a new tool. 01:07:27.040 |
- I don't know if speed is significantly improved, 01:07:39.680 |
then you're already seeing another kind of self, 01:07:45.680 |
Because previously, one major generation of improvement 01:07:50.480 |
of the code would happen on the human R&D timescale. 01:07:57.960 |
than it otherwise would to develop the next level 01:08:09.040 |
There can be humans in the loop a lot in the early stages, 01:08:11.760 |
and then eventually humans are needed less and less, 01:08:16.440 |
But what you said there is just an exact example 01:08:27.520 |
imagining I'm on a psychiatrist's couch here, 01:08:29.680 |
saying, "What are my fears that people would do 01:08:33.040 |
So I mentioned three that I had fears about many years ago 01:08:37.080 |
that they would do, namely teach you the code, 01:08:48.200 |
where code can control this super powerful thing, right? 01:08:58.520 |
have going for them is that they are an oracle 01:09:00.840 |
in the sense that they just answer questions. 01:09:07.080 |
GPT-4 can't go and do stock trading based on its thinking. 01:09:16.560 |
processes it to figure out what action to take 01:09:26.460 |
But once you have an API, for example, GPT-4, 01:09:29.800 |
nothing stops Joe Schmo and a lot of other people 01:09:41.720 |
which makes them themselves much more powerful. 01:09:45.600 |
That's another kind of unfortunate development, 01:09:48.920 |
which I think we would have been better off delaying. 01:09:53.360 |
I don't want to pick on any particular companies. 01:09:55.040 |
I think they're all under a lot of pressure to make money. 01:09:58.260 |
And again, the reason we're calling for this pause 01:10:12.860 |
I hope we'll make it clear to people watching this 01:10:32.360 |
That's the definition of an explosion in science. 01:10:36.520 |
If you have two people and they fall in love, 01:10:55.160 |
if it's instead free neutrons in a nuclear reaction, 01:11:24.720 |
And do you have any intuition why it might stop? 01:11:29.400 |
when it bumps up against the laws of physics. 01:11:36.200 |
- 'Cause we don't know the full laws of physics yet, right? 01:11:42.680 |
on the physical limits on computation, for example. 01:11:49.000 |
then in finite space, it'll turn into a black hole. 01:11:53.320 |
faster than the speed of light, stuff like that. 01:11:58.680 |
than a modest number of bits per atom, et cetera. 01:12:02.720 |
But those limits are just astronomically above, 01:12:06.920 |
like 30 orders of magnitude above where we are now. 01:12:28.680 |
to make sure exactly it doesn't blow up out of control. 01:12:32.480 |
When we do experiments with biology and cells and so on, 01:12:37.480 |
we also try to make sure it doesn't get out of control. 01:12:56.360 |
or the other company is gonna catch up with you, 01:12:58.280 |
or the other country is gonna catch up with you. 01:13:23.840 |
you're putting people in a very hard situation. 01:13:26.920 |
is change the whole incentive structure instead. 01:13:31.520 |
Maybe I should say one more thing about this, 01:13:42.320 |
And we came up with some really cool countermeasures. 01:13:46.600 |
First of all, already over 100,000 years ago, 01:13:49.760 |
evolution realized that it was very unhelpful 01:13:52.960 |
that people kept killing each other all the time. 01:14:00.240 |
And made it so that if you get two drunk dudes 01:14:12.760 |
And similarly, if you find a baby lying on the street 01:14:18.160 |
when you go out for your morning jog tomorrow, 01:14:22.120 |
Even though it may make you late for your next podcast. 01:14:28.080 |
that make our own egoistic incentives more aligned 01:14:32.320 |
with what's good for the greater group we're part of. 01:14:51.680 |
moochers, cheaters, because their own incentive now 01:14:57.040 |
is not to do this, because word quickly gets around 01:15:00.880 |
and then suddenly people aren't gonna invite them 01:15:05.640 |
And then when we got still more sophisticated 01:15:07.560 |
and bigger societies, invented the legal system, 01:15:11.440 |
where even strangers who couldn't rely on gossip 01:15:21.080 |
that he actually wants to kill the other guy, 01:15:26.160 |
he also has a little thought in the back of his head 01:15:28.080 |
that, "Do I really wanna spend the next 10 years 01:15:38.760 |
And we similarly have tried to give these incentives 01:15:45.760 |
so that their incentives are aligned with the greater good. 01:16:01.480 |
than the regulators have been able to keep up. 01:16:06.720 |
like European Union right now is doing this AI act, right? 01:16:16.040 |
that GPT-4 would be completely excluded from regulation. 01:16:24.240 |
- Some lobbyists pushed successfully for this. 01:16:30.080 |
Mark Brackel, Christo Ouk, Anthony Aguirre, and others. 01:16:34.160 |
We're quite involved with educating various people 01:16:42.960 |
and pointing out that they would become the laughingstock 01:16:52.520 |
And then there was a huge counter push from lobbyists. 01:16:56.800 |
There were more lobbyists in Brussels from tech companies 01:17:14.240 |
but the challenge we're facing is that the tech 01:17:18.080 |
is generally much faster than what the policymakers are. 01:17:28.160 |
So it's, you know, we really need to work hard 01:17:34.800 |
So we're getting the situation where the first kind of non, 01:17:46.160 |
a company, a corporation is also an artificial intelligence 01:17:50.560 |
because the corporation isn't, it's humans, it's a system. 01:17:56.040 |
if the CEO of a tobacco company decides one morning 01:17:58.760 |
that she or he doesn't wanna sell cigarettes anymore, 01:18:02.840 |
It's not enough to align the incentives of individual people 01:18:08.080 |
or align individual computers' incentives to their owners, 01:18:12.920 |
which is what technically AI safety research is about. 01:18:16.120 |
You also have to align the incentives of corporations 01:18:19.840 |
And some corporations have gotten so big and so powerful 01:18:30.440 |
to what they want rather than the other way around. 01:18:35.600 |
- All right, is the thing that the slowdown hopes to achieve 01:18:40.400 |
is give enough time to the regulators to catch up 01:18:43.560 |
or enough time to the companies themselves to breathe 01:18:46.280 |
and understand how to do AI safety correctly? 01:18:52.000 |
the path to success I see is first you give a breather 01:18:58.040 |
their leadership who wants to do the right thing 01:19:00.240 |
and they all have safety teams and so on on their companies. 01:19:03.080 |
Give them a chance to get together with the other companies 01:19:08.720 |
and the outside pressure can also help catalyze that 01:19:21.200 |
one should put on future systems before they get rolled out? 01:19:35.480 |
that within six months you can get these people coming up, 01:19:48.080 |
but they got together a bunch of people and decided, 01:19:50.200 |
you know, in order to be allowed to sell a car, 01:19:53.920 |
They're the analogous things that you can start requiring 01:20:08.080 |
this intellectual work has been done by experts in the field, 01:20:13.520 |
I think it's going to be quite easy to get policymakers 01:20:19.360 |
And it's, you know, for the companies to fight Moloch, 01:20:29.120 |
they want the regulators to actually adopt it 01:20:31.000 |
so that their competition is going to abide by it too, right? 01:20:33.840 |
You don't want to be enacting all these principles 01:20:46.880 |
and then now they can gradually overtake you. 01:20:54.280 |
knowing that everybody's playing by the same rules. 01:20:56.680 |
- So do you think it's possible to develop guardrails 01:21:09.200 |
while still enabling sort of the capitalist-fueled 01:21:13.640 |
as they develop how to best make money with this AI? 01:21:20.560 |
in many other sectors where you've had the free market 01:21:23.240 |
produce quite good things without causing particular harm. 01:21:35.360 |
for just getting the same things done more efficiently. 01:21:48.160 |
in any country who thinks it was a terrible idea 01:21:55.200 |
- Yeah, but it seems like this particular technology 01:22:02.560 |
to a degree where you could see in the near term 01:22:07.800 |
and to put guardrails, to develop guardrails quickly 01:22:16.640 |
It seems like the opportunity to make a lot of money here 01:22:29.000 |
the more money there is, the more gold ingots 01:22:32.720 |
there are on the ground you can pick up or whatever 01:22:36.080 |
But it's not in anyone's incentive that we go over the cliff 01:22:38.720 |
and it's not like everybody's in their own car. 01:22:40.920 |
All the cars are connected together with a chain. 01:22:43.680 |
So if anyone goes over, they'll start dragging the others down too. 01:22:48.160 |
And so ultimately, it's in the selfish interests 01:22:52.560 |
also of the people in the companies to slow down 01:22:56.200 |
when you start seeing the contours of the cliff 01:23:03.080 |
who are building the technology and the CEOs, 01:23:12.400 |
they are people who don't honestly understand 01:23:19.600 |
to really appreciate how powerful this is and how fast. 01:23:22.560 |
And a lot of people are even still stuck again 01:23:41.440 |
that intelligence is information processing of a certain kind. 01:23:48.000 |
whether the information is processed by carbon atoms 01:24:00.720 |
and there are a lot of people who love capitalism 01:24:02.560 |
and a lot of people who really, really don't. 01:24:07.560 |
And it struck me recently that what's happening 01:24:16.360 |
to the way in which superintelligence might wipe us out. 01:24:20.120 |
So, you know, I studied economics for my undergrad, 01:24:31.000 |
- So I was very interested in how you could use 01:24:34.080 |
market forces to just get stuff done more efficiently, 01:24:51.520 |
where they proved mathematically that if you just take 01:24:59.720 |
that you think is gonna bring you in the right direction. 01:25:05.440 |
in the beginning, it will make things better for you. 01:25:11.320 |
it's gonna start making things worse for you again. 01:25:16.400 |
So just as a simple, the way I think of the proof is, 01:25:20.520 |
suppose you wanna go from here back to Austin, for example, 01:25:25.520 |
and you're like, okay, yeah, let's just, let's go south, 01:25:47.240 |
and eventually you're gonna be leaving the solar system. 01:25:51.000 |
- And they proved, it's a beautiful mathematical proof. 01:25:53.440 |
This happens generally, and this is very important for AI, 01:25:57.800 |
because even though Stuart Russell has written a book 01:26:02.240 |
and given a lot of talks on why it's a bad idea 01:26:12.280 |
that we're just minimizing, or reward function, 01:26:30.440 |
Things got done much more efficiently than they did 01:26:46.480 |
and ever more efficient information processing, 01:26:51.080 |
And eventually a lot of people are beginning to feel, 01:26:55.360 |
wait, we're kind of optimizing a bit too much. 01:26:57.320 |
Like, why did we just chop down half the rainforest? 01:27:11.040 |
If you have an AI that actually has power over the world 01:27:23.480 |
But it's almost impossible to give it exactly 01:27:29.920 |
And then eventually all hay breaks loose, right? 01:27:38.840 |
What if you just wanna tell it to cure cancer or something, 01:27:45.120 |
Maybe it's gonna decide to take over entire continents 01:27:50.120 |
just so it can get more supercomputer facilities in there 01:27:55.960 |
And then you're like, wait, that's not what I wanted, right? 01:28:12.360 |
that is optimizing for only one thing, profit. 01:28:16.680 |
And that worked great back when things were very inefficient 01:28:24.760 |
were small enough that they couldn't capture the regulators. 01:28:28.080 |
But that's not true anymore, but they keep optimizing. 01:28:37.000 |
can make even more profit by building ever more powerful AI 01:28:40.680 |
but optimize more and more and more and more and more. 01:28:50.280 |
And I just wanna, anyone here who has any concerns 01:28:54.200 |
about late-stage capitalism having gone a little too far, 01:29:02.400 |
'cause it's the same villain in both cases, it's Moloch. 01:29:10.040 |
aggressively, blindly is going to take us there. 01:29:16.080 |
and look into our hearts and ask, why are we doing this? 01:29:27.400 |
- And that is the idea behind a halt for six months. 01:29:34.200 |
Can we just linger and explore different ideas here? 01:29:37.680 |
Because this feels like a really important moment 01:29:40.160 |
in human history where pausing would actually 01:29:50.480 |
the number one pushback we were gonna get in the West 01:29:57.960 |
And everybody knows there's no way that China 01:30:01.040 |
is gonna catch up with the West on this in six months. 01:30:05.800 |
and you can forget about geopolitical competition 01:30:13.800 |
But you've already made the case that even for China, 01:30:20.640 |
China too would not be bothered by a longer halt 01:30:47.400 |
- Yeah, it's gonna get better and better and better 01:30:55.000 |
of basically quality of life for human civilization 01:31:06.320 |
is that we're just gonna, there won't be any humans. 01:31:10.560 |
We can see in history, once you have some species 01:31:15.000 |
or some group of people who aren't needed anymore, 01:31:18.180 |
doesn't usually work out so well for them, right? 01:31:26.440 |
for traffic in Boston and then the car got invented 01:31:29.080 |
and most of them got, yeah, well, we don't need to go there. 01:32:05.120 |
aren't needed anymore, I think it's quite naive 01:32:07.760 |
to think that they're gonna still be treated well. 01:32:13.200 |
and the government will always, we'll always protect them. 01:32:24.200 |
And now, in the beginning, so Industrial Revolution, 01:32:38.680 |
and got usually more interesting, better paid jobs. 01:32:42.520 |
But now we're beginning to replace brain work. 01:32:50.920 |
and adding, multiplying numbers anymore at work. 01:33:13.120 |
about this guy who was doing 3D modeling for gaming 01:33:17.320 |
and all of a sudden now they got this new software, 01:33:20.960 |
he just says prompts and he feels his whole job 01:33:27.320 |
And I asked GPT-4 to rewrite "Twinkle, Twinkle Little Star" 01:33:39.920 |
You've seen a lot of the art coming out here. 01:33:42.160 |
So I'm all for automating away the dangerous jobs 01:34:22.880 |
but you'll still end up needing fewer programmers than today. 01:34:29.960 |
So we need to stop and ask ourselves why again 01:34:36.720 |
I feel that AI should be built by humanity for humanity. 01:34:48.800 |
Or what it really is now is kind of by humanity for Moloch, 01:35:00.280 |
if we develop, figure out gradually and safely 01:35:04.920 |
And then we think about what are the kind of jobs 01:35:11.480 |
what are the jobs that people really find meaning in? 01:35:15.240 |
Like maybe taking care of children in the daycare center, 01:35:23.320 |
And even if it were possible to automate that way, 01:35:33.760 |
or rediscover what are the jobs that give us meaning. 01:35:43.920 |
half the time I'm crying as I'm generating code 01:35:58.240 |
and then you bring it to life and it does something, 01:36:00.080 |
especially if there's some intelligence that it does something. 01:36:06.240 |
you made a little machine and it comes to life. 01:36:11.080 |
- And there's a bunch of tricks you learn along the way 01:36:13.840 |
'cause you've been doing it for many, many years. 01:36:29.040 |
It's almost painful, like a loss of innocence maybe. 01:36:36.520 |
I remember before I learned that sugar's bad for you, 01:36:48.760 |
I enjoyed it unapologetically, fully, just intensely. 01:36:55.840 |
Now I feel like a little bit of that is lost for me 01:36:59.400 |
with programming, or being lost with programming, 01:37:06.440 |
no longer being able to really enjoy the art of modeling 01:37:11.840 |
I don't know, I don't know what to make sense of that. 01:37:24.040 |
to really intensify the value from conscious experiences 01:37:36.480 |
And just to inject some optimism in this here, 01:38:09.160 |
including the richest people can be better off as well. 01:38:17.000 |
Again, you can have two countries like Sweden and Denmark 01:38:20.440 |
have all these ridiculous wars century after century. 01:38:49.800 |
basically all the limitations that cause harm today 01:39:05.640 |
And we can talk about ways of making it safe also. 01:39:27.240 |
I just read a heartbreaking study from the CDC 01:39:37.840 |
Those are steps in totally the wrong direction. 01:39:42.600 |
And it's important to keep our eyes on the prize here 01:39:45.880 |
that we have the power now for the first time 01:40:05.840 |
To help us have really fulfilling experiences 01:40:13.680 |
and dictate to future generations what they will be. 01:40:18.680 |
and not foreclose all these possibilities for them 01:40:23.040 |
- And for that, we'll have to solve the AI safety problem. 01:40:29.520 |
So one interesting way to enter that discussion 01:40:37.920 |
You tweeted, "Let's not just focus on whether GPT-4 01:40:40.400 |
"will do more harm or good on the job market, 01:40:44.580 |
"will hasten the arrival of superintelligence." 01:40:47.480 |
That's something we've been talking about, right? 01:41:06.760 |
And in general, what are your different ideas 01:41:09.960 |
to start approaching the solution to AI safety? 01:41:24.520 |
but that's because we made it in a certain way. 01:41:28.800 |
We can use it for great things and bad things. 01:41:33.240 |
And this is part of my vision for success here, 01:41:36.840 |
truth-seeking AI that really brings us together again. 01:41:46.080 |
It's because they each have totally different versions 01:42:14.960 |
for example, a little baby step in this direction 01:42:25.320 |
not for money, but just for their own reputation. 01:42:32.400 |
as you have a loss function where they get penalized 01:42:39.440 |
Whereas if you're kind of humble and then you're like, 01:42:43.120 |
I think it's 51% chance this is gonna happen, 01:42:45.360 |
and then the other happens, you don't get penalized much. 01:42:57.680 |
is an outgrowth of Improve the News Foundation 01:43:00.520 |
is seeing if we can really scale this up a lot 01:43:24.160 |
of different pundits and newspapers, et cetera, 01:43:27.480 |
if they wanna know why someone got a low score, 01:43:29.840 |
they can click on it and see all the predictions 01:43:32.440 |
that they actually made and how they turned out. 01:43:38.160 |
You trust scientists like Einstein who said something 01:43:40.560 |
everybody thought was bullshit and turned out to be right. 01:43:53.800 |
a lot of the rifts we're seeing by creating trust systems. 01:44:05.760 |
and you just trust it because of its reputation. 01:44:13.080 |
they earn their trust and they're completely transparent. 01:44:21.400 |
the very dysfunctional conversation that humanity has 01:44:24.920 |
about how it's gonna deal with all its biggest challenges 01:44:39.400 |
from people who are saying, "We're just screwed. 01:44:40.920 |
"There's no hope," is, well, things like GPT-4 01:44:44.120 |
are way too complicated for a human to ever understand 01:44:53.480 |
There's this very fundamental fact that in math, 01:45:01.760 |
than it is to verify that the proof is correct. 01:45:04.920 |
You can actually write a little proof-checking code 01:45:07.040 |
which is quite short, but you can, as a human, understand it. 01:45:10.640 |
And then it can check the most monstrously long proof 01:45:26.880 |
with virus-checking software that it looks to see 01:45:29.680 |
if there's something, if you should not trust it. 01:45:33.160 |
that you should not trust that code, it warns you. 01:45:40.000 |
And this is an idea I give credit to Steve on Mahindra for. 01:45:44.240 |
So that it will only run the code if it can prove, 01:45:49.000 |
that it's not trustworthy, if it will only run it 01:45:52.920 |
So it asks the code, "Prove to me that you're gonna do 01:46:06.480 |
that's much more intelligent than you are, right? 01:46:08.880 |
Because it's a problem to come up with this proof 01:46:13.440 |
that you could never have found, but you should trust it. 01:46:17.760 |
I agree with you, but this is where Eliezer Yakovsky 01:46:30.680 |
would be able to know how to lie to you with such a proof. 01:46:45.160 |
- So his general idea is a super-intelligent system 01:47:05.400 |
He really focuses on this weak AGI to strong AGI jump 01:47:11.680 |
where the strong AGI can make all the weak AGIs 01:47:15.760 |
think that it's just one of them, but it's no longer that. 01:47:25.720 |
I think no matter how super-intelligent an AI is, 01:47:30.880 |
that there are only finitely many primes, for example. 01:47:42.840 |
that say, trust me, the way your proof-checker works 01:47:47.840 |
is too limited, and we have this new hyper-math, 01:48:00.000 |
I'm only gonna go with the ones that I can prove 01:48:05.320 |
There's still, of course, this is not something 01:48:08.520 |
anyone has successfully implemented at this point, 01:48:10.360 |
but I think I just give it as an example of hope. 01:48:17.160 |
This is exactly the sort of very boring and tedious task 01:48:35.840 |
- Well, for starters, begin with a simple problem 01:48:39.160 |
of just making sure that the system that you own 01:48:44.320 |
has to prove to itself that it's always gonna do 01:48:48.240 |
And if it can't prove it, maybe it's still gonna do it, 01:49:06.880 |
what direction we should go with humanity, right? 01:49:10.840 |
- And you've talked a lot about geopolitical things 01:49:19.120 |
from the fact that there are actually a lot of things 01:49:21.680 |
that everybody in the world virtually agrees on 01:49:25.920 |
that, hey, you know, like having no humans on the planet 01:49:36.000 |
the United Nations Sustainable Development Goals, 01:49:47.960 |
So instead of quibbling about the little things 01:49:50.960 |
we don't agree on, let's start with the things 01:49:56.720 |
Instead of being so distracted by all these things 01:50:07.840 |
it feels like a war on life playing out in front of our eyes. 01:50:15.960 |
we're on this planet, beautiful, vibrant ecosystem. 01:50:24.680 |
even though most people thought that was a bad idea. 01:50:36.720 |
And we're replacing more and more of the biosphere 01:50:45.600 |
a lot of the things which were so valuable to humanity. 01:50:49.120 |
A lot of social interactions now are replaced 01:50:51.360 |
by people staring into their rectangles, right? 01:50:54.320 |
And I'm not a psychologist, I'm out of my depth here, 01:50:58.640 |
but I suspect that part of the reason why teen suicide 01:51:04.760 |
the record breaking levels is actually caused by, 01:51:16.320 |
We've all seen a bunch of good looking people 01:51:22.240 |
instead of looking into each other's eyes, right? 01:51:41.640 |
The technology that was supposed to connect us 01:51:43.800 |
is actually distancing us, ourselves from each other. 01:51:52.680 |
These large corporations are not living things, right? 01:52:01.960 |
I think we humans, together with all our fellow living things 01:52:35.560 |
very close to one, depending on the day he puts it at one, 01:52:42.800 |
That there's just, he does not see a trajectory 01:52:47.320 |
which it doesn't end up with that conclusion. 01:52:50.920 |
What trajectory do you see that doesn't end up there? 01:53:01.300 |
- First of all, I tremendously respect Eliezer Yudkowsky 01:53:22.280 |
We just had a little baby, and I keep asking myself, 01:53:39.200 |
it feels a little bit like I was just diagnosed 01:53:59.220 |
I think, but I absolutely don't think it's hopeless. 01:54:05.360 |
I think there is, first of all, a lot of momentum now, 01:54:16.960 |
since I and many others started warning about this, 01:54:34.440 |
and he's like, "I think we're getting replaced." 01:54:37.880 |
So that's positive, that we're finally seeing this reaction, 01:54:44.360 |
which is the first step towards solving the problem. 01:54:57.880 |
it's really just virus checking in reverse again. 01:55:05.120 |
We might have to forfeit some of the technology 01:55:08.960 |
that we could get if we were putting blind faith in our AIs, 01:55:14.360 |
- Do you envision a process with a proof checker, 01:55:18.840 |
would go through a process of rigorous interrogation? 01:55:23.200 |
That's like trying to prove Vero about five spaghetti. 01:55:27.720 |
What I think, well, the vision I have for success 01:55:38.160 |
Galileo, when his dad threw him an apple when he was a kid, 01:55:43.360 |
'cause his brain could in this funny spaghetti kind of way, 01:55:49.640 |
Then he got older and it's like, wait, this is a parabola. 01:55:56.960 |
and today you can easily program it into a computer 01:56:05.640 |
where we use the amazing learning power of neural networks 01:56:09.120 |
to discover the knowledge in the first place, 01:56:12.320 |
but we don't stop with a black box and use that. 01:56:19.160 |
where we use automated systems to extract out the knowledge 01:56:28.280 |
into a completely different kind of architecture 01:56:33.640 |
that's made in a way that it can be both really efficient 01:56:37.040 |
and also is more amenable to very formal verification. 01:56:48.600 |
but I don't think, the chance is certainly not zero either, 01:56:58.920 |
So we can have a lot of the fun that we're excited about 01:57:26.240 |
because there's no more guaranteed way to fail 01:57:29.840 |
than to convince yourself that it's impossible and not try. 01:57:40.920 |
that that's how you do psychological warfare. 01:57:44.680 |
You persuade the other side that it's hopeless 01:57:51.520 |
Let's not do this psychological warfare on ourselves 01:57:58.400 |
It's sadly, I do get that a little bit sometimes 01:58:08.160 |
I'm just gonna play computer games and do drugs 01:58:20.120 |
and makes it more likely that we're gonna succeed. 01:58:22.720 |
- It seems like the people that actually build solutions 01:58:25.680 |
to a problem seemingly impossible to solve problems 01:58:33.040 |
And it seems like there's some fundamental law 01:58:36.560 |
to the universe where fake it 'til you make it 01:58:40.000 |
Like, believe it's possible and it becomes possible. 01:58:46.000 |
if you tell yourself that it's impossible, it is? 01:58:57.400 |
Everybody's so gloomy and the media are also very biased 01:58:59.920 |
towards if it bleeds, it leads and gloom and doom. 01:59:07.400 |
are dystopian, which really demotivates people. 01:59:12.600 |
We wanna really, really, really focus on the upside also 01:59:16.000 |
to give people the willingness to fight for it. 01:59:18.800 |
And for AI, you and I mostly talked about gloom here again, 01:59:23.800 |
but let's not forget that we have probably both lost 01:59:30.000 |
someone we really cared about to some disease 01:59:52.280 |
So if we can get this right, just be a little more chill 01:59:56.760 |
and slow down a little bit till we get it right. 01:59:59.160 |
It's mind blowing how awesome our future can be. 02:00:04.480 |
We talked a lot about stuff on earth, it can be great. 02:00:11.280 |
there's no reason we have to be stuck on this planet 02:00:22.480 |
let life spread out into space to other solar systems, 02:00:29.960 |
And this to me is a very, very hopeful vision 02:00:42.920 |
is one of the things that also really gives meaning 02:00:46.440 |
If there's ever been an epic struggle, this is it. 02:00:50.080 |
And isn't it even more epic if you're the underdog? 02:00:53.320 |
If most people are telling you this is gonna fail, 02:01:01.040 |
That's what we can do together as a species on this one. 02:01:05.280 |
A lot of pundits are ready to count this out. 02:01:17.640 |
that's how we're gonna get multi-planetary very efficiently. 02:02:04.280 |
Should we open source how to make bioweapons? 02:02:09.680 |
Should we open source how to make a new virus 02:02:19.600 |
It's already that powerful that we have to respect 02:02:46.440 |
Baby, sort of baby proto, almost little bit AGI 02:02:50.720 |
according to what Microsoft's recent paper said. 02:02:55.640 |
What we're scared about is people taking that 02:03:12.200 |
There are many things which are not open sourced 02:03:27.000 |
you don't open source those things for a reason. 02:03:34.920 |
I have to say it feels in a way a bit weird to say it 02:03:38.120 |
because MIT is like the cradle of the open source movement. 02:03:59.640 |
all the most dangerous things he can do in the house. 02:04:07.040 |
because this is one of the first moments in history 02:04:15.740 |
This is when the software has become too dangerous. 02:04:21.160 |
that we didn't wanna open source a technology. 02:04:28.400 |
about how to get the release of such systems right? 02:04:33.980 |
So OpenAI went through a pretty rigorous effort 02:04:40.300 |
but nevertheless it's longer than you would have expected 02:04:54.300 |
Or how do I say I hate a certain group on Twitter 02:05:00.180 |
in a way that doesn't get me blocked from Twitter, 02:05:02.260 |
banned from Twitter, those kinds of questions. 02:05:15.460 |
having thought about this problem of AI safety, 02:05:29.900 |
that the two biggest risks from large language models are 02:05:42.380 |
And second, being used for offensive cyber weapon. 02:05:48.620 |
So I think those are not the two greatest threats. 02:05:53.300 |
They're very serious threats and it's wonderful 02:06:00.220 |
is how is this just gonna disrupt our economy 02:06:03.620 |
and maybe take away a lot of the most meaningful jobs. 02:06:17.860 |
- Write code, connect it to the internet, manipulate humans. 02:06:21.120 |
- Yeah, and before we know it, we have something else, 02:06:26.860 |
but which is way more intelligent and capable and has goals. 02:06:37.920 |
that's not something that's easy for them to verify 02:06:42.460 |
And the only way to really lower that risk a lot 02:06:48.940 |
never let it read any code, not train on that, 02:06:54.400 |
and not give it access to so much information 02:07:13.720 |
Microsoft is rolling out the new Office suite 02:07:17.680 |
where you go into Microsoft Word and give it a prompt, 02:07:35.920 |
and whether society is prepared to deal with this disruption, 02:07:43.560 |
that keeps me awake at night for wiping out humanity. 02:07:46.200 |
And I think that's the biggest misunderstanding we have. 02:07:56.560 |
That's not what Eliezer was freaked out about either. 02:08:09.600 |
you've talked about a lot, is autonomous weapon systems. 02:08:17.200 |
Is that one of the things that still you carry concern for 02:08:21.200 |
as these systems become more and more powerful? 02:08:24.120 |
not that all humans are going to get killed by slaughterbots, 02:08:26.480 |
but rather just as express route into Orwellian dystopia 02:08:31.480 |
where it becomes much easier for very few to kill very many 02:08:38.200 |
If you want to know how AI could kill all people, 02:08:58.000 |
and stepping on them or shooting them or anything like that. 02:09:04.240 |
In some cases, we did it by putting more carbon dioxide 02:09:15.800 |
So if you're an AI and you just want to figure something out 02:09:20.800 |
then you decide, we just really need the space here 02:09:34.240 |
we are just the sort of accidental roadkill along the way. 02:09:54.920 |
We've driven a number of the elephant species extinct. 02:10:04.420 |
The basic problem is you just don't want to give, 02:10:11.040 |
you don't want to cede control over your planet 02:10:23.720 |
which AI safety researchers have been grappling with 02:10:27.440 |
How do you make AI first of all understand our goals 02:10:32.760 |
and then retain them as they get smarter, right? 02:10:44.080 |
Like a human child, first they're just not smart enough 02:10:59.080 |
But there is fortunately a magic phase in the middle 02:11:03.760 |
where they're smart enough to understand our goals 02:11:06.840 |
with good parenting teach them right from wrong 02:11:12.280 |
So those are all tough challenges with computers. 02:11:17.960 |
And then even if you teach your kids good goals 02:11:20.560 |
when they're little, they might outgrow them too. 02:11:22.300 |
And that's a challenge for machines to keep improving. 02:11:25.720 |
So these are a lot of hard challenges we're up for, 02:11:30.380 |
but I don't think any of them are insurmountable. 02:11:33.240 |
The fundamental reason why Eliezer looked so depressed 02:11:46.000 |
- He was hoping that humanity was gonna take this threat 02:11:53.360 |
That's why the open letter is calling for more time. 02:11:56.360 |
- But even with time, the AI alignment problem 02:12:11.660 |
the most important problem for humanity to ever solve. 02:12:15.940 |
that aligned AI can help us solve all the other problems. 02:12:20.740 |
- 'Cause it seems like it has to have constant humility 02:12:23.940 |
about its goal, constantly questioning the goal. 02:12:26.440 |
Because as you optimize towards a particular goal 02:12:32.580 |
that's when you have the unintended consequences, 02:12:35.940 |
So how do you enforce and code a constant humility 02:12:40.000 |
as your ability become better and better and better and better 02:12:42.920 |
- Stewart, Professor Stewart Russell at Berkeley, 02:12:44.760 |
who's also one of the driving forces behind this letter, 02:12:59.080 |
Although he calls it inverse reinforcement learning 02:13:15.220 |
but I'm not gonna tell you right now what it is 02:13:19.260 |
So then you give the incentive to be very humble 02:13:25.660 |
And oh, this other thing I tried didn't work, 02:13:33.240 |
What's nice about this is it's not just philosophical 02:13:35.600 |
mumbo jumbo, it's theorems and technical work 02:13:38.320 |
that with more time I think you can make a lot of progress. 02:13:47.840 |
- But also not that many relative to the scale of the problem. 02:13:51.960 |
There should be, at least just like every university 02:13:56.400 |
worth its name has some cancer research going on 02:14:17.180 |
to what's happening here, even though I think many people 02:14:20.820 |
would wish it would have been rolled out more carefully, 02:14:31.540 |
stop fantasising about this being 100 years off 02:14:35.080 |
and stop fantasising about this being completely 02:14:37.600 |
controllable and predictable because it's so obvious 02:15:00.080 |
It was not 'cause the engineers that built it 02:15:02.780 |
was like, "Heh heh heh heh heh, let's put this in here 02:15:29.840 |
I would encourage them to just play a bit more 02:15:31.960 |
with these tools that are out there now, like GPT-4. 02:15:42.240 |
Once you've woken up, then gotta slow down a little bit 02:15:52.600 |
- You know, what's interesting is, you know, MIT, 02:15:58.680 |
but let's just even say computer science curriculum. 02:16:01.920 |
How does the computer science curriculum change now? 02:16:06.300 |
When I was coming up, programming is a prestigious position. 02:16:13.600 |
Like, why would you be dedicating crazy amounts of time 02:16:19.240 |
Like, the nature of programming is fundamentally changing. 02:16:30.840 |
and like think about, 'cause it's really turning. 02:16:33.960 |
- I mean, some English professors, some English teachers 02:16:44.880 |
And then they realize they have to completely rethink. 02:16:48.080 |
And even, you know, just like we stopped teaching, 02:16:52.920 |
writing a script, is that what you say in English? 02:16:59.160 |
- Yeah, when everybody started typing, you know, 02:17:01.200 |
like so much of what we teach our kids today. 02:17:09.960 |
everything is changing and it's changing very, 02:17:23.960 |
And if the education system is being turned on its head, 02:17:27.960 |
It feels like having these kinds of conversations 02:17:35.480 |
I don't think there's even, speaking of safety, 02:17:40.880 |
I don't think most universities have courses on AI safety. 02:18:01.580 |
and then they're gonna come out of high school 02:18:03.020 |
12 years later, and you've already pre-planned now 02:18:06.380 |
what they're gonna learn when you're not even sure 02:18:08.700 |
if there's gonna be any world left to come out to. 02:18:32.560 |
- If we just linger on the GPT-4 system a little bit, 02:18:36.160 |
you kind of hinted at it, especially talking about 02:18:41.920 |
the importance of consciousness in the human mind 02:18:57.560 |
because in my experience, like 90% of all arguments 02:19:00.880 |
about consciousness boil down to the two people 02:19:03.320 |
arguing, having totally different definitions 02:19:05.060 |
of what it is, and they're just shouting past each other. 02:19:08.240 |
I define consciousness as subjective experience. 02:19:13.720 |
Right now, I'm experiencing colors and sounds 02:19:22.680 |
That's the question about whether it's conscious or not. 02:19:26.400 |
Other people think you should define consciousness 02:19:34.960 |
I'm gonna use consciousness for this, at least. 02:19:50.060 |
Short answer, I don't know, because we still don't know 02:19:53.240 |
what it is that gives us wonderful subjective experience 02:19:56.680 |
that is kind of the meaning of our life, right? 02:20:13.760 |
Giulio Tononi, a professor, has stuck his neck 02:20:23.080 |
for what's the essence of conscious information processing. 02:20:45.460 |
where some of the output gets fed back in again. 02:20:56.360 |
where information only goes in one direction, 02:20:58.680 |
like from your eye, retina, into the back of your brain, 02:21:04.600 |
isn't conscious of anything, or a video camera. 02:21:11.360 |
is it's also just one-way flow of information. 02:21:14.800 |
So, if Tononi is right, GPT-4 is a very intelligent zombie 02:21:32.280 |
about turning off GPT-4 and wiping its memory 02:21:48.240 |
very high intelligence, perhaps, that's not conscious. 02:21:53.400 |
and it's sad enough that humanity isn't here anymore, 02:22:04.280 |
I could be like, well, but they're our descendants 02:22:06.200 |
and maybe they have our values, they're our children. 02:22:09.280 |
But if Tononi is right, and these are all transformers 02:22:32.580 |
That would be like the ultimate depressing future. 02:22:44.880 |
what kind of information processing actually has experience, 02:22:49.620 |
And I completely don't buy the dismissal that some people, 02:22:54.280 |
some people will say, well, this is all bullshit 02:23:03.160 |
when you're not really accomplishing any goals at all, 02:23:24.800 |
that we will not allow AI systems to have consciousness. 02:23:29.120 |
We'll come up with theories about measuring consciousness 02:23:37.120 |
because maybe we humans will create something 02:23:47.080 |
which is they have a deeper subjective experience 02:23:52.220 |
Not only are they smarter, but they feel deeper. 02:24:02.100 |
they'll be the other, we'll try to suppress it, 02:24:04.800 |
they'll create conflict, they'll create war, all of this. 02:24:46.200 |
where they show that lobsters actually do feel pain 02:24:49.580 |
So they banned lobster boiling in Switzerland now, 02:25:01.800 |
- We do the same thing with cruelty to farm animals, 02:25:03.880 |
also all these self-serving arguments for why they're fine. 02:25:20.160 |
where it's really information being aware of itself 02:25:23.160 |
And let's study it and give ourselves a little bit of time. 02:25:28.200 |
actually what it is that causes consciousness. 02:25:34.560 |
that do the boring jobs that we would feel immoral 02:25:38.440 |
But if you have a companion robot taking care of your mom 02:25:44.160 |
she would probably want it to be conscious, right? 02:25:45.760 |
So that the emotions it seems to display aren't fake. 02:25:59.400 |
- Is there something you could say to the timeline 02:26:02.000 |
that you think about, about the development of AGI? 02:26:05.920 |
Depending on the day, I'm sure that changes for you. 02:26:09.160 |
But when do you think there'll be a really big leap 02:26:13.560 |
in intelligence where you would definitively say 02:26:37.760 |
But I think that cat has really left the bag now. 02:26:46.600 |
I don't think the Microsoft paper is totally off 02:26:50.640 |
when they say that there are some glimmers of AGI. 02:27:09.720 |
because if there's ever been a time to pause, it's today. 02:27:14.280 |
- There's a feeling like this GPT-4 is a big transition 02:27:19.360 |
into waking everybody up to the effectiveness 02:27:33.040 |
And there are many companies trying to do these things. 02:27:54.280 |
- You have spoken about and thought about nuclear war a lot 02:28:13.700 |
- What do you learn about human nature from that? 02:28:23.480 |
America doesn't want there to be a nuclear war. 02:28:26.800 |
Russia doesn't want there to be a global nuclear war either. 02:28:40.240 |
that this is the closest we've come since 1962? 02:28:54.920 |
saying that we have to drive Russia out of Ukraine. 02:29:10.080 |
if we don't drive Russia out entirely of the Ukraine. 02:29:24.840 |
out of Ukraine, it's not just gonna be very humiliating 02:29:36.400 |
that things don't go so well for their leadership either. 02:29:39.640 |
You remember when Argentina invaded the Falkland Islands? 02:29:51.840 |
And then when they got their butt kicked by the British, 02:30:01.320 |
And I believe those who are still alive are in jail now. 02:30:04.600 |
So the Russian leadership is entirely cornered 02:30:09.320 |
where they know that just getting driven out of Ukraine 02:30:17.160 |
And so this to me is a typical example of Moloch. 02:30:30.920 |
If Russia starts losing in the conventional warfare, 02:30:36.480 |
since they're back against the war is to keep escalating. 02:30:39.280 |
And the West has put itself in the situation now 02:30:43.040 |
where we've sort of already committed to drive Russia out. 02:30:45.520 |
So the only option the West has is to call Russia's bluff 02:30:52.160 |
because Moloch can sometimes drive competing parties 02:30:55.480 |
to do something which is ultimately just really bad 02:31:02.720 |
is not just that it's difficult to see an ending, 02:31:12.320 |
that doesn't involve some horrible escalation, 02:31:21.640 |
There was an amazing paper that was published 02:31:30.480 |
who've been studying nuclear winter for a long time. 02:31:31.960 |
And what they basically did was they combined climate models 02:31:42.800 |
So instead of just saying, yeah, it gets really cold, 02:31:46.700 |
how many people would die in different countries. 02:31:52.280 |
So basically what happens is the thing that kills 02:31:56.000 |
it's not the radioactivity, it's not the EMP mayhem, 02:31:59.800 |
it's not the rampaging mobs foraging for food. 02:32:06.680 |
coming up from the burning cities and stratosphere 02:32:09.720 |
that spreads around the earth from the jet streams. 02:32:14.720 |
So in typical models, you get like 10 years or so 02:32:25.960 |
and their models, the temperature drops in Nebraska 02:32:38.920 |
No, yeah, 20, 30 Celsius, depending on where you are, 02:32:42.600 |
40 Celsius in some places, which is 40 Fahrenheit 02:32:46.160 |
to 80 Fahrenheit colder than what it would normally be. 02:32:48.840 |
So I'm not good at farming, but if it's snowing, 02:32:53.840 |
if it drops below freezing pretty much most days in July, 02:32:59.160 |
So they worked out, they put this into their farming models. 02:33:12.680 |
they had about 99% of all Americans starving to death. 02:33:21.120 |
So you might be like, oh, it's kind of poetic justice 02:33:31.360 |
But that doesn't particularly cheer people up in Sweden 02:33:45.880 |
not understanding very much just like how bad this is. 02:33:55.720 |
in decision-making positions still think of nuclear weapons 02:33:59.880 |
Scary, powerful, they don't think of it as something 02:34:05.000 |
where, yeah, just to within a percent or two, 02:34:11.880 |
- And starving to death is the worst way to die, 02:34:17.040 |
as a lot of more, as all the famines in history show, 02:34:27.160 |
- Probably brings out the worst in people also, 02:34:39.240 |
they'd rather be at ground zero and just get vaporized. 02:34:42.000 |
But I think people underestimate the risk of this 02:34:54.800 |
'cause humans don't want this, so it's not gonna happen. 02:35:04.360 |
- Exactly, and it applies to some of the things 02:35:09.000 |
that people have gotten most upset with capitalism for also, 02:35:14.920 |
It's not to see if some company does something 02:35:25.560 |
but she or he knew that all the other companies 02:35:55.440 |
or other mechanism that lets us instead realize 02:36:04.120 |
- It's not us versus them, it's us versus it. 02:36:06.480 |
- We are fighting, Moloch, for human survival. 02:36:27.540 |
but that's the closest I've seen to a movie about Moloch. 02:36:39.560 |
you can interpret him as the devil or whatever, 02:36:41.600 |
but he doesn't actually ever go around and kill people 02:36:44.200 |
or torture people with burning coal or anything. 02:36:49.200 |
makes everybody fear each other, hate each other, 02:36:57.460 |
That seems to be one of the ways to fight Moloch 02:37:02.460 |
is by compassion, by seeing the common humanity. 02:37:12.300 |
like what's it, Kumbaya tree huggers here, right? 02:37:21.860 |
understand the true facts about the other side. 02:38:01.240 |
If someone is like, "I really disagree with you, Lex, 02:38:07.860 |
"You're not a bad person who needs to be destroyed, 02:38:13.560 |
"and I'm happy to have an argument about it." 02:38:15.920 |
That's a lot of progress compared to where we are 02:38:18.760 |
at 2023 in the public space, wouldn't you say? 02:38:22.120 |
- If we solve the AI safety problem, as we've talked about, 02:38:26.560 |
and then you, Max Tegmark, who has been talking about this 02:38:31.040 |
for many years, get to sit down with the AGI, 02:38:35.120 |
with the early AGI system, on a beach with a drink, 02:39:02.760 |
I know I'm just a meat bag with all these flaws, 02:39:05.440 |
but I have, we talked a lot about homo sentiens. 02:39:09.920 |
I've already tried that for a long time with myself. 02:39:12.920 |
So that is what's really valuable about being alive for me, 02:39:19.760 |
It's not that I'm good at this or good at that or whatever. 02:39:30.760 |
In fact, my son reminds me of that pretty frequently. 02:39:34.160 |
- You could find out how dumb you are in terms of physics, 02:39:40.200 |
I think, so I can't waffle my way out of this question. 02:39:49.280 |
I think, given that I'm a really, really curious person, 02:40:13.360 |
I would just really, really love to understand also. 02:40:18.720 |
that I've been obsessing about a lot recently. 02:40:21.000 |
So I believe that, so suppose Tononi is right. 02:40:27.720 |
And suppose there are some information processing systems 02:40:32.560 |
Suppose you can even make reasonably smart things 02:40:38.840 |
Here's the question that keeps me awake at night. 02:40:41.280 |
Is it the case that the unconscious zombie systems 02:40:47.800 |
that are really intelligent are also really efficient? 02:40:51.600 |
So that when you try to make things more efficient, 02:40:54.960 |
which there'll naturally be a pressure to do, 02:41:02.480 |
And do you want me to give you a hand wavey argument for it? 02:41:11.160 |
how these large language models do something, 02:41:17.640 |
If you, we have loops in our computer language for a reason. 02:41:37.000 |
even an operating system knows things about itself. 02:41:53.240 |
of implementing a given level of intelligence 02:41:55.840 |
has loops in it, self-reflection, and will be conscious. 02:42:09.560 |
And I think if you look at our brains, actually, 02:42:12.160 |
our brains are part zombie and part conscious. 02:42:16.920 |
When I open my eyes, I immediately take all these pixels 02:42:29.880 |
But I have no freaking clue of how I did that computation. 02:42:36.640 |
we could even do it well with machines, right? 02:42:39.720 |
You get a bunch of information processing happening 02:42:42.160 |
in my retina, and then it goes to the lateral geniculate 02:42:44.480 |
nucleus, my thalamus, and the area V1, V2, V4, 02:42:48.520 |
and the fusiform face area here that Nancy Kenwisher 02:42:51.040 |
at MIT invented, and blah, blah, blah, blah, blah. 02:42:53.400 |
And I have no freaking clue how that worked, right? 02:42:56.320 |
It feels to me subjectively like my conscious module 02:43:18.440 |
because this was all one-way information processing mainly. 02:43:46.840 |
that if you want to fake that with a zombie system 02:43:49.120 |
that just all goes one way, you have to unroll those loops 02:43:55.840 |
So I'm actually hopeful that AI, if in the future 02:43:59.320 |
we have all these very sublime and interesting machines 02:44:14.520 |
to great consciousness or a deep kind of consciousness. 02:44:21.760 |
'cause the zombie apocalypse really is my worst nightmare 02:44:29.040 |
but we frigging replaced ourselves by zombies. 02:44:37.080 |
That's one that we humans can intuit and prove 02:44:42.520 |
as we start to understand what it means to be intelligent 02:44:55.200 |
Most of my colleagues, when I started going on 02:44:57.440 |
about consciousness, tell me that it's all bullshit 02:45:04.240 |
and from my mom saying, "Keep talking about it," 02:45:08.040 |
And the main way to convince people like that, 02:45:13.040 |
that they're wrong, if they say that consciousness 02:45:17.120 |
is just equal to intelligence, is to ask them, 02:45:23.960 |
If it's just about these particles moving this way 02:45:28.920 |
rather than that way, and there is no such thing 02:45:31.680 |
as subjective experience, what's wrong with torture? 02:45:36.520 |
- No, it seems like suffering imposed onto other humans 02:45:46.120 |
- And if someone tells me, "Well, it's just an illusion, 02:45:50.720 |
"consciousness, whatever," I like to invite them 02:46:03.760 |
If you have it, you can have it at local anesthesia 02:46:12.560 |
It just removed my subjective experience of pain. 02:46:15.560 |
It didn't change anything about what was actually happening 02:46:26.680 |
- It could be fundamental to whatever this thing 02:46:31.320 |
- It is fundamental because what we feel is so fundamental 02:46:36.080 |
is suffering and joy and pleasure and meaning. 02:46:41.080 |
That's all, those are all subjective experiences there. 02:46:47.880 |
And let's not, those are the elephant in the room. 02:46:58.400 |
- And let's not make the mistake of not instilling 02:47:02.640 |
the AI systems with that same thing that makes us special. 02:47:09.600 |
- Max, it's a huge honor that you will sit down to me 02:47:12.800 |
the first time on the first episode of this podcast. 02:47:18.280 |
and talk about this, what I think is the most important 02:47:21.400 |
topic, the most important problem that we humans 02:47:28.740 |
- Yeah, well the honor is all mine and I'm so grateful 02:47:31.600 |
to you for making more people aware of the fact 02:47:34.960 |
that humanity has reached the most important fork 02:47:45.440 |
To support this podcast, please check out our sponsors 02:47:59.120 |
Thank you for listening and hope to see you next time.