back to indexNick Bostrom: Simulation and Superintelligence | Lex Fridman Podcast #83
Chapters
0:0 Introduction
2:48 Simulation hypothesis and simulation argument
12:17 Technologically mature civilizations
15:30 Case 1: if something kills all possible civilizations
19:8 Case 2: if we lose interest in creating simulations
22:3 Consciousness
26:27 Immersive worlds
28:50 Experience machine
41:10 Intelligence and consciousness
48:58 Weighing probabilities of the simulation argument
61:43 Elaborating on Joe Rogan conversation
65:53 Doomsday argument and anthropic reasoning
83:2 Elon Musk
85:26 What's outside the simulation?
89:52 Superintelligence
107:27 AGI utopia
112:41 Meaning of life
00:00:00.000 |
The following is a conversation with Nick Bostrom, 00:00:05.520 |
and the director of the Future of Humanity Institute. 00:00:08.700 |
He has worked on fascinating and important ideas 00:00:16.860 |
and the risks of superintelligent AI systems, 00:00:23.200 |
I can see talking to Nick multiple times in this podcast, 00:00:30.440 |
in artificial intelligence, in technology space, 00:00:40.480 |
before the outbreak of the coronavirus pandemic, 00:00:45.800 |
will have a lot to say about next time we speak. 00:00:52.920 |
only in retrospect, when the storm has passed. 00:01:09.360 |
For everyone feeling the medical, psychological, 00:01:29.080 |
or simply connect with me on Twitter at Lex Friedman, 00:01:34.840 |
As usual, I'll do one or two minutes of ads now, 00:01:58.800 |
Since Cash App does fractional share trading, 00:02:02.240 |
let me mention that the order execution algorithm 00:02:06.480 |
to create the abstraction of fractional orders 00:02:15.840 |
provides an easy interface that takes a step up 00:02:18.480 |
to the next layer of abstraction over the stock market, 00:02:21.120 |
making trading more accessible for new investors 00:02:37.040 |
an organization that is helping to advance robotics 00:02:39.880 |
and STEM education for young people around the world. 00:02:42.720 |
And now, here's my conversation with Nick Bostrom. 00:02:48.080 |
At the risk of asking the Beatles to play "Yesterday" 00:02:51.600 |
or the Rolling Stones to play "Satisfaction," 00:02:59.320 |
- That we are living in a computer simulation. 00:03:04.280 |
How are we supposed to even think about that? 00:03:07.240 |
- Well, so the hypothesis is meant to be understood 00:03:12.640 |
not that we can kind of metaphorically view the universe 00:03:17.160 |
as an information processing physical system, 00:03:30.560 |
of what's going on inside one of those computers, 00:03:37.520 |
everything we see and perceive and think and feel 00:03:51.000 |
as something similar to the computers of today, 00:03:54.200 |
these deterministic sort of Turing machine type things? 00:04:06.760 |
something much bigger, something much more complicated, 00:04:09.160 |
something much more mysterious from our current perspective? 00:04:18.760 |
I don't think anything else would be required. 00:04:24.360 |
maybe they have quantum computers and other things 00:04:30.600 |
but I don't think it's a necessary assumption 00:04:40.760 |
would be able to create these kinds of computer simulations 00:05:06.040 |
I guess it's more informative and interesting 00:05:12.200 |
but it's not designed to be useful for something else. 00:05:29.160 |
Certainly it would be of interest in philosophy, 00:05:38.920 |
in the fundamental building blocks of the world 00:05:47.000 |
there is then the possibility that say physics 00:05:49.760 |
at the level where the computer running the simulation 00:05:52.960 |
could be different from the physics governing phenomena 00:06:08.680 |
So we mentioned the simulation hypothesis so far. 00:06:43.760 |
equals simulation hypothesis, just terminology wise. 00:06:51.600 |
So the hypothesis that we're living in a simulation, 00:06:53.600 |
simulation argument has these three complete possibilities 00:07:05.640 |
Although it doesn't on its own tell us which one. 00:07:10.640 |
So the first one is that almost all civilizations 00:07:15.240 |
at our current stage of technological development 00:07:17.800 |
go extinct before they reach technological maturity. 00:07:47.480 |
kill themselves too eagerly or destroy themselves too eagerly 00:07:50.600 |
or it might be super difficult to build a simulation. 00:07:58.320 |
Now, I think it looks like we would technologically 00:08:04.360 |
that is short compared to, say, the lifetime of planets 00:08:13.680 |
- So your intuition is to build a simulation is not... 00:08:25.040 |
We can see, even based on our current limited understanding, 00:08:29.240 |
what some lower bound would be on the capabilities 00:08:32.680 |
that you could realize by just developing technologies 00:08:38.480 |
So for example, one of my research fellows here, 00:08:48.320 |
That is, you could analyze using theoretical tools 00:09:04.960 |
if we could put these atoms together in this way, 00:09:12.200 |
and have these computational characteristics. 00:09:20.840 |
of molecularly manufacturing in the fullness of time. 00:09:35.880 |
We have an upper limit, which is the speed of light. 00:09:42.120 |
We know we can go faster than that by just, you know, 00:09:45.840 |
making them bigger and have more fuel and stuff. 00:09:52.080 |
the technological affordances that would exist 00:09:54.880 |
once a civilization has had enough time to develop, 00:09:58.400 |
at least those technologies we already know are possible. 00:10:07.800 |
But at least there is this kind of basic set of capabilities. 00:10:17.560 |
to deep space exploration to mature technology? 00:10:35.600 |
and that a civilization that was allowed to continue 00:10:45.240 |
the set of breakthroughs that are likely to happen. 00:10:55.640 |
The one is we could just imagine bigger computers 00:11:00.000 |
So you can kind of scale things that way, right? 00:11:02.680 |
But you could also make processors a bit faster. 00:11:37.960 |
You could have this kind of nanomolecular ability 00:11:58.200 |
that a technologically mature civilization would have. 00:12:26.080 |
What is a technologically mature civilization? 00:12:27.480 |
- Well, okay, so that means it's a stronger concept 00:12:29.080 |
than we really need for the simulation hypothesis. 00:12:31.160 |
I just think it's interesting in its own right. 00:12:40.880 |
that you developed all those general purpose, 00:12:43.920 |
widely useful technologies that could be developed, 00:12:55.120 |
You can think either that there is such a ceiling, 00:13:09.880 |
- So new things won't keep springing up, new ceilings. 00:13:13.800 |
- In terms of basic technological capabilities, 00:13:16.320 |
I think that, yeah, there is like a finite set of those 00:13:22.080 |
Moreover, I mean, I wouldn't be that surprised 00:13:30.080 |
fairly shortly after we have, say, machine superintelligence. 00:13:33.880 |
So I don't think it would take millions of years 00:13:36.360 |
for a human originating civilization to begin to do this. 00:13:51.320 |
I mean, for the purpose of the simulation argument, 00:13:56.080 |
indefinitely far up or whether there's a ceiling, 00:13:58.280 |
as long as we know we can at least get to a certain level. 00:14:01.600 |
And it also doesn't matter whether that's gonna happen 00:14:04.400 |
in 100 years or 5,000 years or 50 million years. 00:14:08.360 |
Like the timescales really don't make any difference 00:14:13.200 |
Like there's a big difference between 100 years 00:14:44.760 |
- Yeah, well, so it would shift around the probabilities 00:14:54.680 |
if it's like say the billions of years into the future, 00:14:56.880 |
then it's more likely that we will fail ever to get there. 00:15:06.280 |
- So it is important to think about how hard it is 00:15:09.880 |
- In terms of, yeah, figuring out which of the disjuncts. 00:15:16.320 |
which is agnostic as to which of these three alternatives 00:15:24.800 |
whether or not we thought this could be done in 500 years 00:15:31.800 |
I mean, I'm sure there might be some people who oppose it, 00:15:36.120 |
I mean, it's very nice those three cases cover it. 00:15:54.200 |
but let's actually, I don't think we went through them. 00:16:02.440 |
but we have to think not just what might destroy us. 00:16:09.560 |
whatever disasters or meteorites slamming the earth 00:16:13.400 |
a few years from now that could destroy us, right? 00:16:23.320 |
that almost all civilizations throughout the cosmos 00:16:32.040 |
- And the underlying assumption there is that there is 00:16:41.880 |
then they would virtually all have to succumb 00:16:48.640 |
I guess there are a lot of little digressions 00:16:50.880 |
- Definitely, let's go there, let's go there. 00:16:55.040 |
there is a set of basic questions that always come up 00:17:03.600 |
you could almost define whether a person is interesting, 00:17:14.480 |
it looks to me that the universe is very big. 00:17:17.920 |
I mean, in fact, according to the most popular 00:17:21.480 |
current cosmological theories, infinitely big. 00:17:28.360 |
that it would contain a lot of other civilizations, 00:17:32.600 |
If you have some local stochasticity and infinitely many, 00:17:37.800 |
it's like, you know, infinitely many lumps of matter, 00:17:54.200 |
Even short of that, if the universe is very big, 00:18:08.000 |
then all of civilizations at our current stage 00:18:17.440 |
that a very high fraction of those went extinct. 00:18:21.240 |
But if we think there are many, I mean, it's interesting 00:18:24.520 |
because there are certain things that plausibly 00:18:26.920 |
could kill us, like if you look at existential risks. 00:18:32.320 |
And it might be a different, like the best answer 00:18:43.640 |
if there is something that kills almost everyone, 00:18:47.720 |
'Cause that would have to be some risk factor 00:18:49.880 |
that was kind of uniform overall possible civilization. 00:18:53.520 |
- Yeah, so in this, for the sake of this argument, 00:19:02.920 |
- Yeah, or something very close to everybody. 00:19:10.080 |
- Well, so number two is the convergence hypothesis 00:19:21.320 |
they all lose interest in creating these simulations. 00:19:26.320 |
So they just, they have the capability of doing it, 00:19:39.360 |
maybe not even a single one of them would do it. 00:19:46.920 |
because it's like they get bored or whatever, 00:19:49.400 |
but it could be so many possibilities within that. 00:20:10.080 |
to ethical concerns, all those kinds of things 00:20:15.200 |
- Well, certainly, I mean, yeah, ethical concerns. 00:20:21.160 |
I mean, in a sense, that's the first assumption 00:20:27.920 |
using only a tiny fraction of your resources, 00:20:34.400 |
So it wouldn't be the case that they would need 00:20:42.880 |
about whether they should invest half of their GDP for this. 00:20:47.160 |
well, if any little fraction of the civilization 00:20:51.680 |
during maybe their millions of years of existence, 00:21:00.200 |
But certainly, there could be many conceivable reasons 00:21:07.280 |
many possible reasons for not running ancestor simulations 00:21:17.080 |
- Well, that would be the type of computer simulation 00:21:24.600 |
we think have lived on our planet in the past 00:21:27.560 |
and like ourselves in terms of the types of experiences 00:21:30.120 |
that they have and where those simulated people are conscious. 00:21:36.440 |
that a non-player character would be simulated 00:21:49.600 |
but something where the simulated being has a brain, 00:21:57.960 |
that it would have the same subjective experiences 00:22:20.960 |
Is it enough to simulate just the brain, just the minds, 00:22:24.660 |
and not the simulation, not the big universe itself? 00:22:27.560 |
Like is there different ways to think about this? 00:22:48.280 |
is conscious or not is not like whether it's built 00:23:04.480 |
or as far as my view, that it would be sufficient, say, 00:23:17.400 |
So if you had a simulation with 100 billion neurons 00:23:19.800 |
connected in the same way as the human brain, 00:23:24.400 |
with the same kind of synaptic weights and so forth, 00:23:27.560 |
so you actually had the same behavior coming out of this 00:23:35.840 |
Now, it's possible you could also generate consciousness 00:23:47.000 |
exactly how much you could simplify or abstract away. 00:23:53.680 |
I missed where you're placing consciousness in the second. 00:24:14.400 |
such that when they are run, consciousness emerges? 00:24:24.160 |
you think maybe that's not gonna have any consciousness. 00:24:43.360 |
there is self-representations of other cognitive processes 00:24:56.520 |
But exactly how much less than the full computation 00:25:01.200 |
that the human brain is performing would be required 00:25:04.560 |
is a little bit, I think, of an open question. 00:25:07.340 |
You asked another interesting question as well, 00:25:12.560 |
which is, would it be sufficient to just have, 00:25:17.720 |
say, the brain or would you need the environment 00:25:20.760 |
in order to generate the same kind of experiences 00:25:34.080 |
one thing that's clear is that we don't have to simulate 00:25:42.440 |
to have the perception that there is a full reality in there. 00:25:49.920 |
when it's actually within the view of the player character. 00:26:03.800 |
it might be that all of the parts that come into our view 00:26:10.240 |
And a lot of aspects that never come into view, 00:26:13.600 |
say, the details of this microphone I'm talking into, 00:26:17.320 |
exactly what each atom is doing at any given point in time, 00:26:42.880 |
but to construct a reality that is sufficiently real to us 00:26:46.960 |
to be immersive in the way that the physical world is. 00:26:52.120 |
I think that's actually probably an answerable question 00:26:59.440 |
of where's the line where it becomes so immersive 00:27:08.240 |
- Yeah, or that you don't realize while you're in it 00:27:20.600 |
what's interesting is it doesn't have to be real, 00:27:29.800 |
- Yeah, I mean, I think that might be too low a bar. 00:27:32.520 |
I mean, if you think, say, when people first had Pong 00:27:38.160 |
who wanted to keep playing it for a long time 00:27:48.160 |
but an absorbing activity doesn't even have to be. 00:28:04.400 |
or whatever more sophisticated games for hours, 00:28:09.680 |
while the World of Warcraft could be in a big addiction, 00:28:18.400 |
you would choose to spend your entire life in there. 00:28:21.040 |
- And then thereby changing the concept of what reality is 00:28:35.400 |
People might have different preferences regarding that. 00:28:38.480 |
Some might, even if you had a perfect virtual reality, 00:28:49.120 |
I mean, in philosophy, there's this experience machine, 00:28:58.840 |
where you imagine some crazy, super duper neuroscientists 00:29:11.120 |
you can kind of pre-programmed it in different ways. 00:29:26.080 |
like have a wonderful life, all of these things. 00:29:36.240 |
But you would kind of disconnect from the rest of reality 00:29:44.600 |
would choose not to enter the experience machine. 00:29:49.000 |
I mean, many might wanna go there for a holiday, 00:30:11.880 |
So therefore what we value depends on other things 00:30:18.440 |
So, okay, can you take that argument further? 00:30:21.280 |
I mean, what about the fact that maybe what we value 00:30:25.040 |
So you could have up and downs in the experience machine. 00:30:27.840 |
But what can't you have in the experience machine? 00:30:35.320 |
but for example, real connection with other people, 00:30:42.000 |
like that's something you wouldn't have there. 00:30:51.840 |
that wouldn't be anybody there who actually got happy. 00:30:54.360 |
It would just be a little simulation of somebody smiling, 00:30:58.360 |
but the simulation would not be the kind of simulation 00:31:38.560 |
these different possible explanations for why it is 00:31:43.120 |
you wouldn't wanna go into the experience machine 00:31:52.680 |
and the conclusions he wanted to draw from it 00:31:54.400 |
is how much is a kind of a status quo effect. 00:32:01.760 |
on current reality to plug into this dream machine. 00:32:09.920 |
well, what you've experienced up to this point 00:32:13.040 |
was a dream, now, do you wanna disconnect from this 00:32:20.440 |
when you have no idea maybe what the real world is? 00:32:42.640 |
If the status quo was that they were actually 00:32:57.880 |
- That's interesting, the change itself, the leap. 00:33:21.440 |
or whether inside a computer, outside a computer, 00:33:36.600 |
but I find myself quite willing to take the leap 00:33:48.440 |
And I think more people would take that leap. 00:34:01.080 |
like before this whole experience machine started. 00:34:04.440 |
- Well, I kind of assumed from that description, 00:34:16.920 |
about your kind of distinction between real and illusion. 00:34:21.920 |
Because when you can have an illusion that feels, 00:34:28.720 |
I mean, what, I don't know how you can definitively say 00:34:33.760 |
Like what's a good way to prove that something is real 00:34:43.880 |
with these wires by the super duper neuroscientists 00:34:58.200 |
by other people all around the world who buy the exports. 00:35:01.880 |
- So these are two different possible situations 00:35:11.040 |
when you're in a vat with wires and the neuroscientists, 00:35:22.400 |
you could have the experience of farming in Peru, 00:35:24.440 |
but that wouldn't actually be any peanuts grown. 00:35:37.280 |
And why can't all of that be done in a simulation? 00:35:42.480 |
that they actually have peanut farms in Peru. 00:35:45.200 |
I guess we'll get a lot of comments otherwise from Angry. 00:35:52.800 |
- You should know you can't grow peanuts in that climate. 00:35:57.800 |
- No, I mean, I think, I mean, in the simulation, 00:36:02.600 |
I think there is a sense, the important sense, 00:36:09.840 |
between inside a simulation and outside a simulation, 00:36:12.720 |
or in the case of NOSIG's thought experiment, 00:36:16.160 |
whether you're in the vat or outside the vat. 00:36:19.400 |
And some of those differences may or may not be important. 00:36:22.480 |
I mean, that comes down to your values and preferences. 00:36:29.800 |
only gives you the experience of growing peanuts, 00:36:32.760 |
but you're the only one in the experience machines. 00:36:37.920 |
within the experience machine, others can plug in. 00:36:41.280 |
- Well, there are versions of the experience machine. 00:36:48.320 |
So in like in the original thought experiment, 00:36:52.480 |
So, and you think, I wouldn't wanna go in there. 00:36:56.160 |
about what you value and what you care about. 00:36:58.040 |
Then you could say, well, what if you add the fact 00:37:03.280 |
Well, it starts to make it more attractive, right? 00:37:07.720 |
what if you could also have important long-term effects 00:37:17.640 |
Like you could actually have a life that had a purpose 00:37:25.360 |
it becomes more similar to the baseline reality 00:37:32.840 |
- Yeah, but I just think inside the experience machine 00:37:35.120 |
and without taking those steps that you just mentioned, 00:37:39.240 |
you still have an impact on long-term history 00:37:47.680 |
of the quote unquote fake creatures that live inside 00:38:17.400 |
she starts to have more of an independent existence. 00:38:21.320 |
But it depends, I think, on how she's implemented 00:38:29.360 |
is a static picture on the wall, a photograph. 00:38:33.160 |
So you think, well, I can look at her, right? 00:38:37.960 |
but then you think, well, it doesn't really matter much 00:39:00.640 |
with the same behavioral repertoire as you have, 00:39:06.160 |
she would be a conscious person like you are. 00:39:08.840 |
And then you would, what you do in this experience machine 00:39:18.880 |
which of these experience machines you're talking about. 00:39:23.520 |
that it would be possible to have an experience machine 00:39:28.000 |
that gave you a normal set of human experiences, 00:39:31.320 |
which include experiences of interacting with other people, 00:39:49.120 |
not just when you say hello, they say hello back, 00:40:02.120 |
instantiated this other person in enough detail 00:40:05.240 |
that you would have a second consciousness there. 00:40:07.720 |
I think that's to some extent an open question. 00:40:18.240 |
if you have a very limited interaction with somebody, 00:40:47.480 |
ask open-ended questions and probe from different angles, 00:40:54.040 |
to all of the possible ways that you could probe it, 00:41:02.240 |
in such a way that you would get the right answer 00:41:08.440 |
where you also instantiated a conscious mind. 00:41:10.640 |
- Yeah, I'm with you on the intelligence part, 00:41:15.800 |
Like I've recently gotten my hands on a lot of Roombas. 00:41:24.360 |
there's just a nice robotic mobile platform for experiments. 00:41:28.360 |
And I made them scream or moan in pain and so on 00:41:35.520 |
And it's just a sort of psychological experiment on myself. 00:41:38.920 |
And I think they appear conscious to me pretty quickly. 00:41:42.960 |
To me, at least my brain can be tricked quite easily. 00:41:59.560 |
and having certain qualities of the interaction, 00:42:03.360 |
like being able to suffer, like being able to hurt, 00:42:13.720 |
creating the illusion that you're wandering about it 00:42:17.480 |
is enough to create the feeling of consciousness 00:42:23.040 |
And because of that, create a really immersive experience 00:42:26.080 |
to where you feel like that is the real world. 00:42:29.280 |
between appearing conscious and being conscious? 00:42:33.120 |
Or is it that you think it's very easy to be conscious? 00:42:36.120 |
- I'm not actually sure what it means to be conscious. 00:42:37.960 |
All I'm saying is the illusion of consciousness 00:42:47.600 |
that's as good as if the thing was conscious. 00:42:53.440 |
I mean, I guess there are a few different things. 00:42:56.440 |
which might, I mean, if you don't really care about, 00:42:59.600 |
like probing hard for whether the thing is conscious, 00:43:02.280 |
maybe it would be a satisfactory interaction. 00:43:07.800 |
Whether or not you really thought it was conscious. 00:43:10.360 |
Now, if you really care about it being conscious 00:43:30.360 |
Like it's much more widely spread in the world 00:43:33.160 |
and we have thought it doesn't require a big human brain 00:43:43.160 |
Like in that case, I guess you still have a close coupling. 00:43:46.160 |
That I guess a data case would be where they can come apart, 00:43:56.080 |
with actually not being another conscious mind. 00:44:03.360 |
I think one observation that makes it plausible, 00:44:06.680 |
that you could have very realistic appearances 00:44:13.440 |
which also is relevant for the simulation argument. 00:44:32.080 |
during the wee hours of the night when we are dreaming. 00:44:38.680 |
but often you also don't realize that you're in a dream. 00:44:46.640 |
three pound lumps of neural matter effortlessly. 00:44:53.560 |
the virtual reality that seems pretty real to us, 00:45:02.520 |
with planetary sized computers optimized over the eons 00:45:05.920 |
to create a realistic environment for you to interact with? 00:45:21.120 |
It's also possible that the brain is the epitome, 00:45:30.640 |
- Meaning like this is the smartest possible thing 00:45:39.960 |
I mean, for some of these reasons we alluded to earlier 00:45:43.920 |
in terms of designs we already have for computers 00:45:48.920 |
that would be faster by many orders of magnitude 00:45:55.120 |
- Yeah, but it could be that the constraints, 00:46:05.360 |
the less likely it is to become super intelligent. 00:46:07.920 |
This is where I say dumb things to push back on. 00:46:14.160 |
No, I mean, so there are different dimensions 00:46:20.000 |
Like if you could solve the same challenge faster 00:46:25.240 |
So there, I think we have very strong evidence 00:46:34.680 |
than the human brain and therefore have speed 00:46:37.880 |
super intelligence, like be completely superior, 00:46:48.520 |
And there, the concepts are a little bit less clear cut. 00:46:56.160 |
firmly logical argument for why that could be 00:47:00.200 |
qualitative super intelligence as opposed to just things 00:47:14.520 |
like there seems to be like Einstein versus random person. 00:47:18.520 |
Like it's not just that Einstein was a little bit faster, 00:47:21.920 |
but like how long would it take a normal person 00:47:26.480 |
It's like, it's not 20% longer than it took Einstein 00:47:30.040 |
It's like, I don't know whether they would do it at all 00:47:31.840 |
or it would take millions of years or some totally bizarre. 00:47:36.840 |
- But your intuition is that the compute size 00:47:39.080 |
will get you, increasing the size of the computer 00:47:45.520 |
might create some much more powerful levels of intelligence 00:47:51.160 |
we've been talking about with like the simulation, 00:47:53.320 |
being able to simulate an ultra realistic environment, 00:48:01.280 |
- Yeah, strictly speaking, it would not be necessary 00:48:04.160 |
to have super intelligence in order to have say, 00:48:09.160 |
ancestor simulations or other kinds of simulations. 00:48:11.720 |
As a matter of fact, I think if we are in a simulation, 00:48:19.160 |
it would most likely be one built by a civilization 00:48:28.960 |
larger scale structures if you had super intelligence. 00:48:39.360 |
So I'd expect by the time they could make these 00:48:45.360 |
with human brains in there, like before that, 00:48:47.680 |
they got to that stage, they would have figured out 00:48:52.520 |
or maybe biological enhancements of their own brains 00:48:56.120 |
if they were biological creatures to start with. 00:49:04.360 |
One, we destroy ourselves before we ever create 00:49:07.520 |
Two, we somehow, everybody somehow loses interest 00:49:16.080 |
So you've kind of, I don't know if your thinking 00:49:20.120 |
has evolved on this point, but you kind of said 00:49:22.040 |
that we know so little that these three cases 00:49:31.720 |
- Yeah, I mean, I don't think equal necessarily 00:49:34.880 |
would be the most supported probability assignment. 00:49:39.880 |
- So how would you, without assigning actual numbers, 00:49:48.000 |
- Well, I mean, I've historically tended to punt 00:49:50.120 |
on the question of like as between these three. 00:49:55.120 |
- So maybe you ask another way is which kind of things 00:49:59.600 |
would make each of these more or less likely? 00:50:04.960 |
- Certainly in general terms, if you take anything 00:50:07.480 |
that say increases or reduces the probability 00:50:10.960 |
of one of these, we tend to slosh probability around 00:50:17.280 |
So if one becomes less probable, like the other 00:50:19.160 |
would have to, 'cause it's gotta add up to one. 00:50:25.080 |
the first alternative that there's this filter 00:50:29.040 |
that makes it so that virtually no civilization 00:50:39.960 |
And if that's true, then it's like very unlikely 00:50:44.280 |
just because if almost no civilization at our stage does it, 00:50:49.880 |
- Sorry, can you linger on that for a second? 00:50:51.240 |
- Well, if it's the case that almost all civilizations 00:50:53.920 |
at our current stage of technological development 00:51:07.760 |
- And also sort of the flip side of that is the fact 00:51:10.280 |
that we've reached it means that many other civilizations 00:51:14.040 |
- Yeah, so that means if we get closer and closer 00:51:18.280 |
there's less and less distance left where we could 00:51:25.040 |
And therefore the probability that we will reach 00:51:30.880 |
And that would make it less likely to be true 00:51:32.840 |
that almost all civilizations at our current stage 00:51:37.080 |
Like we would have this, the one case we'd started ourselves 00:51:42.680 |
That would be strong evidence that it's not so hard 00:51:46.320 |
So to the extent that we feel we are moving nearer 00:51:50.520 |
to technical maturity, that would tend to reduce 00:51:56.280 |
and increase the probability of the other two. 00:52:01.800 |
Like if every once in a while some new threat 00:52:05.040 |
comes into view, some bad new thing you could do 00:52:11.160 |
that could change our probabilities in the other direction. 00:52:14.880 |
- But that technology, again, you have to think about 00:52:21.640 |
in an even way affect every civilization out there. 00:52:30.720 |
I mean, that could be two different existential risks 00:52:36.160 |
- As in one of them. - Not from one or the other. 00:52:53.800 |
and how we must make sure to handle that wisely 00:52:59.480 |
It's not the right kind of existential catastrophe 00:53:13.760 |
as a result of it being shaped by some process 00:53:17.560 |
that optimized for some completely non-human value. 00:53:20.880 |
But even if we got killed by machine superintelligence 00:53:33.400 |
This could be any intelligent species that achieves, 00:53:36.840 |
like it's all about the technological maturity. 00:53:44.800 |
'cause it replaced us and that's just as well 00:53:46.440 |
for the simulation argument. - And that could still, 00:53:48.480 |
it could interact with the second alternative. 00:54:07.840 |
we won't produce any simulations 'cause we are dead. 00:54:19.800 |
if might not that thing then use some of its resources 00:54:27.560 |
This is given how little we know about the universe. 00:54:30.600 |
Is it reasonable to reason about these probabilities? 00:54:35.600 |
So like how little, well, maybe you can disagree, 00:54:49.520 |
We've also don't know, like as we try to start building it, 00:54:54.240 |
like start creating virtual worlds and so on, 00:55:01.560 |
that can fundamentally change just so many aspects 00:55:21.760 |
have a theory and everything, how that changes the stuff, 00:55:23.840 |
how that changes deep space exploration and so on. 00:55:27.480 |
So like, is it still possible to reason about probabilities 00:55:32.600 |
- Yes, I think though there will be a large residual 00:55:37.800 |
of uncertainty that we'll just have to acknowledge. 00:55:41.800 |
And I think that's true for most of these big picture 00:55:49.680 |
It's just, we are small, short-lived, small-brained, 00:55:54.680 |
cognitively very limited humans with little evidence. 00:55:59.000 |
And it's amazing we can figure out as much as we can 00:56:08.960 |
that seems to happen when I look at the simulation argument, 00:56:11.840 |
which for me, it seems like case one and two feel unlikely. 00:56:19.440 |
sort of, it's not like I have too much scientific evidence 00:56:26.920 |
It just seems unlikely that every single civilization 00:56:37.000 |
So naturally, without necessarily explicitly doing it, 00:56:45.600 |
it's very likely we're living in a simulation. 00:56:51.800 |
I think the mind goes there for a lot of people. 00:57:00.860 |
which has to do with the motivations and interest 00:57:11.000 |
I think there is much we don't understand about that. 00:57:15.600 |
- Yeah, can you talk about that a little bit? 00:57:27.840 |
Do you think it'll fundamentally transform our motivations? 00:57:33.160 |
that once you take this leap to technological maturity, 00:57:45.040 |
that would be sort of on the path for basically 00:57:50.880 |
to create large numbers of ancestry simulations. 00:58:00.680 |
of what a civilization is in fact optimizing for. 00:58:20.080 |
our own experiences are not under our direct control. 00:58:23.660 |
So for example, if you want to experience a pleasure 00:58:30.140 |
and happiness, you might have to do a whole host of things 00:58:35.800 |
in the external world to try to get into the stage, 00:58:39.260 |
into the mental state where you experience pleasure. 00:58:42.320 |
You're like, like some people get some pleasure 00:58:47.040 |
They have to kind of actually go to a nice restaurant 00:58:53.100 |
that maybe arises from the fact that we are trying 00:59:02.000 |
But the only way to do that is by a whole host 00:59:04.280 |
of complicated activities in the external world. 00:59:06.920 |
Now, at some level of technological development, 00:59:09.320 |
I think we'll become auto potent in the sense 00:59:15.440 |
our own internal configuration and enough knowledge 00:59:22.680 |
So then it could turn out that there are a lot 00:59:24.880 |
of instrumental goals that would drop out of the picture 00:59:31.480 |
because we could now serve some of these final goals 00:59:41.240 |
after civilizations reflect on that and converge 00:59:45.640 |
and different attractors and so on and so forth. 00:59:48.040 |
And that could be new instrumental considerations 00:59:54.520 |
that come into view as well that we are just oblivious to 00:59:57.800 |
that would maybe have a strong shaping effect on actions, 01:00:08.280 |
because we are so dumb, fumbling through the universe. 01:00:11.000 |
But if almost inevitably on route to attaining the ability 01:00:17.480 |
you do have this cognitive enhancement or advice 01:00:26.320 |
And you have to realize it's obvious that the thing 01:00:30.680 |
Whereas right now it seems, hey, you could X, Y or Z 01:00:32.960 |
and different people will do different things. 01:00:38.920 |
- Yeah, because at this time with our limited technology, 01:00:45.200 |
I mean, that's starting to change in some ways, but- 01:00:49.360 |
- Well, I'm not sure it follows that the impact 01:00:55.640 |
I mean, I suppose a hundred years ago it was minor. 01:01:14.360 |
- Yeah, so it might be that the greatest impact 01:01:18.080 |
of individuals is not at technological maturity 01:01:22.840 |
It might be earlier on when there are different tracks, 01:01:35.120 |
that those could be bigger than the direct effects 01:01:51.820 |
That I think you had a conversation with Joe Rogan. 01:02:01.040 |
How does that lead to where likely living in a simulation? 01:02:12.600 |
why does that mean that we're now in a simulation? 01:02:15.600 |
- What you get to if you accept alternative three first 01:02:23.320 |
with our kinds of experiences than non simulated ones. 01:02:31.320 |
by the end of time as it were, you just count it up. 01:02:34.960 |
That would be more simulated ones than non simulated ones. 01:02:39.440 |
Then there is an extra step to get from that. 01:02:43.120 |
If you assume that suppose for the sake of the argument 01:02:49.320 |
to the statement we are probably in a simulation? 01:02:55.320 |
So here you're introducing an indexical statement 01:02:57.600 |
like it's that this person right now is in a simulation. 01:03:09.840 |
But what probability should you have that you yourself 01:03:18.240 |
So yeah, so I call it the bland principle of indifference, 01:03:25.920 |
when you have two, I guess, sets of observers, 01:03:33.920 |
And you can't from any internal evidence you have 01:03:40.720 |
You should assign a probability that's proportional 01:03:48.160 |
So that if there are 10,000 people in a simulation, 01:03:55.200 |
you would be 10 times more likely to be one of those. 01:04:04.800 |
you should rationally just assign the same probability 01:04:31.400 |
for a universe to produce an intelligent civilization 01:04:39.800 |
At some point, when the first simulation is created, 01:04:57.720 |
and then some subset of those get to intelligent life, 01:05:05.000 |
they might get started at quite different times. 01:05:07.360 |
Like maybe on some planet, it takes a billion years longer 01:05:10.360 |
before you get monkeys, or before you get even bacteria, 01:05:24.880 |
- Is there a connection here to the doomsday argument 01:05:32.080 |
they both involve an application of anthropic reasoning, 01:05:36.960 |
that is reasoning about these kind of indexical propositions. 01:05:55.040 |
And maybe you can speak to the anthropic reasoning 01:06:02.960 |
But the doomsday argument is this really first discovered 01:06:07.920 |
by Brandon Carter, who was a theoretical physicist 01:06:11.160 |
and then developed by philosopher John Leslie. 01:06:16.720 |
I think it might've been discovered initially 01:06:18.240 |
in the '70s or '80s, and Leslie wrote this book, 01:06:27.280 |
but let's focus on the Carter-Leslie version, 01:06:38.400 |
the probability that humanity will go extinct soon. 01:06:47.720 |
at the end of the day, there is something wrong 01:06:49.120 |
with this doomsday argument that it doesn't really hold. 01:06:53.480 |
but it's proved hard to say exactly what is wrong with it. 01:06:57.200 |
And different people have different accounts. 01:07:22.760 |
and they have balls in them that have numbers. 01:07:29.800 |
ball number one, two, three, up to ball number 10. 01:07:32.360 |
And then in the other urn, you have a million balls 01:07:40.360 |
And somebody puts one of these urns in front of you 01:07:44.240 |
and ask you to guess what's the chance it's the 10 ball urn. 01:07:49.000 |
And you say, well, 50-50, I can't tell which urn it is. 01:07:57.800 |
And let's suppose you find that it's ball number seven. 01:08:00.560 |
So that's strong evidence for the 10 ball hypothesis. 01:08:05.400 |
Like it's a lot more likely that you would get 01:08:16.560 |
it would be very unlikely you would get number seven. 01:08:22.520 |
And if your prior was 50-50 that it was the 10 ball urn, 01:08:33.200 |
So in the case of the urns, this is uncontroversial, 01:08:37.360 |
The Doomsday Argument says that you should reason 01:08:40.360 |
in a similar way with respect to different hypotheses 01:08:44.040 |
about how many balls there will be in the urn of humanity. 01:08:49.040 |
I said, for how many humans there will ever be 01:08:52.880 |
So to simplify, let's suppose we only consider 01:08:56.640 |
two hypotheses, either maybe 200 billion humans in total 01:09:09.280 |
So it's easiest to see if we just consider these two. 01:09:18.800 |
And maybe you say it's a 5% chance that we will go extinct 01:09:22.520 |
by the time there will have been 200 billion only. 01:09:30.000 |
But then according to this Doomsday Argument, 01:09:40.080 |
So your birth rank is your sequence in the position 01:09:47.680 |
It turns out you're about a human number of 100 billion, 01:10:01.160 |
I mean, obviously the exact number would depend 01:10:05.280 |
like which ancestors was human enough to count as human. 01:10:16.080 |
Now, if they're only gonna be 200 billion in total, 01:10:23.160 |
It's run-of-the-mill human, completely unsurprising. 01:10:47.560 |
you thought after finding this low-numbered random sample, 01:10:51.920 |
you updated in favor of the urn having few balls. 01:10:56.160 |
you should update in favor of the human species 01:11:06.800 |
Well, that would be the hypothesis in this case, 01:11:11.680 |
- I just like that term for that hypothesis, yeah. 01:11:23.840 |
from the set of all humans that will ever have existed. 01:11:30.960 |
The question then is why should you make that assumption? 01:11:40.360 |
with different ways of supporting that assumption. 01:11:43.680 |
- That's just one example of a theropic reasoning, right? 01:11:52.520 |
When you think about sort of even like existential threats 01:12:00.240 |
that you should assume that you're just an average case. 01:12:03.040 |
- Yeah, that you're a kind of a typical or randomly sample. 01:12:09.600 |
it seems to lead to what intuitively we think 01:12:15.920 |
that there's gotta be something fishy about this argument, 01:12:27.000 |
of reaching size 200 trillion humans in the future. 01:12:35.440 |
It seems you would need sophisticated arguments 01:12:37.480 |
about the impossibility of space colonization, blah, blah. 01:12:40.480 |
So one might be tempted to reject this key assumption, 01:12:46.520 |
as if you're a random sample from all observers 01:13:00.120 |
to make sense of bona fide scientific inferences. 01:13:14.920 |
So I mean, if you have a sufficiently large universe, 01:13:26.560 |
it could be true according to both of these theories 01:13:32.200 |
that there will be some observers observing the value 01:13:39.720 |
because there will be some observers that have hallucinations 01:13:49.200 |
And if enough observers make enough different observations, 01:14:20.600 |
is evidence for the theories we think are supported. 01:14:29.200 |
of these inferences that clearly seem correct 01:14:34.840 |
and infer what the temperature of the cosmic background is 01:14:39.000 |
and the fine structure constant and all of this. 01:14:42.600 |
But it seems that without rolling in some assumption 01:14:54.720 |
where it looks like this kind of anthropic reasoning 01:14:59.080 |
And yet in the case of the Doomsday Argument, 01:15:02.440 |
and people might think there's something wrong with it there. 01:15:05.680 |
So there's then this project that would consist 01:15:20.400 |
In other words, developing a theory of anthropics. 01:15:23.080 |
And there are different views of looking at that. 01:15:29.120 |
But to tie it back to the simulation argument, 01:15:39.440 |
is much weaker than the self-sampling assumption. 01:15:41.880 |
So if you think about in the case of the Doomsday Argument, 01:15:47.400 |
it says you should reason as if you're a random sample 01:15:59.680 |
Whereas in the case of the simulation argument, 01:16:01.520 |
it says that, well, if you actually have no way 01:16:05.720 |
then you should assign this kind of uniform probability. 01:16:12.840 |
in the simulation argument is different, it seems like. 01:16:17.360 |
I mean, I keep assigning the individual consciousness. 01:16:19.640 |
- Yeah, I mean, well, there are a lot of observers 01:16:33.320 |
So this would be the class of observers that we need. 01:16:40.160 |
So the question is, given that class of observers, 01:17:06.120 |
of the anthropic reasoning cases that we mentioned. 01:17:14.760 |
It's just like, are you simulated or are you not simulated? 01:17:18.160 |
Given this assumption that these two groups of people exist. 01:17:24.600 |
- Yeah, so the key point is the methodological assumption 01:17:28.840 |
you need to make to get the simulation argument 01:17:32.480 |
to where it wants to go is much weaker and less problematic 01:17:36.960 |
than the methodological assumption you need to make 01:17:39.520 |
to get the doomsday argument to its conclusion. 01:17:42.560 |
Maybe the doomsday argument is sound or unsound, 01:17:48.000 |
and more controversial assumption to make it go through. 01:17:57.640 |
to support this bland principle of indifference 01:18:05.400 |
where the fraction of people who are simulated 01:18:12.480 |
So in the limiting case where everybody is simulated, 01:18:23.840 |
If everybody with your experiences is simulated 01:18:32.400 |
You just kind of logically conclude it, right? 01:18:36.480 |
- So then as we move from a case where say 90% of people 01:18:46.480 |
it should seem plausible that the probability assigned 01:19:01.840 |
- Like you wouldn't expect that to be a discrete. 01:19:11.720 |
There are other arguments as well one can use 01:19:14.800 |
to support this blind principle of indifference, 01:19:19.480 |
- But in general, when you start from time equals zero 01:19:22.440 |
and go into the future, the fraction of simulated, 01:19:29.160 |
the fraction of simulated worlds will go to one. 01:19:40.720 |
although maybe a technologically mature civilization 01:19:50.040 |
It probably wouldn't be able to run infinitely many. 01:20:05.520 |
of information processing that any one civilization 01:20:15.680 |
- Well, first of all, there's a limited amount of matter 01:20:18.880 |
because with a positive cosmological constant, 01:20:37.600 |
when you perform an erasure of a computation, 01:20:41.360 |
just matter gradually over cosmological timescales, 01:20:50.480 |
like there's all kinds of seemingly unavoidable losses 01:21:04.320 |
- So it's finite, but of course, we don't know which, 01:21:14.680 |
couldn't there be like an arbitrary number of simulation 01:21:26.640 |
- Sorry, what do you mean that that could be? 01:21:37.440 |
it seems like each new spawn has fewer resources 01:21:42.840 |
But we don't know at which step along the way we are at. 01:21:49.080 |
Any one observer doesn't know whether we're in level 42 01:22:02.080 |
- I mean, it's true that there would be uncertainty as, 01:22:24.680 |
also have to be expanded at the level of the simulation. 01:22:37.760 |
like that has to power this whole tower, right? 01:22:39.880 |
So if there is a finite compute power in basement reality, 01:22:44.280 |
that would impose a limit to how tall this tower can be. 01:22:48.120 |
And if each level kind of imposes a large extra overhead, 01:22:53.040 |
you might think maybe the tower would not be very tall, 01:22:55.560 |
that most people would be low down in the tower. 01:23:13.420 |
just like you said, it comes up every once in a while, 01:23:17.880 |
But I would say one of the biggest popularizers 01:23:42.120 |
why to the extent that one think the argument is credible, 01:23:47.720 |
tell us something very important about the world 01:23:50.880 |
whichever of the three alternatives for a simulation 01:23:58.480 |
Now, interestingly, in the case of somebody like Elon, 01:24:02.240 |
for why you might wanna take the simulation hypothesis 01:24:07.360 |
In the case that if you are actually Elon Musk, let us say, 01:24:13.400 |
in that what are the chances you would be Elon Musk? 01:24:16.000 |
Like, it seems like maybe there would be more interest 01:24:31.760 |
or the whole of human civilization are simulated, 01:24:39.160 |
like in those simulations that only include a subset, 01:24:43.680 |
it might be more likely that that would include subsets 01:24:45.920 |
of people with unusually interesting or consequential life. 01:24:56.760 |
or you're like some particularly like distinctive character, 01:25:04.080 |
I mean, if you just think of yourself into the shoes, right? 01:25:12.320 |
So on a scale of like farmer in Peru to Elon Musk, 01:25:20.600 |
- You'd imagine that would be some extra boost from that. 01:25:26.080 |
So he also asked the question of what he would ask an AGI 01:25:30.000 |
saying the question being, what's outside the simulation? 01:25:34.520 |
Do you think about the answer to this question, 01:25:45.440 |
- Yeah, I mean, I think it connects to the question 01:25:50.280 |
if you had views about the creatures of the simulation, 01:25:59.160 |
what might happen, what happens after the simulation, 01:26:03.480 |
if there is some after, but also like the kind of setup. 01:26:06.600 |
So these two questions would be quite closely intertwined. 01:26:10.720 |
- But do you think it would be very surprising to, 01:26:17.880 |
is it possible for it to be fundamentally different 01:26:31.800 |
or any kind of information processing capabilities 01:26:34.760 |
enough to understand the mechanism that's created them? 01:26:45.600 |
it's kind of there are levels of explanation, 01:26:51.040 |
So does your dog understand what it is to be human? 01:26:59.760 |
And like a normal human would have a deeper understanding 01:27:10.280 |
or a great novelist might understand a little bit more 01:27:18.640 |
So similarly, I do think that we are quite limited 01:27:23.640 |
in our ability to understand all of the relevant aspects 01:27:41.800 |
that changes the significance of all the other aspects. 01:27:44.560 |
So we understand maybe seven out of 10 key insights 01:27:49.720 |
but the answer actually like varies completely 01:27:55.840 |
depending on what like number eight, nine, and 10 insight is. 01:28:12.000 |
And if it's even, the best thing for you to do in life 01:28:16.360 |
And if it's odd, the best thing for you is to go south. 01:28:19.120 |
Now we are in a situation where maybe through our science 01:28:29.800 |
but we are clueless about what the last three digits are. 01:28:38.240 |
and therefore whether we should go north or go south. 01:28:42.880 |
but I feel we're somewhat in that predicament. 01:28:48.000 |
We've come maybe more than half of the way there 01:28:52.560 |
but the parts we're missing are plausibly ones 01:28:55.320 |
that could completely change the overall upshot 01:28:59.160 |
of the thing, and including change our overall view 01:29:02.720 |
about what the scheme of priorities should be 01:29:05.280 |
or which strategic direction would make sense to pursue. 01:29:07.760 |
- Yeah, I think your analogy of us being the dog 01:29:11.160 |
trying to understand human beings is an entertaining one, 01:29:17.560 |
The closer the understanding tends from the dog's viewpoint 01:29:36.120 |
to analogize that a dog's understanding of a human being 01:29:42.320 |
of the fundamental laws of physics in the universe. 01:30:11.040 |
to learn from experience, to plan, to reason, 01:30:21.080 |
Is consciousness mixed up into that or is it-- 01:30:33.400 |
- Yeah, that would be like something that was much more, 01:30:41.600 |
So if we talk about general superintelligence, 01:30:49.640 |
make plans that are more effective at achieving its goals, 01:30:53.000 |
say in a wide range of complex, challenging environments. 01:30:56.880 |
- In terms of, as we turn our eye to the idea 01:31:00.040 |
of sort of existential threats from superintelligence, 01:31:07.400 |
in the physical world or can it be digital only? 01:31:10.680 |
Sort of, we think of our general intelligence as us humans, 01:31:15.120 |
as an intelligence that's associated with a body 01:31:20.080 |
that's able to affect the world directly with physically. 01:31:23.960 |
- I mean, digital only is perfectly fine, I think. 01:31:26.120 |
I mean, it's physical in the sense that obviously 01:31:32.040 |
But its capability to affect the world sort of- 01:31:40.240 |
If it can type text on the screen or something like that, 01:31:45.720 |
- So in terms of the concerns of existential threat of AI, 01:31:50.720 |
how can an AI system that's in the digital world 01:31:58.200 |
and what are the attack vectors for a digital system? 01:32:01.800 |
- Well, I mean, I guess maybe to take one step back, 01:32:10.120 |
from machine intelligence, including superintelligence. 01:32:13.280 |
And I wanna stress that because some of my writing 01:32:20.680 |
And when I wrote the book "Superintelligence," 01:32:23.040 |
at that point, I felt that there was a kind of neglect 01:32:33.520 |
a more granular understanding of where the pitfalls are 01:32:37.440 |
I think that since the book came out in 2014, 01:32:43.360 |
there has been a much wider recognition of that. 01:32:50.040 |
say, AI alignment techniques and so on and so forth. 01:32:52.640 |
So I'd like, yeah, I think now it's important 01:33:02.320 |
- And there's a little bit of a neglect now on the upside, 01:33:08.880 |
if you look at the amount of information there's available 01:33:13.720 |
about the positive possibilities of general intelligence, 01:33:28.880 |
- What are, can you link on that for a little bit? 01:33:30.920 |
What are some, to you, possible big positive impacts 01:33:42.800 |
these two different contexts of thinking about AI 01:33:45.920 |
and AI impacts, the kind of near-term and long-term, 01:33:49.200 |
if you want, both of which I think are legitimate things 01:33:53.000 |
to think about, and people should discuss both of them. 01:34:10.160 |
And so I think as long as we keep them apart, 01:34:18.560 |
- Can you clarify just the two things we were talking about, 01:34:27.960 |
but say the things I wrote about in this book, 01:34:39.920 |
or even things, self-driving cars and drones and stuff, 01:34:45.360 |
And then, of course, you could imagine some medium-term 01:35:19.240 |
- Yeah, I mean, I guess it's a little hard to articulate 01:35:24.560 |
has a lot of problems as it currently stands. 01:35:34.720 |
a friendly aligned superintelligence working on. 01:35:59.480 |
I think AI, especially artificial general intelligence 01:36:04.160 |
is really the ultimate general purpose technology. 01:36:09.600 |
this one area where it will have a big impact, 01:36:17.560 |
in all fields where human creativity and intelligence 01:36:39.520 |
if we just had more resources and cooler tech, 01:36:50.040 |
just by having more technological capability. 01:37:02.720 |
Now, again, that doesn't mean I'm like thinking, 01:37:14.000 |
but I mean, those are perfectly consistent views to hold 01:37:38.120 |
like for me sort of building machine learning systems today, 01:37:44.880 |
but there's some intuition of exponential growth, 01:37:47.120 |
of exponential improvement of intelligence explosion. 01:37:58.920 |
about the possibility of intelligence explosion, 01:38:16.160 |
it seems fairly likely to me that at some point 01:38:21.240 |
like some period of time where progress in AI 01:38:45.280 |
to be for it to be called an intelligence explosion? 01:38:48.920 |
Like, does it have to be like overnight, literally, 01:38:55.960 |
if you plotted the opinions of different people in the world, 01:39:00.000 |
I guess that would be somewhat more probability 01:39:05.360 |
than probably the average AI researcher, I guess. 01:39:09.480 |
- So, and then the other part of the intelligence explosion, 01:39:23.040 |
to proceed beyond it to get to superintelligence? 01:39:27.040 |
- Yeah, that seems, I mean, as much as any of these things 01:39:49.440 |
But it does seem, as far as I'm judging things 01:39:55.120 |
it seems very unlikely that there would be a ceiling 01:40:11.400 |
to create a system that's beyond our intelligence. 01:40:24.520 |
it feels like there's a line beyond which it steps, 01:40:31.040 |
and therefore it feels like a step where we lose control. 01:40:51.920 |
while still making sure that they are aligned, 01:40:54.800 |
like they are in the service of human values. 01:40:57.500 |
I mean, so losing control, I think, is not a given, 01:41:08.040 |
I mean, to some extent, I've lived with this for so long, 01:41:19.640 |
- And so I actually misspoke, I didn't mean control. 01:41:24.640 |
I meant, because the control problem is an interesting thing, 01:41:30.720 |
at least we should be able to maintain control 01:41:51.840 |
that humans aren't very good at dealing with. 01:41:55.720 |
I mean, I value my intelligence as a human being. 01:42:05.840 |
I mean, you don't see that as such a fundamentally-- 01:42:09.600 |
- I think yes, a lot, I think it would be small. 01:42:13.160 |
I mean, I think there are already a lot of things out there 01:42:18.160 |
the universe is big, there's gonna be other civilizations 01:42:22.880 |
or that just naturally have brains the size of beach balls, 01:42:26.720 |
and are like completely leaving us in the dust. 01:42:36.840 |
what would happen in a kind of post-human world, 01:42:41.720 |
like how much day to day would these super intelligences 01:42:52.520 |
where it would be more like a background thing 01:42:58.880 |
this intrusive kind of like making you feel bad 01:43:10.600 |
You don't wanna be the dumbest kid in your class, 01:43:12.600 |
everybody picks it, like a lot of those things 01:43:19.440 |
where we have infrastructure that is in some sense 01:43:27.240 |
I mean, it's a little bit like say the scientific community 01:43:33.280 |
but I mean, obviously it's gotta be like way more capacious 01:43:39.400 |
So in some sense, there is this mind like thing 01:43:42.520 |
already out there that's just vastly more intelligent 01:44:00.560 |
- I mean, there's already Google and Twitter and Facebook, 01:44:03.840 |
these recommender systems that are the basic fabric of our, 01:44:12.840 |
I mean, do you think of the collective intelligence 01:44:27.160 |
the kind of vagueness and indeterminacy of those concepts 01:44:32.760 |
starts to dominate how you would answer that question. 01:44:53.880 |
that you have a kind of string, a word string key, 01:45:00.320 |
but a vast set of other things it can't even do at all, 01:45:14.160 |
and then like radically subhuman in all other domains. 01:45:19.160 |
Same with a chess, like, or just a simple computer 01:45:22.320 |
that can multiply really large numbers, right? 01:45:24.280 |
So it's gonna have this like one spike of super intelligence 01:45:27.320 |
and then a kind of a zero level of capability 01:45:32.200 |
- Yeah, I don't necessarily think the generalness, 01:45:36.720 |
but I could sort of, it's a gray area and it's a feeling, 01:45:52.920 |
well, you could say, well, these are both just board games, 01:45:59.100 |
But there's something about the learning, the self play-- 01:46:09.680 |
In the same way, Google is much closer to Deep Blue 01:46:24.320 |
but being able to learn a lot without the constraints 01:46:27.880 |
of being heavily constrained by human interaction, 01:46:37.640 |
seems to be an important facet of general intelligence. 01:46:47.040 |
and you weren't specifically pre-programmed for, 01:46:58.880 |
And in fact, I mean, systems like alpha zero can learn, 01:47:05.800 |
in fact, probably beat Deep Blue in chess and so forth. 01:47:20.800 |
that have even more general purpose learning ability, 01:47:22.760 |
it might also trigger an even stronger intuition 01:47:24.720 |
that they are actually starting to get smart. 01:47:29.520 |
what do you think a utopia looks like with AGI systems? 01:47:39.400 |
where we're kind of really closely interlinked 01:47:43.600 |
Is it possibly where AGI systems replace us completely 01:47:48.080 |
while maintaining the values and the consciousness? 01:47:53.080 |
Is it something like it's a completely invisible fabric, 01:47:55.960 |
like you mentioned, a society where it's just AIDS 01:48:10.320 |
I don't have all the answers yet, or might never have, 01:48:15.000 |
but I think there are some different observations 01:48:19.320 |
One is if this scenario actually did come to pass, 01:48:23.600 |
it would open up this vast space of possible modes of being. 01:48:28.600 |
On one hand, material and resource constraints 01:48:36.120 |
So there would be a lot of, a big pie, let's say. 01:48:51.680 |
this much larger design space and option space 01:48:54.680 |
than we have ever had access to in human history. 01:49:04.360 |
a fairly fundamental rethink of what ultimately we value. 01:49:09.360 |
Like think things through more from first principles. 01:49:11.880 |
The context would be so different from the familiar 01:49:13.720 |
that we could have just take what we've always been doing, 01:49:16.240 |
and then like, oh, well, we have this cleaning robot 01:49:24.600 |
Like, I think we would have to go back to first principles. 01:49:29.000 |
go back to the first principles of what is the meaning 01:49:31.640 |
of life, what is happiness, what is fulfillment? 01:49:35.440 |
And then also connected to this large space of resources 01:49:43.200 |
and I think something we should aim for is to do well 01:49:53.960 |
That is, we wouldn't have to choose only one value system. 01:50:00.120 |
We wouldn't have to choose only one value criterion 01:50:05.120 |
and say, we're gonna do something that scores really high 01:50:20.280 |
and it's like a lot of pleasure, that's good, 01:50:27.520 |
I think to some significant, not unlimited sense, 01:50:32.640 |
it would be possible to do very well by many criteria. 01:50:36.160 |
Like maybe you could get like 98% of the best 01:50:41.160 |
according to several criteria at the same time, 01:50:43.960 |
given this great expansion of the option space. 01:51:00.160 |
there seems to be this always multiple parties 01:51:05.520 |
even though it might seem dysfunctional inside the moment, 01:51:11.200 |
seems to be beneficial for, I guess, a balance of power. 01:51:16.200 |
- So that's, yeah, not exactly what I have in mind, 01:51:23.560 |
but that if you had the chance to do something 01:51:27.920 |
that scored well on several different metrics, 01:51:38.040 |
which ones of these value systems are we gonna screw over? 01:51:42.200 |
let's first try to do very well by all of them. 01:51:44.440 |
Then it might be that you can't get 100% of all, 01:51:47.120 |
and you would have to then like have the hard conversation 01:51:56.200 |
But you're saying, maybe it's not such a bad trade-off. 01:52:02.320 |
in which at least some of the constraints would be removed. 01:52:08.800 |
- So there's probably still be trade-offs in the end. 01:52:11.960 |
we at least take advantage of this abundance. 01:52:15.960 |
So in terms of thinking about this, like, yeah, 01:52:19.280 |
one should think, I think in this kind of frame of mind 01:52:24.280 |
of generosity and inclusiveness to different value systems 01:52:33.560 |
And I think one could do something that would be very good 01:52:41.760 |
- We kind of talked about AGI fundamentally transforming 01:52:46.120 |
the value system of our existence, the meaning of life. 01:52:51.120 |
But today, what do you think is the meaning of life? 01:52:55.280 |
The silliest or perhaps the biggest question, 01:53:01.920 |
What makes, what gives your life fulfillment, 01:53:10.600 |
a bunch of different but related questions in there 01:53:29.680 |
of some television series while eating junk food. 01:53:33.320 |
but they wouldn't think it had a lot of meaning. 01:53:35.760 |
Whereas conversely, something that might be quite 01:53:38.440 |
loaded with meaning might not be very fun always. 01:53:47.720 |
And so these things can, I think, come apart, 01:53:57.200 |
when if you're thinking about these utopia questions 01:54:15.360 |
Make sure you can sort of clearly perceive each one of them. 01:54:18.680 |
And then you can think about how you can combine them. 01:54:22.040 |
- And just as you said, hopefully come up with a way 01:54:29.840 |
or get like a very high score on a wide range of them, 01:54:36.440 |
that are exactly opposed to one another, right? 01:54:39.240 |
But I think for many values, they are kind of opposed with, 01:54:43.440 |
if you place them within a certain dimensionality 01:54:47.400 |
of your space, like there are shapes that are kind of, 01:54:50.200 |
you can't untangle like in a given dimensionality, 01:54:56.640 |
then it might in many cases just be that they are easy 01:55:02.000 |
So we'll see how much space there is for that. 01:55:10.640 |
- I don't think there's a better way to end it, Nick. 01:55:30.600 |
And thank you to our presenting sponsor, Cash App. 01:55:35.440 |
by downloading Cash App and using code LexPodcast. 01:55:39.960 |
If you enjoy this podcast, subscribe on YouTube, 01:55:45.960 |
or simply connect with me on Twitter @LexFriedman. 01:55:49.320 |
And now let me leave you with some words from Nick Bostrom. 01:56:00.400 |
"There's no opportunity to learn from errors. 01:56:05.720 |
"limit damages, and learn from experience, is unworkable. 01:56:12.880 |
"This requires foresight to anticipate new types of threats 01:56:16.320 |
"and a willingness to take decisive, preventative action 01:56:19.600 |
"and to bear the costs, moral and economic, of such actions." 01:56:24.280 |
Thank you for listening and hope to see you next time.