back to indexHow to Build AGI? (Ilya Sutskever) | AI Podcast Clips
Chapters
0:0 Intro
0:48 Selfplay
2:0 Simulation
2:41 Reinforcement Learning
3:27 Simulation to Real World
5:40 Having a Body
6:40 Consciousness
7:24 Existence Proof
7:49 Test of Intelligence
9:9 Criticism
10:54 Judging Progress
11:20 Creating an AI System
11:40 Asking Questions
12:13 Stalins Story
13:6 Ideal World
15:31 George Washington
16:27 Aligning Values
00:00:11.580 |
but in general, what does it take, do you think? 00:00:16.480 |
but I think that deep learning plus maybe another 00:00:25.880 |
So like you've spoken about the powerful mechanism 00:00:31.760 |
sort of exploring the world in a competitive setting 00:00:36.960 |
against other entities that are similarly skilled as them 00:00:50.600 |
I think is going to be deep learning plus some ideas, 00:00:55.320 |
and I think self-play will be one of those ideas. 00:01:00.760 |
self-play has this amazing property that it can surprise us 00:01:18.560 |
I don't know if OpenAI had a release about multi-agent 00:01:22.800 |
where you had two little agents who were playing hide and seek 00:01:31.360 |
They all produce behaviors that we didn't expect. 00:01:39.040 |
that our systems don't exhibit routinely right now. 00:01:45.240 |
I like this direction because of its ability to surprise us. 00:01:48.720 |
And an AGI system would surprise us fundamentally. 00:01:51.520 |
- Yes, and to be precise, not just a random surprise, 00:01:54.840 |
but to find the surprising solution to a problem 00:02:00.320 |
Now, a lot of the self-play mechanisms have been used 00:02:03.840 |
in the game context, or at least in the simulation context. 00:02:17.000 |
How much faith, promise do you have in simulation 00:02:24.800 |
in the real world, whether it's the real world 00:02:37.840 |
It has certain strengths and certain weaknesses 00:02:44.720 |
That's true, but one of the criticisms of self-play, 00:02:53.000 |
one of the criticisms of reinforcement learning 00:03:11.040 |
and be able to learn in non-simulated environments? 00:03:13.680 |
Or do you think it's possible to also just simulate 00:03:17.300 |
in a photorealistic and physics realistic way 00:03:21.400 |
the real world in a way that we can solve real problems 00:03:31.960 |
and has been exhibited many times by many different groups. 00:03:38.920 |
Also, open AI in the summer has demonstrated a robot hand 00:03:52.960 |
- I wasn't aware that was trained in simulation. 00:04:05.080 |
And the policy that was learned in simulation 00:04:11.200 |
it could very quickly adapt to the physical world. 00:04:14.200 |
- So the kind of perturbations with the giraffe 00:04:28.400 |
but not the kind of perturbations we've had in the video. 00:04:32.920 |
it's never been trained with a stuffed giraffe. 00:04:37.280 |
- So in theory, these are novel perturbations. 00:04:46.240 |
That's a clean, small scale, but clean example 00:04:58.460 |
And the better the transfer capabilities are, 00:05:10.620 |
which you could then carry with you to the real world. 00:05:13.820 |
As humans do all the time when they play computer games. 00:05:17.240 |
- So let me ask sort of an embodied question, 00:05:23.880 |
Do you think AGI says that we need to have a body? 00:05:33.280 |
sort of fear of mortality, sort of self-preservation 00:05:36.960 |
in the physical space, which comes with having a body? 00:05:44.620 |
but I think it's very useful to have a body for sure, 00:05:49.180 |
you can learn things which cannot be learned without a body. 00:05:54.800 |
if you don't have a body, you could compensate for it 00:06:01.400 |
For example, there are many people who were born deaf 00:06:08.540 |
I'm thinking about Helen Keller specifically. 00:06:11.860 |
- So even if you're not able to physically interact 00:06:22.960 |
I'm not sure if it's connected to having a body or not, 00:06:28.140 |
and a more constrained version of that is self-awareness. 00:06:31.540 |
Do you think an AGI system should have consciousness? 00:06:34.840 |
We can't define, whatever the heck you think consciousness is. 00:06:58.100 |
from the representation that's stored within your networks? 00:07:05.420 |
you're able to represent more and more of the world. 00:07:07.340 |
- Well, I'd say I'd make the following argument, 00:07:14.100 |
and if you believe that artificial neural nets 00:07:19.860 |
then there should at least exist artificial neural nets 00:07:24.560 |
- You're leaning on that existence proof pretty heavily. 00:07:37.420 |
if there's not some magic in the brain that we're not, 00:07:41.100 |
I mean, I don't mean a non-materialistic magic, 00:07:43.940 |
but that the brain might be a lot more complicated 00:07:50.180 |
- If that's the case, then it should show up. 00:07:54.860 |
- We will find out that we can't continue to make progress. 00:08:00.460 |
but let me talk about another poorly defined concept 00:08:08.380 |
What do you think is a good test of intelligence for you? 00:08:11.940 |
Are you impressed by the test that Alan Turing formulated 00:08:28.300 |
There's a certain frontier of capabilities today. 00:08:33.540 |
And there exists things outside of that frontier. 00:08:39.260 |
For example, I would be impressed by a deep learning system 00:08:47.540 |
like machine translation or computer vision task 00:08:53.700 |
a human wouldn't make under any circumstances. 00:09:03.040 |
- Yeah, so right now they make mistakes in different, 00:09:05.180 |
they might be more accurate than human beings, 00:09:06.900 |
but they still, they make a different set of mistakes. 00:09:09.420 |
- So my, I would guess that a lot of the skepticism 00:09:16.080 |
is when they look at their mistakes and they say, 00:09:35.700 |
But I also just don't like that human instinct 00:09:43.420 |
when we criticize any group of creatures as the other. 00:09:53.740 |
is much smarter than human beings at many things. 00:10:05.220 |
- It's kind of hard to judge what depth means, 00:10:14.780 |
- Yes, the same is applied to autonomous vehicles. 00:10:18.060 |
The same is probably gonna continue being applied 00:10:29.740 |
is the search for one case where the system fails 00:10:37.260 |
And then many people writing articles about it. 00:10:40.940 |
And then broadly, the public generally gets convinced 00:10:46.860 |
And we pacify ourselves by thinking it's not intelligent 00:11:01.100 |
But I think this connects to the earlier point 00:11:08.180 |
- And you have a new robot demonstrating something. 00:11:13.020 |
And I think that people will start to be impressed 00:11:16.260 |
once AI starts to really move the needle on the GDP. 00:11:19.640 |
- So you're one of the people that might be able 00:11:22.300 |
to create an AGI system here, not you, but you and OpenAI. 00:11:29.300 |
and you get to spend sort of the evening with it, him, her, 00:11:45.980 |
And I would be amazed that it doesn't make mistakes 00:11:55.200 |
would they be factual or would they be personal, 00:12:18.060 |
that might be in the room where this happens. 00:12:21.500 |
So let me ask sort of a profound question about, 00:12:30.280 |
Been talking to a lot of people who are studying power. 00:12:33.440 |
Abraham Lincoln said, "Nearly all men can stand adversity, 00:12:37.960 |
"but if you want to test a man's character, give him power." 00:12:41.640 |
I would say the power of the 21st century, maybe the 22nd, 00:12:46.640 |
but hopefully the 21st, would be the creation 00:12:49.520 |
of an AGI system and the people who have control, 00:12:53.700 |
direct possession and control of the AGI system. 00:12:56.560 |
So what do you think, after spending that evening 00:13:36.380 |
for what the AGI that represents them should do, 00:13:39.420 |
and an AGI that represents them goes and does it. 00:13:41.700 |
I think a picture like that, I find very appealing. 00:13:47.540 |
you would have an AGI for a city, for a country, 00:13:54.300 |
take the democratic process to the next level. 00:14:09.380 |
as long as it's possible to press the reset button. 00:14:16.680 |
- So I think that it's definitely will be possible to build. 00:14:23.260 |
so the question that I really understand from you is, 00:14:29.580 |
humans people have control over the AI systems that they build? 00:14:44.340 |
so it's not that just they can't help it be controlled, 00:14:52.040 |
one of the objectives of their existence is to be controlled 00:15:06.320 |
They are excited to help the children and to feed them 00:15:27.380 |
and the drive will be to help humans flourish. 00:15:39.300 |
And between that moment and the Democratic board members 00:15:56.760 |
one of the big things he did is he relinquished power. 00:15:59.640 |
He, first of all, didn't want to be president. 00:16:09.340 |
Do you see yourself being able to relinquish control 00:16:16.540 |
given how much power you can have over the world? 00:16:19.500 |
At first financial, just make a lot of money, right? 00:16:22.980 |
And then control by having possession of this AGI system. 00:16:29.260 |
I'd find it trivial to relinquish this kind of power. 00:16:31.700 |
I mean, you know, the kind of scenario you are describing 00:16:39.220 |
I would absolutely not want to be in that position. 00:16:45.900 |
or the minority of people in the AI community? 00:16:50.980 |
- It's an open question and an important one. 00:16:54.020 |
Are most people good is another way to ask it. 00:17:09.540 |
Are there specific mechanism you can think of 00:17:16.940 |
of continued alignment as we develop the AI systems? 00:17:21.640 |
In some sense, the kind of question which you are asking is, 00:17:27.580 |
so if I were to translate the question to today's terms, 00:17:30.940 |
it would be a question about how to get an RL agent 00:17:36.880 |
that's optimizing a value function which itself is learned. 00:17:41.440 |
And if you look at humans, humans are like that 00:17:43.440 |
because the reward function, the value function of humans 00:17:59.360 |
and as objective as possible perception system 00:18:12.260 |
And then that component would then be integrated