back to indexDaniel Kahneman: Deep Learning (System 1 and System 2) | AI Podcast Clips
Chapters
0:0 Intro
1:20 System 1 advances
2:28 Humans learn quickly
3:32 Benefits of System 1
5:26 Current Architecture
6:10 Neural Networks
7:0 Grounding
7:24 What is Grounding
8:23 Active Learning
9:32 Building a System
10:23 Human Perception
12:3 Pedestrians
13:44 Understanding mortality
00:00:00.000 |
So we're not talking about humans, but if we think about building artificial intelligence 00:00:07.280 |
systems, robots, do you think all the features and bugs that you have highlighted in human 00:00:14.420 |
beings are useful for constructing AI systems? 00:00:18.540 |
So both systems are useful for perhaps instilling in robots? 00:00:23.480 |
What is happening these days is that actually what is happening in deep learning is more 00:00:33.960 |
like a system one product than like a system two product. 00:00:37.800 |
I mean, deep learning matches patterns and anticipate what's going to happen, so it's 00:00:46.160 |
But what deep learning doesn't have, and many people think that this is the critical, it 00:00:54.800 |
doesn't have the ability to reason, so there is no system two there. 00:00:59.840 |
But I think very importantly, it doesn't have any causality or any way to represent meaning 00:01:08.960 |
So until that is solved, what can be accomplished is marvelous and very exciting, but limited. 00:01:19.840 |
That's actually really nice to think of current advances in machine learning as essentially 00:01:29.940 |
If we think of deep learning and artificial intelligence systems? 00:01:34.360 |
It's very clear that deep mind has already gone way beyond what people thought was possible. 00:01:41.640 |
I think the thing that has impressed me most about the developments in AI is the speed. 00:01:48.080 |
It's that things, at least in the context of deep learning, and maybe this is about 00:01:53.520 |
to slow down, but things moved a lot faster than anticipated. 00:01:59.200 |
The transition from solving chess to solving Go, that's bewildering how quickly it went. 00:02:09.760 |
The move from AlphaGo to AlphaZero is sort of bewildering the speed at which they accomplished 00:02:16.840 |
Now, clearly, there are many problems that you can solve that way, but there are some 00:02:29.280 |
Well, reasoning and also, you know, one of the real mysteries, psychologist Gary Marcus, 00:02:37.480 |
who is also a critic of AI, I mean, what he points out, and I think he has a point, is 00:02:53.160 |
Humans don't need a million examples, they need two or three examples. 00:02:59.200 |
So clearly, there is a fundamental difference. 00:03:02.440 |
And what enables a machine to learn quickly, what you have to build into the machine, because 00:03:11.160 |
it's clear that you have to build some expectations or something in the machine to make it ready 00:03:16.860 |
to learn quickly, that at the moment seems to be unsolved. 00:03:23.320 |
I'm pretty sure that DeepMind is working on it, but if they have solved it, I haven't 00:03:32.400 |
They're trying to actually, them and OpenAI are trying to start to get to use neural networks 00:03:40.160 |
So assembled knowledge, of course, causality is, temporal causality is out of reach to 00:03:49.640 |
You mentioned the benefits of System 1 is essentially that it's fast, allows us to 00:04:00.460 |
You know, in a sense, I mean, there was the earlier phase of AI attempted to model reasoning, 00:04:10.680 |
and they were moderately successful, but, you know, reasoning by itself doesn't get 00:04:18.360 |
Deep learning has been much more successful in terms of, you know, what they can do. 00:04:22.920 |
But now, it's an interesting question, whether it's approaching its limits. 00:04:30.620 |
So I just talked to Gian Lacoon, you mentioned, you know, so he thinks that the limits, we're 00:04:38.800 |
not going to hit the limits with neural networks, that ultimately this kind of System 1 pattern 00:04:43.840 |
matching will start to start to look like System 2 without significant transformation 00:04:53.960 |
So I'm more with the majority of the people who think that yes, neural networks will hit 00:05:01.140 |
He, on the one hand, I have heard him tell the Mises-Sabies essentially that, you know, 00:05:07.860 |
what they have accomplished is not a big deal, that they have just touched, that basically, 00:05:13.220 |
you know, they can't do unsupervised learning in an effective way. 00:05:18.900 |
But you're telling me that he thinks that the current, within the current architecture, 00:05:26.600 |
So he's very much a pragmatist in a sense that's saying that we're very far away, that 00:05:31.520 |
there's still, I think there's this idea that he says is we can only see one or two mountain 00:05:39.120 |
peaks ahead and there might be either a few more after or thousands more after. 00:05:48.160 |
Right. But nevertheless, it doesn't see a, the final answer not fundamentally looking 00:05:59.340 |
So neural networks being a huge part of that. 00:06:03.600 |
I mean, that's very likely because, because pattern matching is so much of what's going 00:06:11.200 |
And you can think of neural networks as processing information sequentially. 00:06:14.200 |
Yeah, I mean, you know, there is, there is an important aspect to, for example, you get 00:06:22.360 |
systems that translate and they do a very good job, but they really don't know what 00:06:33.840 |
For that, you would need, you would need an AI that has sensation, an AI that is in touch 00:06:42.200 |
Yes, self-awareness and maybe even something that resembles consciousness kind of ideas. 00:06:48.880 |
Certainly awareness of, you know, awareness of what's going on so that the words have 00:06:54.160 |
meaning or can get, are in touch with some perception or some action. 00:07:00.920 |
So that's a big thing for Jan and as what he refers to as grounding to the physical 00:07:07.920 |
So, so that's what we're talking about the same. 00:07:13.440 |
I mean the grounding, without grounding, then you get, you get a machine that doesn't know 00:07:18.960 |
what it's talking about because it is talking about the world ultimately. 00:07:24.360 |
The question, the open question is what it means to ground. 00:07:26.800 |
I mean, we're very human centric in our thinking, but what does it mean for a machine to understand 00:07:38.800 |
Does it need to have a finiteness like we humans have? 00:07:42.360 |
All of these elements, it's a very, it's an open question. 00:07:46.360 |
You know, I'm not sure about having a body, but having a perceptual system, having a body 00:07:52.280 |
I mean, if, if you think about human mimicking human, but having a perception that seems 00:07:59.920 |
to be essential so that you can build, you can accumulate knowledge about the world. 00:08:06.800 |
However, you can, you can imagine a human completely paralyzed and there is a lot that 00:08:13.400 |
the human brain could learn, you know, with a paralyzed body. 00:08:17.640 |
So if we got a machine that could do that, that would be a big deal. 00:08:23.520 |
And then the flip side of that, something you see in children and something in machine 00:08:31.440 |
Maybe it is also is being able to play with the world. 00:08:36.780 |
How important for developing system one or system two, do you think it is to play with 00:08:45.040 |
Well, certainly a lot, a lot of what you learn as you learn to anticipate the outcomes of 00:08:52.880 |
I mean, you can see that how babies learn it, you know, with their hands, they, how 00:08:57.400 |
they learn, you know, to connect, you know, the movements of their hands with something 00:09:03.600 |
that clearly is something that happens in the brain and, and, and the ability of the 00:09:11.660 |
So you know, it's the kind of thing that you get with artificial limbs that you connected 00:09:17.200 |
and then people learn to operate the artificial limb, you know, really impressively quickly, 00:09:28.120 |
So we have a system that is ready to learn the world through action. 00:09:33.200 |
At the risk of going into way too mysterious of land, what do you think it takes to build 00:09:42.920 |
Obviously we're very far from understanding how the brain works, but how difficult is 00:09:52.880 |
You know, I mean, I think that Jan LeCun's answer that we don't know how many mountains 00:10:00.560 |
I think that, you know, if you, if you look at what Ray Kurzweil is saying, that strikes 00:10:06.640 |
me as off the wall, but, but I think people are much more realistic than that. 00:10:13.480 |
We're actually, Demis Hassabis is and Jan is, and so the people are actually doing the 00:10:23.840 |
To maybe phrase it another way, from a perspective, not of building it, but from understanding 00:10:28.960 |
it, how complicated are human beings in the following sense? 00:10:36.080 |
You know, I work with autonomous vehicles and pedestrians, so we tried to model pedestrians. 00:10:41.400 |
How difficult is it to model a human being, their perception of the world, the two systems 00:10:48.800 |
they operate under, sufficiently to be able to predict whether the pedestrian is going 00:10:55.040 |
I'm, you know, I'm fairly optimistic about that, actually, because what we're talking 00:11:00.800 |
about is a huge amount of information that every vehicle has, and that feeds into one 00:11:13.220 |
And so anything that any vehicle learns becomes part of what the whole system knows. 00:11:19.200 |
And with a system multiplier like that, there is a lot that you can do. 00:11:25.240 |
So human beings are very complicated, but, and, you know, system is going to make mistakes, 00:11:34.160 |
I think that they'll be able to, I think they are able to anticipate pedestrians, otherwise 00:11:42.680 |
They're able to, you know, they're able to get into a roundabout and into traffic, so 00:11:51.360 |
they must know both to expect or to anticipate how people will react when they're sneaking 00:11:59.840 |
And there's a lot of learning that's involved in that. 00:12:03.840 |
Currently, the pedestrians are treated as things that cannot be hit, and they're not 00:12:12.040 |
treated as agents with whom you interact in a game-theoretic way. 00:12:18.800 |
So I mean, it's not, it's a totally open problem, and every time somebody tries to solve it, 00:12:27.120 |
And nobody's really tried to seriously solve the problem of that dance, because I'm not 00:12:32.360 |
sure if you've thought about the problem of pedestrians, but you're really putting your 00:12:39.040 |
You know, there is a dance, there's part of the dance that would be quite complicated, 00:12:45.000 |
but for example, when I cross the street and there is a vehicle approaching, I look the 00:12:49.840 |
driver in the eye, and I think many people do that. 00:12:54.000 |
And you know, that's a signal that I'm sending, and I would be sending that machine to an 00:13:00.320 |
autonomous vehicle and it had better understand it, because it means I'm crossing. 00:13:06.080 |
So and there's another thing you do that actually, so I'll tell you what you do, because I've 00:13:12.040 |
watched hundreds of hours of video on this, is when you step in the street, you do that 00:13:17.000 |
before you step in the street, and when you step in the street, you actually look away. 00:13:25.880 |
What that's saying is, I mean, you're trusting that the car, who hasn't slown down yet, will 00:13:37.760 |
So I'm committed, and if I'm committed, I'm looking away. 00:13:44.960 |
So the question is whether a machine that observes that needs to understand mortality. 00:13:49.760 |
Here, I'm not sure that it's got to understand so much as it's got to anticipate. 00:13:59.680 |
So and here, but you know, you're surprising me, because here I would think that maybe 00:14:08.280 |
you can anticipate without understanding, because I think this is clearly what's happening 00:14:14.120 |
in playing Go or in playing chess, there's a lot of anticipation and there is zero understanding. 00:14:20.440 |
So I thought that you didn't need a model of the human and a model of the human mind 00:14:30.480 |
to avoid hitting pedestrians, but you are suggesting that actually.