back to indexTomaso Poggio: Brains, Minds, and Machines | Lex Fridman Podcast #13
Chapters
0:0
4:0 Time Travel Is Possible
22:38 Parts of the Brain
26:28 Cortex
28:12 The Human Visual Cortex
41:19 Stochastic Gradient Descent
43:18 Bazoo Theorem
44:53 The Universality a Universal Approximation Theorem
47:15 The Curse of Dimensionality
47:56 Challenges of Unsupervised Learning
55:8 Object Recognition Problem
56:31 The Existential Threat of Ai
61:21 Levels of Understanding
63:37 Ethics of Neuroscience
65:42 The Hard Problem of Consciousness
70:7 Next Breakthrough
71:56 Virtual Reality Experiment
78:51 Intelligence Is a Gift or Curse
00:00:00.000 |
The following is a conversation with Tomasso Poggio. 00:00:05.540 |
of the Center for Brains, Minds and Machines. 00:00:08.360 |
Cited over 100,000 times, his work has had a profound impact 00:00:13.360 |
on our understanding of the nature of intelligence 00:00:16.160 |
in both biological and artificial neural networks. 00:00:19.880 |
He has been an advisor to many highly impactful researchers 00:00:28.000 |
Amnon Shashua of Mobileye and Christophe Koch 00:00:40.240 |
If you enjoy it, subscribe on YouTube, iTunes, 00:00:42.760 |
or simply connect with me on Twitter @LexFriedman, 00:00:47.960 |
And now, here's my conversation with Tomasso Poggio. 00:00:59.720 |
and that Einstein was also a childhood hero to you. 00:01:07.920 |
the nature of his genius, do you think was essential 00:01:15.960 |
and I'm sure to many people because he was able to make, 00:01:20.960 |
of course, a major, major contribution to physics 00:01:25.180 |
with simplifying a bit just a Gedanken experiment, 00:01:33.600 |
You know, imagining communication with lights 00:01:38.880 |
between a stationary observer and somebody on a train. 00:01:43.320 |
And I thought, you know, the fact that just with the force 00:01:48.320 |
of his thought, of his thinking, of his mind, 00:02:04.100 |
It was the power of intelligence, the power of the mind. 00:02:11.120 |
to visualize as he did, as a lot of great physicists do, 00:02:15.180 |
do you think that's in all of us, human beings, 00:02:27.100 |
and have, in principle, similar breakthroughs. 00:02:32.100 |
There are lessons to be learned from Einstein. 00:02:42.180 |
the Eidgenossische Technische Hochschule in Zurich, 00:02:47.540 |
in physics, and he was the worst of the five, 00:02:50.820 |
the only one who did not get an academic position 00:03:04.820 |
for the patent office, but the fact that, obviously, 00:03:19.980 |
So there is a lot to be said about trying to be, 00:03:24.760 |
to do the opposite or something quite different 00:03:39.660 |
- So you've also mentioned, staying on the theme of physics, 00:03:47.620 |
by the mysteries of the universe that physics could uncover. 00:03:50.820 |
Such, as I saw mentioned, the possibility of time travel. 00:03:55.460 |
So the most out-of-the-box question, I think, 00:03:59.460 |
I'll get to ask today, do you think time travel is possible? 00:04:02.460 |
- Well, it would be nice if it were possible right now. 00:04:12.860 |
- But your understanding of the nature of time. 00:04:15.060 |
- Yeah, it's very likely that it's not possible 00:04:35.940 |
But in terms of actively traveling, for instance, 00:04:49.180 |
of the engineering intelligence that will build systems 00:05:00.700 |
that would be required to travel through time? 00:05:08.660 |
I don't think whether, there are certain problems 00:05:14.460 |
depending what you believe about the physical reality. 00:05:18.340 |
Like, maybe totally impossible to create energy 00:05:43.220 |
is happening already with the computers we have. 00:05:53.260 |
at some point could not become more intelligent 00:06:18.740 |
the problem of intelligence is the greatest problem 00:06:30.700 |
said that you're open to arguments against you. 00:06:46.500 |
- Well, originally, I think one of the motivation 00:06:54.980 |
when I was infatuated with theory of relativity 00:07:02.740 |
the problem of time and space and general relativity, 00:07:09.940 |
of the same level of difficulty and importance 00:07:16.620 |
it was difficult to hope to solve all of them. 00:07:19.500 |
So what about solving a problem whose solution 00:07:31.060 |
to an intelligence 10 times better or faster than Einstein? 00:07:35.900 |
- So that's sort of seeing artificial intelligence 00:07:43.220 |
but is there just an inherent curiosity in you 00:08:07.140 |
became a real focus of my science and my research 00:08:20.140 |
the most interesting problem is really asking 00:08:31.700 |
but even about the very tool we are using to do science, 00:08:45.020 |
- And that, in many ways, is the ultimate question 00:09:12.100 |
the understanding how the human brain creates intelligence? 00:09:17.660 |
a strong AI system without really getting at the core, 00:09:21.060 |
the understanding of the functional nature of the brain? 00:09:42.460 |
It was important, I guess, to know that you could have 00:09:54.580 |
But beyond that, probably we did not learn very much. 00:10:01.140 |
The Brothers Wright did learn a lot of observation 00:10:11.580 |
But you can argue we did not use much of biology 00:10:25.500 |
If you ask, okay, we all agree we'll get at some point, 00:10:42.180 |
say, in terms of what I can ask the machine to do. 00:10:51.980 |
do you think we'll get there without any knowledge 00:10:59.260 |
to get there is to understand better the human brain? 00:11:05.980 |
that different people with different background 00:11:15.060 |
in the last, I would say, five years or 10 years 00:11:35.900 |
is one of the algorithms at the core of AlphaGo, 00:11:43.700 |
the kind of an official world champion of Go, 00:12:06.700 |
And deep learning started, which is at the core, again, 00:12:12.580 |
of AlphaGo and systems like autonomous driving systems 00:12:22.500 |
which is a company started by one of my ex-post-docs, 00:12:25.620 |
Amnon Shashua, that is at the core of those things. 00:12:43.140 |
and David Hubel at Harvard, up the river in the '60s. 00:12:54.340 |
My personal bet is that there is a good chance 00:13:12.180 |
You said these mechanisms that underlie deep learning, 00:13:18.460 |
but there is nevertheless significant differences 00:13:23.920 |
between biological and artificial neural networks 00:13:44.900 |
too simplistic relative to real neural networks. 00:14:09.660 |
that computer science used as model of thinking, 00:14:31.620 |
as I said, caricature of the biological neurons. 00:14:53.140 |
in artificial neural networks added over time 00:14:57.280 |
as we try to figure out ways to improve them? 00:15:03.900 |
and problems in terms of deep learning today, 00:15:12.820 |
and the brain is the need for deep learning techniques 00:15:24.700 |
you have a training set which is one million images, 00:15:59.620 |
I think that there is this interesting challenge 00:16:03.900 |
that today deep learning and related techniques 00:16:09.500 |
big data meaning a lot of examples labeled by humans. 00:16:30.260 |
But I think the biological world is more n going to one. 00:16:58.460 |
there's because the world of Go is so simplistic 00:17:05.940 |
through self play, you can play against each other. 00:17:12.500 |
is a lot more complicated than the game of Go. 00:17:18.260 |
which are fascinatingly good at learning new stuff, 00:17:29.540 |
is in a sense is the old question of nurture and nature, 00:17:35.620 |
and how much is in the experience of an individual. 00:17:49.820 |
gives, put prior information, so to speak, hardwired, 00:18:05.300 |
has almost necessarily, if you believe in Darwin, 00:18:14.860 |
And think about our DNA and the DNA of Drosophila, 00:18:19.860 |
our DNA does not have many more genes than Drosophila. 00:18:32.500 |
Now we know that the fruit fly does not learn very much 00:18:49.820 |
But since we don't have many more genes than Drosophila, 00:19:09.860 |
Like for instance, let me give a specific example 00:19:19.300 |
We know because of work of other people in our group 00:19:26.700 |
in a part of our brain, neurons, that are tuned to faces. 00:19:31.220 |
They seem to be involved in face recognition. 00:19:33.860 |
Now this face area exists, seems to be present 00:19:42.780 |
And one question is, is there from the beginning, 00:20:06.420 |
you're a good person to guess at some of these. 00:20:16.420 |
that seems to be concentrated on face recognition, 00:20:21.220 |
Or you just, it's designed to learn that quickly? 00:20:32.260 |
And it turns out that Marge Livingstone at Harvard 00:20:42.980 |
depriving them of faces during the first weeks of life. 00:20:47.220 |
So they see technicians, but the technicians have a mask. 00:20:53.860 |
- And so when they looked at the area in the brain 00:21:00.060 |
of these monkeys, that where usually you find faces, 00:21:09.140 |
So my guess is that what evolution does in this case 00:21:15.820 |
is there is a plastic, an area which is plastic, 00:21:20.380 |
which is kind of predetermined to be imprinted very easily. 00:21:25.380 |
But the command from the gene is not a detailed circuitry 00:21:32.300 |
Could be, but this will require probably a lot of bits. 00:21:39.700 |
Instead, the command from the gene is something like imprint, 00:22:00.440 |
It'd be interesting if a variant of that experiment 00:22:10.240 |
- There are indications that during that experiment, 00:22:22.240 |
that were giving to the baby monkeys the milk. 00:22:25.560 |
And some of the cells, instead of being face sensitive 00:22:34.840 |
Can you talk about what are the different parts 00:22:39.240 |
of the brain and in your view, sort of loosely, 00:22:45.800 |
Do you see the brain as a bunch of different modules 00:22:56.160 |
mush of the same kind of fundamental architecture? 00:23:13.460 |
back in the 1950 or so in which it was believed 00:23:18.460 |
for a while that the brain was equipotential, 00:23:22.980 |
You could cut out a piece and nothing special happened 00:23:32.400 |
There was a surgeon, Lashley, who did a lot of experiments 00:23:37.400 |
of this type with mice and rats and concluded 00:23:43.240 |
that every part of the brain was essentially equivalent 00:23:56.120 |
There are very specific modules in the brain, as you said, 00:24:07.520 |
or may lose control of their legs in another region. 00:24:14.540 |
The brain is also quite flexible and redundant, 00:24:26.040 |
It can take over functions from one part of the brain 00:24:29.240 |
to the other, but really there are specific modules. 00:24:33.840 |
So the answer that we know from this old work, 00:24:43.720 |
either on animals or very often there were a mine of, 00:24:48.720 |
well, there was a mine of very interesting data 00:24:53.000 |
coming from the war, from different types of-- 00:25:08.760 |
which allow you to check which part of the brain 00:25:13.840 |
are active when you are doing different tasks, 00:25:32.280 |
- But sort of taking a step back to that part of the brain 00:25:45.320 |
is it possible that sort of from a physicist's perspective, 00:25:52.720 |
when you're born, it's plastic and quickly figures out, 00:26:02.000 |
Do you have an intuition that that kind of learning 00:26:06.280 |
or is it really kind of solidified in hardware? 00:26:31.200 |
which is the most developed part of the brain in humans. 00:26:35.200 |
And in the cortex, you have different regions of the cortex 00:26:43.400 |
for audition, for motor control, for language. 00:26:50.800 |
is that in the cortex, is the cortex, is the cortex, 00:26:55.280 |
looks like it is the same in terms of hardware, 00:27:14.280 |
like spinal cord, hippocampus, cerebellum and so on, 00:27:20.440 |
about hardware and software and learning and so on, 00:27:35.600 |
computer architecture that is good for vision 00:27:48.240 |
- But the underlying mechanism might be the same 00:27:55.200 |
So you've done a lot of great work in vision, 00:28:00.620 |
And you mentioned the problem of human vision 00:28:07.480 |
And maybe that connects to the cortex discussion. 00:28:15.340 |
and how the humans begin to understand the world 00:28:36.620 |
What is it that we know about the human visual cortex 00:29:14.620 |
- Do you think, so taking a step back on that, 00:29:35.580 |
as you've done a lot of excellent work on that side? 00:29:37.860 |
Which psychology, which level of abstraction do you think? 00:30:00.180 |
I could say well I understand how to use PowerPoint. 00:30:21.420 |
I can write the equation for transistor and diodes 00:30:29.340 |
And I can ask this guy, do you know how to operate PowerPoint? 00:30:41.580 |
that are also operating under Windows and have PowerPoint, 00:30:45.580 |
do you think it's digging in a little bit more, 00:30:49.980 |
how useful is it to understand the transistor 00:30:57.900 |
and these higher level intelligent processes? 00:31:11.900 |
They are separate modules so that the engineer 00:31:19.220 |
does not need to know what is inside PowerPoint. 00:31:30.260 |
So in that case, I don't think understanding the transistor 00:31:35.260 |
help you understand PowerPoint, or very little. 00:31:39.940 |
If you want to understand the computer, this question, 00:31:43.940 |
I would say you have to understand it at different levels. 00:31:51.580 |
But for the brain, I think this levels of understanding, 00:31:57.300 |
so the algorithms, which kind of computation, 00:32:04.620 |
the transistors, I think they are much more intertwined 00:32:15.900 |
And so that's why I think in the case of the brain, 00:32:21.780 |
more than for computers requires the interaction, 00:32:25.620 |
the collaboration between different types of expertise. 00:32:35.180 |
- I think you can, but it's much more difficult, 00:32:40.860 |
And as I said, I think he's one of the person 00:32:53.300 |
That said, you do talk about compositionality 00:32:58.300 |
And when you discuss why these neural networks 00:33:01.740 |
in artificial or biological sense learn anything, 00:33:07.540 |
See, there's a sense that nature can be disentangled. 00:33:12.540 |
Well, all aspects of our cognition could be disentangled 00:33:27.740 |
and why do you think it exists at all in nature? 00:33:31.660 |
- I spoke about, I used the term compositionality 00:33:36.660 |
when we looked at deep neural networks, multi-layers, 00:34:00.020 |
And what we found is that in terms of approximating 00:34:12.220 |
like from an image to the label in the image, 00:34:33.940 |
If the function is made up of functions of functions, 00:34:49.140 |
at all pixels at once, but you can compute something 00:34:57.180 |
and then you can compute something on the output 00:35:03.660 |
It is similar to what you do when you read a sentence. 00:35:07.300 |
You don't need to read the first and the last letter, 00:35:11.340 |
but you can read syllables, combine them in words, 00:35:22.580 |
of why deep neural networks may be more effective 00:35:27.820 |
And is your sense, for most things we can use 00:35:32.260 |
neural networks for, those problems are going 00:35:37.780 |
to be compositional in nature, like language, like vision? 00:36:10.180 |
His conclusion is that, for images, for instance, 00:36:19.460 |
that we have to learn or to solve these problems 00:36:35.380 |
between particle of matter and other particles, 00:36:45.700 |
And that's true, but you could push this argument 00:37:02.820 |
but maybe what happens is kind of the opposite, 00:37:06.820 |
is that our brain is wired up as a deep network. 00:37:26.500 |
that don't have this compositional structure. 00:37:42.700 |
- And that's, in a sense, an evolutionary perspective 00:37:48.260 |
that weren't dealing with a compositional nature 00:38:00.340 |
why we have this local connectivity in the brain, 00:38:09.020 |
at the small part of the image, each one of them, 00:38:11.860 |
and then other cells looking at the small number 00:38:16.340 |
The reason for this may be purely that it was difficult 00:38:28.660 |
it's possible to grow short-range connectivity, 00:38:33.620 |
but not long-range also because there is a limited 00:38:39.740 |
And so you have this limitation from the biology. 00:38:44.020 |
And this means you build a deep convolutional network. 00:38:50.060 |
This would be something like a deep convolutional network. 00:38:53.660 |
And this is great for solving certain class of problems. 00:38:57.780 |
These are the ones we find easy and important for our life. 00:39:13.300 |
With Mobileye, driving is a compositional problem. 00:39:17.420 |
So on the learning task, I mean, we don't know much 00:39:21.940 |
about how the brain learns in terms of optimization, 00:39:25.720 |
but so the thing that's stochastic gradient descent 00:39:29.100 |
is what artificial neural networks use for the most part 00:39:33.820 |
to adjust the parameters in such a way that it's able 00:39:42.580 |
So what's your intuition about why it works at all, 00:39:47.580 |
how hard of a problem it is to optimize a neural network, 00:40:07.820 |
So I find neuroscience, the architecture of cortex 00:40:12.820 |
is really similar to the architecture of deep networks. 00:40:20.380 |
between the biology and this kind of local connectivity, 00:40:28.220 |
The stochastic gradient descent, as you said, 00:40:34.360 |
It seems pretty unlikely that biology could do that 00:40:50.260 |
So it's a big question open whether there are other 00:40:58.820 |
that can replace stochastic gradient descent. 00:41:18.360 |
The fact that stochastic gradient descent is so successful, 00:41:23.360 |
this has become clearly is not so mysterious. 00:41:27.760 |
And the reason is that it's an interesting fact 00:41:44.960 |
when you had data and you had, say, a model with parameters, 00:41:56.000 |
Typically, the kind of crowd wisdom type idea 00:42:01.000 |
was you should have at least twice the number of data 00:42:15.560 |
Now, the way you train neural network these days 00:42:19.600 |
is that they have 10 or 100 times more parameters than data. 00:42:25.440 |
And which, you know, it has been one of the puzzles 00:42:40.680 |
- From that little data, it can generalize somehow. 00:42:48.200 |
- So I think we have some initial understanding 00:42:52.280 |
But one nice side effect of having this over-parameterization 00:43:00.040 |
is that when you look for the minima of a loss function, 00:43:14.200 |
some old basic theorem of algebra called the Bezout theorem 00:43:19.200 |
that gives you an estimate of the number of solution 00:43:26.000 |
Anyway, the bottom line is that there are probably 00:43:44.880 |
- A more global minima, zero minima, good minima. 00:43:54.520 |
So it's not so surprising that you can find them 00:43:59.240 |
And this is because of the over-parameterization. 00:44:04.280 |
- The over-parameterization sprinkles that entire space 00:44:11.240 |
It's like, if you have a system of linear equation 00:44:22.000 |
And the question is to pick one, that's another story. 00:44:25.400 |
But you have an infinite number of solutions. 00:44:33.120 |
- But it's possible that there's a lot of those solutions 00:44:37.360 |
What's surprising is that they're pretty good. 00:44:43.680 |
- But that's a separate question with separate answers. 00:44:52.680 |
of the power of neural networks is the universality, 00:44:57.840 |
that you can approximate any computable function 00:45:07.680 |
Do you find it useful, interesting, inspiring? 00:45:11.560 |
- No, this one, you know, I never found it very surprising. 00:45:16.440 |
It was known since the '80s, since I entered the field, 00:45:21.440 |
because it's basically the same as Weierstrass theorem, 00:45:27.160 |
which says that I can approximate any continuous function 00:45:34.600 |
with a sufficient number of terms, monomials. 00:45:38.320 |
- It's basically the same, and the proofs are very similar. 00:45:41.680 |
- So your intuition was there was never any doubt 00:45:48.040 |
- Right, the question, the interesting question is that 00:45:51.680 |
if this theorem says you can approximate, fine, 00:45:58.280 |
but when you ask how many neurons, for instance, 00:46:03.160 |
or in the case of polynomial, how many monomials, 00:46:09.440 |
Then it turns out that that depends on the dimensionality 00:46:16.360 |
of your function, how many variables you have. 00:46:20.560 |
But it depends on the dimensionality of your function 00:46:25.080 |
It's, for instance, suppose you want an error 00:46:28.680 |
which is no worse than 10% in your approximation. 00:46:35.080 |
You come up with a network that approximates your function 00:46:40.480 |
Then it turns out that the number of units you need 00:46:44.520 |
are in the order of 10 to the dimensionality, D, 00:46:57.320 |
But if you have, say, 200 by 200 pixel images, 00:47:06.880 |
- We again go to the size of the universe pretty quickly. 00:47:12.500 |
And so this is called the curse of dimensionality. 00:47:28.040 |
- What we proved is that if you have deep layers, 00:47:32.280 |
hierarchical architecture with local connectivity 00:47:41.760 |
that has this kind of hierarchical architecture, 00:47:50.760 |
- You've spoken a lot about supervised deep learning. 00:48:01.240 |
with GANs, with generative adversarial networks? 00:48:05.680 |
Do you see those as distinct, the power of GANs, 00:48:09.200 |
do you see those as distinct from supervised methods 00:48:21.280 |
of probability densities, which is a somewhat new way 00:48:39.120 |
It's interesting, I'm less enthusiastic about it 00:48:48.680 |
I have the feeling that many people in the field 00:49:01.160 |
- Which describes the popularity of the methods, 00:49:03.080 |
but you're saying that while that's exciting and cool 00:49:06.360 |
to look at, it may not be the tool that's useful for-- 00:49:18.160 |
and we really have to figure out how to go to N to one. 00:49:29.360 |
I think they may help, they certainly have applications, 00:49:35.880 |
I did work long ago, which was a little bit similar 00:49:51.800 |
so input is images and output is, for instance, 00:49:55.360 |
the pose of the image, a face, how much is smiling, 00:50:06.360 |
with the same data set, but now I invert input and output. 00:50:14.960 |
a number, set of numbers, and the output is the image, 00:50:22.520 |
in terms of producing very realistic looking images. 00:50:35.320 |
but the output was pretty much of the same quality. 00:50:38.840 |
So I think for a computer graphics type application, 00:51:13.960 |
- So what do you think, what's your intuition? 00:51:25.080 |
So, for example, Mobileye has very successfully, 00:51:29.880 |
I mean, essentially, annotated large amounts of data 00:51:34.680 |
Now, one thought is, so we're trying to teach machines, 00:51:52.360 |
again, one caricature of the history of computer science, 00:51:58.200 |
you could say, is, it begins with programmers, 00:52:33.600 |
but how can we pick better examples with which to learn? 00:52:58.160 |
evolution is opportunistic and has weak priors. 00:53:03.160 |
You know, the way I think the intelligence of a child, 00:53:18.560 |
you can assume that you have in most organisms, 00:53:46.400 |
Even in the retinas, in the very peripheral part, 00:53:58.200 |
It may be the reason why babies tend to look, 00:54:26.720 |
or the eye of the baby is tracking the moving object, 00:54:32.900 |
- Yeah, so just purely on the visual characteristics 00:54:36.080 |
of the scene, that seems to be the most useful. 00:54:50.400 |
So suppose you do this at the beginning, first weeks, 00:55:05.760 |
- So that's the, by the way, I just wanna ask, 00:55:10.880 |
so there is this being responsive to movement, 00:55:22.640 |
detecting where it is, and understanding the scene? 00:55:35.080 |
I think present algorithm, with all the success 00:55:40.040 |
that we have, and the fact that there are a lot 00:55:43.160 |
of very useful, I think we are in a golden age 00:55:56.840 |
There are many more things of similar level to be done, 00:56:02.020 |
but we are far from what we call understanding 00:56:05.540 |
of a scene, of language, of actions, of people. 00:56:10.540 |
That is, despite the claims, that's, I think, very far. 00:56:19.460 |
So, in popular culture, and among many researchers, 00:56:27.480 |
in and out of the AI field, there's a concern 00:56:37.900 |
And is it valuable to think about large-scale, 00:57:14.480 |
But, as I said, I think it's good to put in place 00:57:23.400 |
What I find a bit misleading are things like, 00:57:30.300 |
like Elon Musk, and what is, Bostrom, in particular? 00:57:41.680 |
for instance, AI is more dangerous than nuclear weapons. 00:57:52.740 |
Because in terms of priority, we should still 00:57:59.520 |
and, you know, what people are doing about it, 00:58:20.600 |
Do you have any updates for those predictions? 00:58:32.880 |
What's your prediction about when you'll be truly surprised? 00:58:41.860 |
- You know, it's so difficult to predict the future, 00:59:11.240 |
do you think it will be, the underlying design 00:59:16.360 |
of such a system is something we'll be able to understand? 00:59:37.100 |
it depends what you really mean for understanding. 00:59:53.340 |
I think we're beginning to have a theory now, 01:00:01.540 |
simpler kernel machines or linear classifier, 01:00:06.340 |
we really don't understand the individual units also. 01:00:11.340 |
But we understand, you know, what the computation 01:00:15.980 |
and the limitations and the properties of it are. 01:00:22.500 |
You know, we, what does it mean to understand 01:00:29.620 |
How many of us, you know, many of us understand 01:00:40.660 |
- In that sense, understanding is, as a community, 01:00:43.460 |
as a civilization, can we build another copy of it? 01:00:47.500 |
- And in that sense, do you think there'll be, 01:00:50.740 |
there'll need to be some evolutionary component 01:00:56.260 |
Or do you think it could be engineered from the ground up? 01:00:59.340 |
The same way you go from the transistor to PowerPoint. 01:01:02.300 |
- Right, so many years ago, this was actually, 01:01:13.420 |
who was one of the founding father of computer vision, 01:01:20.580 |
I wrote a paper about levels of understanding, 01:01:23.820 |
which is related to the question we discussed earlier 01:01:48.300 |
and one level I added to those three was learning. 01:01:53.060 |
So, and you can imagine, you could have a good understanding 01:01:59.340 |
of how you construct learning machine, like we do. 01:02:08.820 |
what the learning machines will discover, right? 01:02:13.740 |
Now, that would be still a powerful understanding 01:02:25.300 |
- Just like our children, if they start listening 01:02:32.020 |
Miley Cyrus or something, you don't understand 01:02:50.420 |
it has a certain, one of the challenging things 01:03:06.860 |
What is the origin on the neural and low level of ethics? 01:03:16.280 |
- I think, yeah, ethics is learnable, very likely. 01:03:27.180 |
think understanding the neuroscience of ethics. 01:03:34.740 |
You know, people discuss there is an ethics of neuroscience. 01:03:42.540 |
- You know, how a neuroscientist should or should not behave. 01:03:45.900 |
You can think of a neurosurgeon and the ethics 01:03:53.940 |
But I'm more interested on the neuroscience of-- 01:04:01.100 |
- Yeah, and I think that would be important to understand 01:04:09.420 |
that are ethical machines in our sense of ethics. 01:04:14.420 |
- And you think there is something in neuroscience, 01:04:21.540 |
that could help us shed some light on ethics? 01:04:25.340 |
Or is it more so on the psychologist's sociology 01:04:29.860 |
- No, there is psychology, but there is also, 01:04:41.140 |
that are involved in certain ethical judgment. 01:04:44.500 |
And not only this, you can stimulate those area 01:04:47.620 |
with magnetic fields and change the ethical decisions. 01:04:56.380 |
- So that's work by a colleague of mine, Rebecca Sachs, 01:05:00.780 |
and there is other researchers doing similar work. 01:05:08.260 |
but ideally at some point we'll have an understanding 01:05:17.420 |
- The big why question, yeah, it must have some purpose. 01:05:21.980 |
- Yeah, obviously it has some social purposes, probably. 01:05:26.980 |
- If neuroscience holds the key to at least illuminate 01:05:33.600 |
some aspect of ethics, that means it could be 01:05:38.860 |
- And as we're getting into harder and harder questions, 01:05:42.020 |
let's go to the hard problem of consciousness. 01:05:45.460 |
Is this an important problem for us to think about 01:05:49.020 |
and solve on the engineering of intelligence side 01:06:02.700 |
partly because it's very difficult to define consciousness. 01:06:11.880 |
and about whether consciousness, and philosophers, 01:06:21.200 |
of course, whether consciousness is something 01:06:31.480 |
or could be, you know, that we could have silicon devices 01:06:36.480 |
that are conscious, or up to statement like everything 01:06:43.360 |
has some degree of consciousness and some more than others. 01:06:56.280 |
- Okay, yeah, Christoph was my first graduate student. 01:07:00.680 |
- Do you think it's important to illuminate aspects 01:07:04.960 |
of consciousness in order to engineer intelligence systems? 01:07:21.000 |
in artificial intelligence, I think, would answer 01:07:31.840 |
because it's a very engineering answer to the question. 01:07:36.520 |
- Pass the Turing test, we don't need consciousness. 01:07:38.160 |
But if you were to go, do you think it's possible 01:07:42.600 |
that we need to have that kind of self-awareness? 01:07:49.920 |
So for instance, I personally think that when test 01:08:00.520 |
in an extended Turing test, I think consciousness 01:08:18.560 |
Despite many other romantic notions he holds, 01:08:36.840 |
do you think mortality and those kinds of things 01:08:46.920 |
and for intelligence, the finiteness of life, 01:08:51.500 |
finiteness of existence, or is that just a side effect 01:09:03.280 |
this interview is gonna run out of time soon, 01:09:08.080 |
do you think that's needed to make this conversation good 01:09:15.920 |
I think Steve Jobs in his commencement speech 01:09:31.680 |
- You live every day like it's your last, right? 01:09:46.200 |
They seem to go together in our biological system, right? 01:09:56.160 |
AlphaGo immobilized the big recent success stories in AI. 01:10:06.060 |
So what do you think will be the next breakthrough? 01:10:10.680 |
What's your intuition about the next breakthrough? 01:10:13.720 |
- Of course, I don't know where the next breakthrough is. 01:10:16.800 |
I think that there is a good chance, as I said before, 01:10:21.440 |
that the next breakthrough would also be inspired by, 01:10:32.320 |
- And there's, so MIT has this quest for intelligence. 01:10:35.840 |
And there's a few moonshots, which, in that spirit, 01:10:44.120 |
- Well, of course I'm excited about one of the moonshots, 01:10:48.760 |
which is our Center for Brains, Minds, and Machines, 01:10:51.760 |
which is the one which is fully funded by NSF. 01:11:02.680 |
- And that one is particularly about understanding. 01:11:36.760 |
These are all things that we perceive very quickly. 01:11:41.240 |
And it's something actually quite close to being conscious, 01:11:46.240 |
not quite, but there is this interesting experiment 01:11:50.360 |
that was run at Google X, which is in a sense, 01:11:58.860 |
but in which they had subjects sitting, say, in a chair, 01:12:07.140 |
earphones, and they were seeing through the eyes 01:12:53.600 |
have a self-awareness about your position in the world 01:13:07.760 |
So you've been an advisor to some incredible minds, 01:13:24.240 |
and from perspective as a mentor of these researchers, 01:13:34.200 |
what does it take to be successful in science 01:13:37.880 |
Whether you're talking to somebody in their teens, 01:15:13.200 |
would be that you'll actually end up discovering 01:15:16.320 |
So as you've led many incredible efforts here, 01:15:30.200 |
Or yeah, what advice could you give to people, 01:16:02.440 |
of when somebody comes with a new idea in the group, 01:16:30.480 |
Sometimes people are very critical from the beginning. 01:16:51.120 |
Is that a part of science and academic pursuits 01:17:31.840 |
So, as long as this does not become personal, 01:17:45.480 |
So, if you somehow find yourself in a position 01:18:06.000 |
it could be, how could I become 10 times more intelligent? 01:18:11.760 |
- And so, but see, you only get a clear, short answer. 01:18:16.200 |
So, do you think there's a clear, short answer to that? 01:18:27.960 |
- As a story that inspired you in your childhood, 01:18:51.440 |
do you think intelligence is a gift or a curse, 01:18:55.360 |
from the perspective of happiness and meaning of life? 01:19:03.880 |
but on an individual level, the meaning of life, 01:19:23.680 |
as one people consider the smartest people in the world, 01:19:33.280 |
- I don't know, it may be invariant to intelligence, 01:19:46.160 |
- You could be smart and happy and clueless and happy. 01:19:51.800 |
- As always, on the discussion of the meaning of life, 01:19:57.320 |
Tomasso, thank you so much for talking today.