back to indexDileep George: Brain-Inspired AI | Lex Fridman Podcast #115
Chapters
0:0 Introduction
4:50 Building a model of the brain
17:11 Visual cortex
27:50 Probabilistic graphical models
31:35 Encoding information in the brain
36:56 Recursive Cortical Network
51:9 Solving CAPTCHAs algorithmically
66:48 Hype around brain-inspired AI
78:21 How does the brain learn?
81:32 Perception and cognition
85:43 Open problems in brain-inspired AI
90:33 GPT-3
100:41 Memory
105:8 Neuralink
111:32 Consciousness
117:59 Book recommendations
126:49 Meaning of life
00:00:00.000 |
The following is a conversation with Dalipe George, 00:00:03.440 |
a researcher at the intersection of neuroscience 00:00:11.000 |
and formerly co-founder of Numenta with Jeff Hawkins, 00:00:14.800 |
who's been on this podcast, and Donna Dubinsky. 00:00:18.880 |
From his early work on hierarchical temporal memory 00:00:24.760 |
Dalipe's always sought to engineer intelligence 00:00:31.240 |
As a side note, I think we understand very little 00:00:41.200 |
that may be more useful for engineering intelligence 00:00:43.680 |
than any idea in mathematics, computer science, physics, 00:00:50.520 |
And so the brain is a kind of existence proof 00:00:56.320 |
I should also say that brain-inspired AI is often overhyped 00:01:00.600 |
and use this fodder, just as quantum computing 00:01:08.080 |
sometimes overhyped areas since where there's smoke, 00:01:15.360 |
Three sponsors, Babbel, Raycon Earbuds, and Masterclass. 00:01:21.760 |
by clicking the special links in the description 00:01:25.880 |
It really is the best way to support this podcast. 00:01:29.160 |
If you enjoy this thing, subscribe on YouTube, 00:01:31.560 |
review it with the five stars on Apple Podcasts, 00:01:33.760 |
support on Patreon, or connect with me on Twitter 00:01:52.040 |
Go to babbel.com and use code Lex to get three months free. 00:01:55.740 |
They offer 14 languages, including Spanish, French, 00:02:03.140 |
Daily lessons are 10 to 15 minutes, super easy, effective, 00:02:10.700 |
Let me read a few lines from the Russian poem, 00:02:13.540 |
"Noch, ulitza, fanar, apteka," by Alexander Bloch, 00:02:18.360 |
that you'll start to understand if you sign up to Babbel. 00:02:26.060 |
Now, I say that you'll only start to understand this poem 00:02:43.700 |
Now, the latter part is definitely not endorsed 00:02:53.600 |
of late night Russian conversation over vodka. 00:03:12.580 |
They've become my main method of listening to podcasts, 00:03:15.060 |
audio books, and music when I run, do pushups and pull-ups, 00:03:20.740 |
In fact, I often listen to Brown Noise with them 00:03:27.060 |
They're super comfortable, pair easily, great sound, 00:03:34.080 |
I've been putting in a lot of miles to get ready 00:03:38.220 |
and listening to audio books on World War II. 00:03:52.780 |
Sign up at masterclass.com/lex to get a discount 00:04:03.820 |
For 180 bucks a year, you get an all access pass 00:04:06.760 |
to watch courses from, to list some of my favorites, 00:04:24.420 |
Carlos Santana on guitar, Garry Kasparov on chess, 00:04:33.500 |
and the experience of being launched into space alone 00:04:37.540 |
By the way, you can watch it on basically any device. 00:04:40.740 |
Once again, sign up at masterclass.com to get a discount 00:04:45.900 |
And now here's my conversation with Dilip George. 00:04:56.260 |
we definitely need to understand how it works. 00:05:02.380 |
is trying to build a brain without understanding it, 00:05:10.700 |
from neuroscience experiments into a giant simulation. 00:05:16.100 |
By putting more and more neurons, more and more details. 00:05:21.500 |
because when it doesn't perform as what you expect it to do, 00:05:39.380 |
what they're going to contribute, you can't build it. 00:05:57.740 |
- Right, human brains and rat brains or cat brains 00:06:03.540 |
That the cortex, the neocortex structure is very similar. 00:06:08.220 |
So initially they were trying to just simulate a cat brain 00:06:22.740 |
you easily get one thing out, which is oscillations. 00:06:27.700 |
If you simulate a large number of neurons, they oscillate. 00:06:32.340 |
And you can adjust the parameters and say that, 00:07:00.460 |
And you interconnect them according to the statistics 00:07:13.100 |
And these neural models are incredibly complicated 00:07:17.940 |
Because these neurons are modeled using this idea 00:07:24.580 |
which are about how signals propagate in a cable. 00:07:28.260 |
And there are active dendrites, all those phenomena, 00:07:46.580 |
we just have to take whatever comes out of it as, 00:07:59.620 |
like with the axons and all the basic models, 00:08:04.500 |
- Oh, well, actually, they are pretty detailed 00:08:16.460 |
and you try to turn on the different channels, 00:08:20.980 |
the calcium channels and the different receptors, 00:08:27.820 |
or off those channels are in the neuron's spike output, 00:08:32.740 |
people have built pretty sophisticated models of that. 00:08:35.420 |
And they are, I would say, in the regime of correct. 00:08:40.420 |
- Well, see, the correctness, that's interesting, 00:08:43.300 |
'cause you've mentioned it at several levels. 00:08:46.980 |
by looking at some kind of aggregate statistics? 00:08:49.580 |
- It would be more the spiking dynamics of the-- 00:08:53.260 |
- Spiking dynamics of the single neuron, okay. 00:08:57.980 |
because they are going to the level of mechanism, right? 00:09:02.620 |
okay, what is the effect of turning on an ion channel? 00:09:06.660 |
And you can model that using electric circuits. 00:09:16.460 |
it is people are looking at the mechanism underlying it, 00:09:19.260 |
and putting that in terms of electric circuit theory, 00:09:23.540 |
signal propagation theory, and modeling that. 00:09:49.220 |
if you did not understand how a microprocessor works, 00:09:52.500 |
but you say, oh, I now can model one transistor well, 00:09:56.180 |
and now I will just try to interconnect the transistors 00:10:02.100 |
guess from the experiments and try to simulate it, 00:10:08.100 |
that you will produce a functioning microprocessor. 00:10:12.340 |
when you want to produce a functioning microprocessor, 00:10:16.860 |
how does, how do the gates work, all those things, 00:10:21.300 |
understand how do those gates get implemented 00:10:29.020 |
that I remember going through in a reading group 00:10:39.140 |
it uses all the tools that we have of neuroscience 00:10:43.620 |
like as if we just aliens showed up to study computers, 00:10:50.940 |
to get any kind of sense of how the microprocessor works. 00:10:57.860 |
at least this initial exploration is that we're screwed. 00:11:02.380 |
There's no way that the tools of neuroscience 00:11:09.700 |
any aspect of the architecture of the function 00:11:21.820 |
you can't figure that out from the tools of neuroscience. 00:11:24.480 |
- Yeah, so I'm very familiar with this particular paper. 00:11:30.340 |
Can a Neuroscientist Understand a Microprocessor? 00:11:44.380 |
So I don't think it is that bad in the sense of saying, 00:12:20.040 |
put those things together and build hypothesis. 00:12:22.920 |
So I don't want to diss all of neuroscientists saying, 00:12:36.740 |
but it has to be put together in a computational framework. 00:12:43.580 |
will be listening to this podcast 100 years from now, 00:13:18.140 |
So the dynamics of the individual neurocommunication 00:13:35.100 |
- Yeah, so timelines are very, very hard to predict, 00:13:57.940 |
And humans might not fly for another 100 years. 00:14:04.780 |
And so, but no, they flew three years after that. 00:14:11.540 |
- Well, and on that point, one of the Wright brothers, 00:14:18.380 |
said that, he said some number, like 50 years, 00:14:22.900 |
he has become convinced that it's impossible. 00:14:27.900 |
- Even during their experimentation, yeah, yeah, yeah. 00:14:34.140 |
that's like the entrepreneurial battle of depression, 00:14:37.140 |
of going through just thinking this is impossible. 00:14:50.500 |
objectively, what are the things that we know about the brain 00:14:57.140 |
which can then go back and inform how the brain works. 00:15:04.220 |
look at the insights neuroscientists have found, 00:15:11.140 |
information processing angle, build models using that. 00:15:15.300 |
And then building that model, which functions, 00:15:20.500 |
which is doing the task that we want the model to do. 00:15:23.780 |
It is not just trying to model a phenomena in the brain. 00:15:26.780 |
It is trying to do what the brain is trying to do 00:15:41.820 |
fills in the rest of the pieces of the puzzle. 00:15:44.700 |
And then you can go and connect that back to biology 00:15:53.660 |
or this layer in the cortical circuit is doing this. 00:16:21.780 |
so neuroscientists alone, just from experimentation, 00:16:27.180 |
will not be able to build a model of the brain, 00:16:36.540 |
in collecting more and more connectivity data 00:16:54.020 |
by themselves, convey the story of how does it work. 00:17:08.100 |
using hints from neuroscience, and repeat the cycle. 00:17:19.300 |
you're both a neuroscientist and an AI person. 00:17:23.140 |
I guess the dream is to both understand the brain 00:17:27.460 |
So you're, it's like an engineer's perspective 00:17:33.860 |
So what aspects of the brain, functionally speaking, 00:17:45.180 |
and the visual cortex is a large part of the brain. 00:18:05.820 |
There are a lot more feedback connections in the brain 00:18:32.500 |
And you will say, yes, they have studied that. 00:18:35.380 |
- So every possible combination has been studied. 00:18:38.820 |
- I mean, it's not a random exploration at all. 00:18:46.420 |
are very, very systematic in how they probe the brain. 00:18:49.340 |
Because experiments are very costly to conduct. 00:18:58.580 |
And often what I find is that when we have a question 00:19:13.140 |
and people have probed it very systematically. 00:19:15.260 |
And they have hypothesis about how those lateral connections 00:19:20.100 |
are supposedly contributing to visual processing. 00:19:30.460 |
sorry to interrupt, do they stimulate like a neuron 00:19:36.540 |
and then see how the signal travels kind of thing? 00:19:39.740 |
- Fascinating, very, very fascinating experiments. 00:19:41.940 |
So I can give you one example I was impressed with. 00:19:50.780 |
of how the layers in the cortex are organized. 00:20:08.260 |
I'm talking about just object recognition pathway. 00:20:14.780 |
so it's, there is a very detailed microcircuit in V1 itself. 00:20:22.580 |
The cortical sheet is organized into multiple layers, 00:20:32.580 |
is repeated in V1, V2, V4, IT, all of them, right? 00:20:37.580 |
And the connections between these layers within a level, 00:20:48.700 |
And now, so one example of an experiment people did is, 00:21:00.180 |
let's say, requires separating the foreground 00:21:05.660 |
So it's a textured triangle on a textured background. 00:21:10.660 |
And you can check, does the surface settle first, 00:21:21.780 |
so when you finally form the percept of the triangle, 00:21:28.220 |
you understand where the contours of the triangle are, 00:21:31.140 |
and you also know where the inside of the triangle is, 00:21:33.980 |
right, that's when you form the final percept. 00:21:58.460 |
- In this case, it turns out that it first settles 00:22:02.260 |
on the edges, it converges on the edge hypothesis first, 00:22:12.180 |
- And the detail to which you can study this, 00:22:15.900 |
it's amazing that you can actually not only find 00:22:23.300 |
and then you can also find which layer in V1, 00:22:51.220 |
I don't know that you are familiar with this one. 00:23:08.780 |
- And then your visual system hallucinates the edges. 00:23:13.020 |
And when you look at it, you will see a faint edge. 00:23:21.820 |
do actually neurons signal the presence of this edge? 00:23:28.580 |
Because they are not receiving anything from the input. 00:23:39.180 |
So if a real contour is present in the input, 00:24:00.020 |
it's the feedback connections that is causing them to fire. 00:24:08.980 |
So these studies are pretty impressive and very detailed. 00:24:16.940 |
you said that there may be more feedback connections 00:24:21.740 |
First of all, just for like a machine learning folks, 00:24:44.780 |
So what the heck are these feedback connections? 00:24:54.060 |
- Yeah, so this fits into a very beautiful picture 00:24:59.380 |
So the beautiful picture of how the brain works 00:25:02.340 |
is that our brain is building a model of the world. 00:25:06.420 |
I know, so our visual system is building a model 00:25:17.180 |
So what we are seeing is not just a feedforward thing 00:25:21.300 |
that just gets interpreted in a feedforward part. 00:25:23.860 |
We are constantly projecting our expectations 00:25:32.620 |
combined with what the actual sensory input is. 00:25:36.860 |
- Almost like trying to calculate the difference 00:25:40.980 |
- Yeah, I wouldn't put it as calculating the difference. 00:26:00.060 |
So the feedback mechanism, it just helps you constantly, 00:26:26.860 |
This happens at the lowest level, at the highest level. 00:26:30.260 |
In fact, feedback connections are more prevalent 00:26:41.180 |
So basically, if you have a model of the world, 00:26:55.460 |
And this inference includes projecting your model 00:27:02.620 |
back into the model and doing an iterative procedure. 00:27:13.180 |
And feedback affects what you see in the world, 00:27:24.660 |
The idea that there can be multiple competing hypothesis 00:27:28.460 |
in our model, trying to explain the same evidence, 00:27:32.580 |
and then you have to kind of make them compete. 00:27:35.940 |
And one hypothesis will explain away the other hypothesis 00:27:47.020 |
that try to explain, what do you mean by explain away? 00:27:50.180 |
- So this is a classic example in graphical models, 00:28:27.500 |
There is another class of models in machine learning 00:28:33.140 |
And you can think of them as each node in that model 00:28:38.140 |
is variable, which is talking about something. 00:28:55.260 |
is there an object present in the world or not? 00:28:58.380 |
And then, so it is another way of encoding knowledge. 00:29:12.500 |
What is the best way to explain some set of evidence 00:29:23.860 |
How is the edge connected to the model of the object? 00:29:27.660 |
How is the surface connected to the model of the object? 00:29:32.700 |
this is a very distributed, complicated model. 00:29:34.900 |
And inference is, how do you explain a piece of evidence 00:29:41.420 |
If somebody tells me there is a 50% probability 00:29:44.460 |
that there is an edge here in this part of the model, 00:29:46.900 |
how does that affect my belief on whether I should think 00:29:51.180 |
that there should be a square present in the image? 00:30:13.940 |
can be triggered by a burglar getting into your house, 00:30:37.500 |
But while driving home, if you hear on the radio 00:30:39.860 |
that there was an earthquake in the vicinity, 00:30:45.980 |
for a burglar getting into their house is diminished. 00:30:50.180 |
Because now that piece of evidence is explained 00:30:57.700 |
explaining at lower level variable, which is alarm, 00:31:10.380 |
And initially it was flowing to a burglar being present, 00:31:14.300 |
but now since there is side evidence for this other cause, 00:31:21.140 |
and evidence will now flow to the other cause. 00:31:28.380 |
- And the brain has a similar kind of mechanism 00:31:45.620 |
is it in the hardware of the actual connections? 00:31:54.340 |
- So this is a paper that we are bringing out soon. 00:32:06.100 |
One hypothesis is that you can think of a cortical column 00:32:13.060 |
A concept, think of it as an example of a concept 00:32:25.180 |
Okay, so you can think of it as a binary variable, 00:32:34.380 |
as representing that one concept, one variable. 00:32:38.140 |
And then the connections between these cortical columns 00:32:45.700 |
And then there are connections within the cortical column. 00:32:57.780 |
There are thousands of neurons in a cortical column. 00:33:09.220 |
So all cortical columns pass through this substructure. 00:33:18.980 |
implement this, that's where the knowledge is stored 00:33:22.900 |
about how these different concepts connect to each other. 00:33:27.300 |
And then the neurons inside this cortical column 00:33:33.060 |
implement this actual computations in data for inference, 00:33:39.980 |
and competing between the different hypothesis. 00:33:47.540 |
neuroscientists have actually done experiments 00:34:03.260 |
it will inhibit through this complicated loop 00:34:11.940 |
- Do they use terminology of concepts, for example? 00:34:21.420 |
it's easy to anthropomorphize and think about concepts, 00:34:29.860 |
So how would you think of concepts in that kind of way? 00:34:34.140 |
Or is it a lot messier, a lot more gray area, 00:34:43.420 |
even more messy than the artificial neural network kinds, 00:34:48.460 |
- Easiest way to think of it is a variable, right? 00:34:52.140 |
which is showing the presence or absence of something. 00:34:58.100 |
is that something that we're supposed to think of something 00:35:01.900 |
that's human interpretable, of that something? 00:35:07.100 |
There's no need for it to be human interpretable. 00:35:12.700 |
you will be able to find some interpretation of it 00:35:30.460 |
in connecting to the other entities that are, 00:35:34.740 |
Okay, so by the way, are these the cortical microcircuits? 00:35:40.020 |
- Correct, these are the cortical microcircuits. 00:35:42.300 |
That's what neuroscientists use to talk about the circuits 00:35:49.140 |
So you can think of, let's think of a neural network, 00:36:03.380 |
And then within a layer of the neural network, 00:36:14.620 |
There's a lot more intricate structure there. 00:36:17.300 |
But even within an artificial neural network, 00:36:20.380 |
you can think of feature detection plus pooling as one level. 00:36:30.980 |
So within a level, whatever is that circuitry 00:36:42.660 |
Machine learning people don't use the circuit terminology, 00:36:53.980 |
So what's interesting about, what can we say, 00:37:00.820 |
propose about the ideas around these cortical microcircuits? 00:37:10.780 |
- So the paper focuses, and your idea in our discussions now 00:37:39.780 |
This is called the recursive cortical network model 00:37:45.140 |
And we are using the same model for robotic picking 00:37:59.260 |
- On one side, it outputs the class of the image 00:38:10.500 |
So it's a model that you build to answer multiple questions. 00:38:16.580 |
for just classification or just segmentation, et cetera. 00:38:19.540 |
It's a joint model that can do multiple things. 00:38:37.940 |
The model actually uses feedback connections. 00:38:43.860 |
- So what the heck is a recursive cortical network? 00:38:50.740 |
which is essentially a brain inspired approach 00:38:57.340 |
So there are multiple layers to this question. 00:38:59.420 |
I can go from the very, very top and then zoom in. 00:39:03.420 |
So one important thing, constraint that went into the model 00:39:26.220 |
And so that means if you finally want to have a system 00:39:32.460 |
and can learn in a very conceptual model of the world 00:39:48.220 |
And one aspect of that is top-down controllability. 00:40:19.340 |
You can think about what will happen if something hits that. 00:40:36.380 |
and being able to simulate scenarios in the world. 00:40:55.940 |
and it is not just some arbitrary generative network. 00:40:58.820 |
It has to be built in a way that it is controllable top-down. 00:41:03.020 |
It is not just trying to generate a whole picture at once. 00:41:10.500 |
You don't have good photorealistic models of the world. 00:41:17.100 |
what is the color of the letter E in the Google logo? 00:41:24.140 |
- You probably have seen it millions of times. 00:41:28.900 |
So it's not, our model is not photorealistic. 00:41:31.300 |
But it has other properties that we can manipulate it. 00:41:42.100 |
So you can imagine the consequence of actions 00:41:52.740 |
So this is one constraint that went into our model. 00:42:03.220 |
this top-down controllability of the generative model. 00:42:06.500 |
- So what does top-down controllability in a model look like? 00:42:11.500 |
It's a really interesting concept, fascinating concept. 00:42:26.620 |
what is the model representing as different pieces 00:42:29.940 |
So in the RCN network, it thinks of the world, 00:42:38.420 |
is modeled separately from the foreground of the image. 00:42:41.700 |
So the objects are separate from the background. 00:42:46.980 |
that's built in fundamentally to the structure. 00:42:49.260 |
And then even that object is composed of parts. 00:42:53.020 |
And also, another one is the shape of the object 00:42:56.700 |
is differently modeled from the texture of the object. 00:43:09.380 |
- He's, so there's, he developed this IQ test 00:43:16.460 |
and it's kind of cool that there's these concepts, 00:43:20.620 |
priors that he defines that you bring to the table 00:43:24.020 |
in order to be able to reason about basic shapes 00:43:44.500 |
It doesn't, you don't need to explicitly put it as, 00:43:47.260 |
oh, objects versus foreground versus background, 00:43:52.220 |
No, these are derivable from more fundamental principles 00:44:01.060 |
- What's the property of continuity of natural signals? 00:44:06.340 |
- By the way, that sounds very poetic, but yeah. 00:44:20.820 |
- Kind of like Francois, I mean, there's objectness, 00:44:23.500 |
there's all these things that it's kind of crazy 00:44:29.620 |
because it's useful for us to perceive the world. 00:44:50.660 |
Which is an artificial signal that we created. 00:44:52.780 |
Humans are not very good at classifying QR codes. 00:44:55.620 |
We are very good at saying something is a cat or a dog, 00:45:00.100 |
where computers are very good at classifying QR codes. 00:45:03.900 |
So our visual system is tuned for natural signals. 00:45:08.460 |
And there are fundamental assumptions in the architecture 00:45:11.340 |
that are derived from natural signals properties. 00:45:15.140 |
- I wonder when you take a hallucinogenic drugs, 00:45:18.340 |
does that go into natural or is that closer to QR code? 00:45:24.780 |
- Yeah, because it is still operating using our brains. 00:45:31.340 |
I think they're becoming legalized in certain, 00:45:33.260 |
I can't wait until they become legalized to a degree 00:45:36.980 |
that you, like vision science researchers could study it. 00:45:40.940 |
- And then through medical, chemical ways, modify it. 00:45:59.180 |
- Yeah, but I think there are studies on that already. 00:46:04.260 |
Because it's not unethical to give it to rats. 00:46:16.540 |
So there's these low level things from natural signals 00:46:36.580 |
so you mentioned the priors Franscho wanted to encode 00:46:45.020 |
but it is not straightforward how to encode those priors. 00:46:51.140 |
like the object recognition, completion challenges 00:46:53.860 |
are things that we purely use our visual system to do. 00:46:59.540 |
but it is purely an output of the vision system. 00:47:05.460 |
completing the lines of that Kaninsa triangle. 00:47:13.140 |
but it is stored in our visual system in a particular way 00:47:19.180 |
And that is one of the things that we tackled in the, 00:47:24.180 |
basically saying, okay, these are the prior knowledge 00:47:29.500 |
but then how is that prior knowledge represented 00:47:37.340 |
can be done very efficiently and in a very distributed way. 00:47:43.300 |
there are so many ways of representing knowledge, 00:47:45.500 |
which is not amenable to very quick inference, 00:47:50.820 |
And so that's one core part of what we tackled 00:48:07.100 |
may be familiar with different kinds of architectures 00:48:22.780 |
- Yeah, so you can think of the delta between the model 00:48:28.700 |
if people are familiar with convolutional neural networks. 00:48:35.260 |
which is called feature detectors and pooling. 00:48:43.660 |
And if you want an intuitive idea of what is happening, 00:48:50.460 |
detecting interesting co-occurrences in the input. 00:49:06.340 |
and making it invariant to local transformations. 00:49:11.580 |
Recursive cortical network has a similar structure 00:49:16.540 |
when you look at just the feed-forward pathway. 00:49:20.020 |
it is also structured in a way that it is generative. 00:49:28.500 |
Another aspect that it has is it has lateral connections. 00:49:37.660 |
so if you have an edge here and an edge here, 00:49:50.300 |
which is to enforce compatibility between them. 00:49:56.060 |
It's basically, if you do just feature detection 00:50:35.940 |
There's some interesting connectivity things. 00:50:45.860 |
And yeah, okay, so the interconnection between adjacent, 00:50:58.300 |
- And then there's this idea of doing inference. 00:51:01.100 |
A neural network does not do inference on the fly. 00:51:05.500 |
So an example of why this inference is important is, 00:51:11.780 |
that we showed in the paper was to crack text-based captures. 00:51:35.580 |
- So captures are those strings that you fill in 00:51:50.660 |
what is that string of characters and type it. 00:51:53.340 |
And the reason captures exist is because, you know, 00:52:02.660 |
You can use a computer to create millions of accounts 00:52:10.740 |
So you want to make sure that, to the extent possible, 00:52:32.780 |
So, and text-based captures was the one which is prevalent 00:52:46.780 |
in the sense of an arbitrary text-based capture 00:52:51.740 |
But with the techniques that we have developed, 00:52:53.940 |
it can be, you know, you can quickly develop a mechanism 00:53:00.740 |
The people, they've been getting cleverer and cleverer 00:53:06.300 |
- So, okay, so that was one of the things you've tested it on 00:53:09.420 |
is these kinds of captures in 2014, '15, that kind of stuff. 00:53:23.580 |
If you want to understand how human perception works 00:53:32.300 |
And I wouldn't say capture is a solved problem. 00:53:34.860 |
We have cracked the fundamental defense of captures, 00:53:37.620 |
but it is not solved in the way that humans solve it. 00:53:48.860 |
and show them any new capture that we create. 00:54:11.100 |
otherwise I will be able to figure that out using this one. 00:54:28.220 |
No training examples from that particular style of capture. 00:54:47.660 |
which I did not show you in the training setup. 00:54:58.020 |
And Doug Hofstadter put this very beautifully 00:55:04.700 |
The central problem in AI is what is the letter A? 00:55:10.500 |
If you can build a system that reliably can detect 00:55:20.860 |
- Yeah, you don't even need to go to the B and the C 00:55:30.060 |
I mean, is it like without training examples, 00:55:39.020 |
that make up the letter A in all of its forms? 00:55:44.460 |
It can be, A can be made with two humans standing, 00:55:47.060 |
leaning against each other, holding the hands. 00:55:52.420 |
- Yeah, you might have to understand everything 00:55:54.780 |
about this world in order to understand the letter A. 00:55:58.100 |
- So it's common sense reasoning, essentially. 00:56:12.100 |
So how does this kind of the RCN architecture 00:56:15.780 |
help us to get, do a better job of that kind of thing? 00:56:21.260 |
one of the important things was being able to do inference, 00:56:30.700 |
'Cause you said like neural networks don't do inference. 00:56:34.020 |
- So what do you mean by inference in this context then? 00:56:44.300 |
- Okay, and when you make the characters crowd together, 00:56:46.740 |
what happens is that you will now start seeing 00:56:49.340 |
combinations of characters as some other new character 00:56:59.140 |
And so locally, there is very strong evidence 00:57:08.660 |
But globally, the only explanation that fits together 00:57:27.780 |
which is conflicting with the local information. 00:57:33.940 |
in the way it's used when you talk about reasoning, 00:57:45.700 |
So like you're basically doing some basic forms of reasoning 00:57:48.740 |
like integration of like how local things fit 00:57:54.140 |
- And things like explaining away coming into this one 00:57:56.940 |
because you are explaining that piece of evidence 00:58:13.260 |
if you want to do this, you can brute force it. 00:58:16.260 |
You can just show it all combinations of things 00:58:23.260 |
And you can, you know, like just train the hell 00:58:31.460 |
but it is really just doing amortized inference. 00:58:35.020 |
It is because you have shown it a lot of these combinations 00:58:40.460 |
So what you want to do is be able to do dynamic inference 00:58:44.540 |
rather than just being able to show all those combinations 00:58:48.620 |
And that's something we emphasized in the model. 00:58:53.900 |
Is that that has to do with the feedback thing? 00:58:58.740 |
I'm trying to visualize what dynamic inference 00:59:08.620 |
- And it's like, what's changing over temporarily? 00:59:12.420 |
What's the dynamics of this inference process? 00:59:14.860 |
- So you can think of it as you have at the top of the model, 00:59:28.900 |
The characters are the things that cause the pixels. 00:59:34.980 |
So the reason you mentioned causality, I guess, 00:59:37.780 |
is because there's a temporal aspect to this whole thing. 00:59:43.380 |
It is more like when, if I turn the character on, 00:59:49.180 |
Yeah, it will be after this a little bit, but yeah. 00:59:53.300 |
of like a logic causality, like hence inference, okay. 01:00:03.860 |
And locally, just when I look at just that patch of the image, 01:00:10.660 |
But when I look at it in the context of all the other causes, 01:00:14.500 |
it might not, A is not something that makes sense. 01:00:20.660 |
- Yeah, so, okay, so, and this thing performed pretty well 01:00:27.100 |
- And I mean, is there some kind of interesting intuition 01:00:36.060 |
Is there visualizations that could be human interpretable 01:00:39.740 |
- Yes, yeah, so the good thing about the model 01:00:44.500 |
so it is not just doing a classification, right? 01:00:46.620 |
It is providing a full explanation for the scene. 01:01:02.140 |
these are the pixels in the input that tells, 01:01:08.100 |
And also these are the portions I hallucinated. 01:01:10.940 |
It provides a complete explanation of that form. 01:01:15.340 |
And then these are the contours, this is the interior, 01:01:32.140 |
And then the kind of errors it makes are also, 01:02:06.820 |
the assumptions, as I mentioned, are general, right? 01:02:09.340 |
It is more, and those themselves can be applied 01:02:13.300 |
in many situations which are natural signals. 01:02:17.180 |
So it's the foreground versus background factorization 01:02:20.700 |
and the factorization of the surfaces versus the contours. 01:02:25.460 |
So these are all generally applicable assumptions. 01:02:29.060 |
So why capture, why attack the capture problem, 01:02:34.060 |
which is quite unique in the computer vision context 01:02:36.700 |
versus like the traditional benchmarks of ImageNet 01:02:43.740 |
or even segmentation tasks and all that kind of stuff. 01:02:48.340 |
what's your thinking about those kinds of benchmarks 01:02:55.220 |
for deep learning kind of algorithms where you, 01:02:58.620 |
so the settings that deep learning works in are, 01:03:02.300 |
here is my huge training set and here is my test set. 01:03:05.940 |
So the training set is almost 100x, 1000x bigger 01:03:17.500 |
The training set is way smaller than the test set. 01:03:22.700 |
- And, you know, capture is a problem that is by definition 01:03:27.700 |
hard for computers and it has these good properties 01:03:33.100 |
of strong generalization, strong out of training 01:03:38.180 |
If you are interested in studying that and putting, 01:03:49.300 |
which I think, I believe there's quite a growing body 01:03:52.460 |
of work on looking at MNIST and ImageNet without training. 01:04:01.460 |
how, what tiny fraction of the training set can we take 01:04:05.760 |
in order to do a reasonable job of the classification task? 01:04:10.460 |
Have you explored that angle in these classic benchmarks? 01:04:19.380 |
So there was also versions of, multiple versions of MNIST, 01:04:33.700 |
how quickly can you get to high level accuracy 01:04:40.060 |
- Is there some performance that you remember, 01:04:51.260 |
you know, on the order of tens or hundreds of examples 01:04:59.580 |
And it was, it was definitely better than the systems, 01:05:05.900 |
I think that's a really interesting space, actually. 01:05:09.340 |
I think there's an actual name for MNIST that, 01:05:19.220 |
I mean, people are like attacking this problem. 01:05:22.900 |
It's funny how like the MNIST will probably be with us 01:05:40.460 |
Not enough people, I don't know, maybe you can correct me, 01:05:43.900 |
but I feel like CAPTCHAs don't show up as often in papers 01:05:49.540 |
Because, you know, usually these things have a momentum, 01:06:12.980 |
- Nobody wants to think outside the box, okay. 01:06:20.700 |
What else is there interesting on the RCN side 01:06:23.940 |
before we talk about the cortical microscope? 01:06:33.860 |
And it's quite robust to out-of-distribution perturbations. 01:06:44.300 |
and advocatiously in many of the robotics tasks 01:06:48.940 |
Well, let me ask you this kind of touchy question. 01:06:51.940 |
I have to, I've spoken with your friend, colleague, 01:06:56.940 |
I mean, I have to kind of ask, there is a bit of, 01:07:08.320 |
there's critics, I mean, machine learning subreddit. 01:07:16.500 |
I mean, criticism is good, but they're a bit over the top. 01:07:21.460 |
There is quite a bit of sort of skepticism and criticism. 01:07:26.660 |
Is this work really as good as it promises to be? 01:07:30.540 |
Do you have thoughts on that kind of skepticism? 01:07:34.860 |
Do you have comments on the kind of criticism 01:07:36.780 |
we might've received about, is this approach legit? 01:07:44.620 |
Or at least as promising as it seems to be advertised as? 01:07:55.300 |
which I would argue is a very high quality journal, 01:08:00.260 |
And usually it is indicative of the quality of the work. 01:08:13.420 |
in terms of the importance of feedback connections, 01:08:23.500 |
trying to solve recognition, segmentation, all jointly 01:08:28.780 |
in a way that is compatible with higher level cognition, 01:08:33.100 |
that we brought together into something coherent 01:08:35.460 |
and workable in the world and tackling a challenging problem, 01:08:39.860 |
I think that will stay and that contribution I stand by. 01:09:02.540 |
which is not a deep neural network, it's a graphical model. 01:09:07.380 |
Now, once the paper was accepted and everything, 01:09:15.180 |
We didn't do any press release when it was published. 01:09:19.060 |
What was the press release that they wrote up? 01:09:27.260 |
And so you can see what was being hyped in that thing. 01:09:32.260 |
So it's like there is a dynamic in the community. 01:09:37.780 |
That especially happens when there are lots of new people 01:09:43.940 |
coming into the field and they get attracted to one thing. 01:09:46.860 |
And some people are trying to think different 01:09:50.660 |
So there is some, I think skepticism in science 01:10:01.540 |
it's mostly bandwagon effect that is happening 01:10:12.220 |
If you look at just companies, OpenAI, DeepMind, 01:10:15.220 |
Vicarious, I mean, there's a little bit of a race 01:10:31.940 |
So like, and the press is just irresponsible often. 01:10:41.380 |
Like, it seems like the people who write articles 01:10:43.580 |
about these things, they literally have not even spent 01:10:51.780 |
They haven't invested just even the language to laziness. 01:11:01.900 |
Like, they write this kind of stuff that just, 01:11:06.020 |
and then of course the researchers are quite sensitive 01:11:12.260 |
They're like, why did this word get so much attention? 01:11:15.020 |
That's over the top and people get really sensitive. 01:11:21.540 |
OpenAI did work with Rubik's Cube with the robot 01:11:37.340 |
and of course with your work, you mentioned deep learning, 01:11:40.260 |
but there's something super sexy to the public 01:11:45.540 |
I mean, that immediately grabs people's imagination. 01:11:48.440 |
Not even like neural networks, but like really brain-inspired. 01:11:57.500 |
That seems really compelling to people and to me as well, 01:12:12.300 |
in the research community and they're skeptical. 01:12:22.940 |
I mean, to me, all these datasets are useless anyway. 01:12:26.820 |
It's nice to have them, but in the grand scheme of things, 01:12:32.180 |
The point is, is there intuition about the ideas, 01:12:51.680 |
I don't treat brain-inspired as a marketing term. 01:13:35.380 |
that's the right, like you should constantly meditate 01:13:46.560 |
- Yes, you need to, so I think it's one input, 01:13:56.280 |
So an example is convolutional neural networks, right? 01:13:59.680 |
Convolution is not an operation brain implements. 01:14:10.180 |
local connectivity, but there is no translation 01:14:15.180 |
in invariance in the network weights in the visual cortex. 01:14:31.780 |
So, and that trick will be with us for some time. 01:14:48.740 |
- So the brain doesn't have translational invariance. 01:14:53.060 |
It has the focal point, like it has a thing it focuses on. 01:14:56.020 |
- Correct, it has a fovea, and because of the fovea, 01:14:59.480 |
the receptive fields are not like the copying of the weights, 01:15:04.140 |
like the weights in the center are very different 01:15:12.540 |
and just gotten a chance to really study peripheral vision, 01:15:24.060 |
at every level the brain does with the periphery. 01:15:29.740 |
So it's another kind of trick than convolutional. 01:15:39.900 |
convolution in neural networks is a trick for efficiency, 01:15:45.180 |
And the brain does a whole nother kind of thing, I guess. 01:15:51.180 |
of processing so that you can still apply engineering tricks 01:16:05.100 |
but it should be the point of really understanding 01:16:23.340 |
How is your just, if you could just give a brief history, 01:16:27.340 |
how is your view of the way the models of the brain changed 01:16:41.660 |
or is it all just building on top of each other? 01:16:46.140 |
especially the ones Jeff wrote about in the book, 01:16:53.380 |
If you blur out the details and if you just zoom out 01:17:09.260 |
multi-level, hierarchical, all of those things, right? 01:17:12.020 |
But in terms of the detail, a lot of things are different 01:17:29.900 |
how much of biological plausibility and realism 01:17:49.940 |
I did not want to be so constrained on saying, 01:17:54.820 |
my learning algorithms need to be biologically plausible 01:17:58.380 |
based on some filter of biological plausibility 01:18:08.060 |
discovering more and more things about the brain 01:18:15.500 |
So I don't want to upfront kill off a learning algorithm 01:18:22.620 |
the full biophysics or whatever of how the brain learns. 01:18:28.900 |
- But let me ask, and I'm sorry to interrupt, 01:18:30.900 |
like, what's your sense, what's our best understanding 01:18:36.620 |
- So things like back propagation, credit assignment, 01:18:43.340 |
learning algorithms have things in common, right? 01:18:45.540 |
It is, back propagation is one way of credit assignment. 01:18:49.380 |
There is another algorithm called expectation maximization, 01:18:52.660 |
which is, you know, another weight adjustment algorithm. 01:18:56.100 |
- But is it your sense the brain does something like this? 01:19:01.380 |
in the sense of saying that you do have to adjust 01:19:08.020 |
you have to reward the connections that were useful 01:19:12.620 |
yeah, I guess, but yeah, it doesn't have to be differentiable. 01:19:17.620 |
- Yeah, it doesn't have to be differentiable. 01:19:24.380 |
you have data comes in, and you have to have a way 01:19:27.700 |
of adjusting the model such that it better fits the data. 01:19:34.660 |
And some of them can be using backprop to do that. 01:19:45.500 |
That can, you know, many of these learning algorithms 01:19:52.820 |
in terms of what the neurons need to do locally. 01:19:57.340 |
- I wonder if small differences in learning algorithms 01:19:59.820 |
can have huge differences in the actual effect. 01:20:02.380 |
So the dynamics of, I mean, sort of the reverse, 01:20:07.220 |
like spiking, like if credit assignment is like a lightning 01:20:15.820 |
like whether there's like a looping local type of situation 01:20:29.460 |
like how it injects robustness into the whole thing, 01:20:34.460 |
like whether it's chemical or electrical or mechanical, 01:20:47.100 |
I feel like those differences could be essential, right? 01:20:53.980 |
on the learning side, you don't know enough to say 01:20:58.460 |
that is definitely not the way the brain does it. 01:21:04.660 |
So you've been open-minded on that side of things. 01:21:07.900 |
On the inference side, on the recognition side, 01:21:14.900 |
because it's like, okay, here's the stimulus. 01:21:17.620 |
How many steps did it get to take the answer? 01:21:21.260 |
I can understand the speed of that computation, et cetera, 01:21:28.740 |
And then you can't do good experiments on the learning side. 01:21:31.700 |
- So let's go right into cortical microcircuits right back. 01:21:40.500 |
beyond recursive cortical network that you're looking at now? 01:21:45.260 |
- So we have made a pass through multiple of the steps 01:21:52.700 |
we were looking at perception from the angle of cognition. 01:21:56.340 |
It was not just perception for perception's sake. 01:22:05.820 |
Similar to some of the things Francois talked about. 01:22:23.060 |
which has a system that learns dynamics of the world, 01:22:29.900 |
program learning system on top of it to learn concepts. 01:22:33.220 |
So we have built one, the version 0.1 of that system. 01:22:49.020 |
- And the application there was on manipulation, 01:22:56.100 |
Suppose you wanted to tell a new person that you met, 01:23:01.100 |
you don't know the language that person uses. 01:23:14.300 |
from the kitchen counter and put it here, right? 01:23:21.060 |
You can basically say, look, this is the starting state. 01:23:23.580 |
The things are here, this is the ending state. 01:23:27.060 |
And what does the person need to understand from that? 01:23:35.940 |
So we are looking at pre-verbal conceptual understanding. 01:23:40.940 |
Without language, how do you have a set of concepts 01:23:48.300 |
And from a set of images of input and output, 01:23:52.340 |
can you infer what is happening in those images? 01:23:55.980 |
- Got it, with concepts that are pre-language, okay. 01:23:59.140 |
So what's it mean for a concept to be pre-language? 01:24:07.500 |
- So I want to make a distinction between concepts 01:24:27.420 |
So those kinds of things you can extract purely from text. 01:24:32.660 |
But that's kind of a simple association thing 01:24:35.860 |
rather than a concept as an abstraction of something 01:24:39.060 |
that happens in the real world, in a grounded way, 01:24:52.180 |
concepts in the visual world are somehow lower level 01:24:58.940 |
- The lower level kind of makes it feel like, 01:25:02.660 |
Like, it's more like, I would say the concepts 01:25:17.660 |
just what we learn by interacting with the world 01:25:21.940 |
that is a prerequisite for any real language understanding. 01:25:29.260 |
'cause he says language is at the bottom of everything. 01:25:32.140 |
- No, yeah, I disagree with Chomsky completely 01:25:34.860 |
on so many levels, from universal grammar to, yeah. 01:25:47.180 |
the open problems in brain-inspired approaches 01:26:05.620 |
but the last thing that you will be actually solved. 01:26:08.620 |
Because if you do not build perception system 01:26:14.380 |
you cannot build concept system in the right way. 01:26:21.180 |
you have to still build that and learn concepts from there 01:26:26.460 |
And finally, perception will get solved fully 01:26:38.540 |
but then maybe on the concept side and like common sense 01:26:44.660 |
is there some intuition you can draw from the brain 01:26:59.060 |
and then ask you a question following that sentence. 01:27:01.260 |
This is a natural language processing problem, right? 01:27:06.100 |
I'm telling you, Sally pounded a nail on the ceiling. 01:27:24.660 |
it was kind of hard to imagine what the hell she was doing, 01:27:26.660 |
but I imagined a visual of the whole situation. 01:27:34.900 |
So here, I posed a question in natural language. 01:27:40.820 |
you got the answer from actually simulating the scene. 01:27:47.100 |
okay, was Sally standing on something while doing this? 01:27:50.380 |
Could she have been standing on a light bulb to do this? 01:27:54.620 |
I could ask more and more questions about this 01:28:02.380 |
Where is all that knowledge that you're accessing stored? 01:28:08.460 |
It was not just by reading text you got that knowledge. 01:28:17.500 |
and by the age of five, you have pretty much all of this, 01:28:27.140 |
such that it can be accessed through language. 01:28:34.300 |
almost serves as the query into the whole visual cortex 01:28:38.620 |
But I mean, is all reasoning kind of connected 01:28:47.460 |
You can still do a lot of it by quick associations 01:28:57.500 |
but I can easily create tricky situations for you 01:29:04.940 |
- So the figuring out how these concepts connect, 01:29:14.140 |
- One of the problems that we are working on. 01:29:15.700 |
And the way we are approaching that is basically saying, 01:29:21.540 |
so the takeaway is that language is simulation control 01:29:34.140 |
And so that's basically the way we are approaching it. 01:29:48.660 |
that puts all these things together into programs, 01:29:52.260 |
as abstractions that you can run and simulate. 01:29:55.420 |
And now we are taking the step of connecting it to language. 01:29:58.140 |
And it will be very simple examples initially. 01:30:05.060 |
but it will be grounded simulation-based language. 01:30:14.900 |
And it will be in some simple world initially 01:30:25.300 |
and run the right simulations to come up with the answer. 01:30:39.060 |
I think it's an interesting thought provoking 01:30:46.140 |
I think it's good for us to talk about the limits 01:31:12.140 |
And of course the text generation part of that 01:31:24.100 |
But of course the weaknesses are also pretty visible 01:31:29.820 |
it is not really carrying a world state around. 01:31:32.700 |
And, you know, sometimes you get sentences like, 01:31:36.180 |
I went up the hill to reach the valley or the thing. 01:31:39.380 |
You know, some completely incompatible statements. 01:31:43.260 |
Or when you're traveling from one place to the other, 01:31:46.260 |
it doesn't take into account the time of travel, 01:31:48.860 |
So those things I think will happen less in GPT-3 01:31:55.100 |
And so, and it can do even more longer distance coherence. 01:32:00.100 |
But it will still have the fundamental limitations 01:32:09.780 |
to find whether something is true in the world or not. 01:32:15.340 |
so it's taking a huge amount of text from the internet 01:32:23.860 |
something that's an approximation of a world model, 01:32:27.780 |
which essentially could be used for reasoning? 01:32:37.580 |
- Yeah, I mean, they will look more impressive than GPT-3. 01:33:00.500 |
second order Markov chains, third order Markov chains, 01:33:02.380 |
and saying that, okay, third order Markov chains 01:33:04.260 |
look better than first order Markov chains, right? 01:33:19.260 |
or more sophisticated structure in the model, 01:33:39.060 |
but just scaling it up is not going to give us AGI 01:33:44.060 |
or natural language understanding or meaning. 01:33:56.900 |
forces you to construct things that are very much like, 01:34:03.180 |
'cause the ideas of concepts and meaning is a spectrum. 01:34:07.180 |
So in order to form that kind of compression, 01:34:12.780 |
maybe it will be forced to figure out abstractions 01:34:18.900 |
which look awfully a lot like the kind of things 01:34:34.460 |
- The information is there behind the text, right? 01:34:38.740 |
- No, unless somebody has written down all the details 01:34:46.660 |
okay, it is easier to walk forward than backward, 01:34:50.460 |
that you have to open the door to go out of the thing, 01:34:55.380 |
unless all these things somebody has written down somewhere 01:35:05.380 |
- That's an argument that like text is a lot lower fidelity 01:35:17.580 |
- Well, in this case, pictures aren't really, 01:35:49.940 |
- So I wonder if there's some interactive element 01:36:03.140 |
so you're making a statement about the limitation of text. 01:36:36.420 |
whether that could store the information needed, 01:37:03.980 |
- You believe in the feedback mechanism, recursion. 01:37:24.180 |
I don't think transformers captures that family. 01:37:28.900 |
It is very good at statistical modeling of text 01:37:32.420 |
and it will become better and better with more data, 01:37:36.540 |
bigger models, but that is only going to get so far. 01:37:40.740 |
Finally, when you, so I had this joke on Twitter 01:37:44.780 |
saying that, "Hey, this is a model that has read 01:37:47.900 |
"all of quantum mechanics and theory of relativity 01:37:54.260 |
"or we are asking it to solve simple puzzles." 01:38:02.460 |
If it does, we'll ask the system to do experiments, 01:38:13.780 |
Those are the things that we want the system to do 01:38:20.100 |
- Like impressive demo, somebody generating a red button 01:38:27.220 |
like there's no dissing the usefulness of it. 01:38:30.060 |
- So I get, by the way, I'm playing a little bit 01:38:32.620 |
of a devil's advocate, so calm down internet. 01:39:03.580 |
even the current GPT-2 and 3 are so surprising. 01:39:15.140 |
And I, reinforcement, the fact that reinforcement learning 01:39:23.740 |
is quite surprising, given how nonlinear the space is, 01:39:30.860 |
that are at all reasonable, it's very surprising. 01:39:54.900 |
with those concepts and connect those concepts 01:40:05.580 |
in human language in this poetic way seems to make sense, 01:40:09.460 |
that that is what intelligence and reasoning are like. 01:40:12.140 |
I wonder if at the core of it, it could be much dumber. 01:40:23.780 |
- So I guess the recursion, the feedback mechanism, 01:40:27.660 |
that does seem to be a fundamental kind of thing. 01:40:45.180 |
which came out recently on how do you form episodic memories 01:40:52.260 |
And we haven't figured out all the connections of that 01:41:22.100 |
And this day is not going to happen ever again. 01:41:24.620 |
And that needs to be stored as just a stream of strings, 01:41:33.140 |
And then the question is about how do you take 01:41:37.420 |
that experience and connect it to the statistical part 01:41:40.660 |
How do you now say that, okay, I experienced this thing. 01:41:43.540 |
Now I want to be careful about similar situations. 01:41:47.460 |
And so you need to be able to index that similarity 01:41:56.420 |
the model of the world that you have learned. 01:41:59.180 |
Although the situation came from the episode, 01:42:08.340 |
as an indexing over the other model that you're building. 01:42:14.220 |
- So the memories remain and they're an index 01:42:20.260 |
into this, like the statistical thing that you formed. 01:42:28.900 |
So it's basically the idea is that the hippocampus 01:42:32.940 |
is just storing or sequencing in a set of pointers 01:42:42.380 |
And then whenever you want to reconstitute that memory 01:43:15.900 |
your cortex is doing inference over in the new situation. 01:43:36.940 |
okay, this is when it happened as a timeline. 01:44:00.980 |
of course affects what you're going to see next 01:44:11.820 |
Yeah, it does seem to be that that's what's happening. 01:44:17.660 |
it's interesting to think of how we actually do that. 01:44:30.420 |
and we will find ways of combining properties 01:44:41.580 |
Graph neural networks are kind of a merge between them. 01:44:46.660 |
So, but to me, the direction is pretty clear. 01:45:08.100 |
- Well, let me ask you, there's a guy named Elon Musk. 01:45:18.300 |
It's kind of an interface between your two loves. 01:45:31.300 |
with different conditions, more in the short term. 01:45:33.620 |
But there's also these sci-fi futuristic kinds of ideas 01:45:41.620 |
in a high bandwidth way with the brain, bi-directional. 01:45:56.540 |
And in fact, when I got interested in brains initially, 01:46:04.700 |
it was through a brain-computer interface talk 01:46:10.540 |
That's when I even started thinking about the problem. 01:46:13.140 |
So, it is definitely a fascinating research area 01:46:27.700 |
Even just the intermediate milestones they're pursuing, 01:46:30.940 |
which are very reasonable as far as I can see, 01:46:40.700 |
and being able to write things into the brain. 01:46:50.620 |
People losing limbs being able to control prosthetics, 01:46:54.620 |
quadriplegics being able to control something, 01:47:03.820 |
They're based on a different electrode array, 01:47:08.260 |
but trying to attack some of the same problems. 01:47:13.780 |
- Correct, surgically implanted electrodes, yeah. 01:47:16.580 |
So, yeah, I think of it as a very, very promising field, 01:47:27.660 |
it will advance the level of being able to communicate. 01:47:43.300 |
- So, like, being able to connect electrodes, 01:47:46.940 |
and not just thousands, but like millions to the brain. 01:47:54.620 |
what will happen to the brain with that, right? 01:48:02.220 |
the brain is quite, in terms of neuroplasticity, 01:48:14.660 |
- Exactly, and then what soup does this land us into? 01:48:17.460 |
- The kind of hallucinations you might get from this 01:48:27.100 |
It's interesting whether we need to be able to figure out 01:48:30.780 |
the basic protocol of the brain's communication schemes 01:48:35.140 |
in order to get them to, the machine and the brain to talk. 01:48:50.380 |
okay, attach electrodes to some part of the cortex, okay? 01:48:59.060 |
that part is not damaged, it was not used for anything. 01:49:10.540 |
And if you do it like that, then it is brain adapting to, 01:49:15.260 |
and of course, your external system is designed 01:49:18.580 |
Just like we design computers or mouse, keyboard, 01:49:26.860 |
So of course, that feedback system is designed 01:49:29.540 |
to be human compatible, but now it is not trying to record 01:49:50.340 |
- Just imagine, it's connecting it to Twitter 01:50:21.740 |
So that, the surgery part of it, biology part of it, 01:50:29.620 |
we often find after a long time in biology that, 01:50:56.660 |
just like, again, with Elon, just like colonizing Mars, 01:51:18.540 |
- The intermediate steps that they are taking 01:51:33.220 |
- Well, we've been talking about cognition a little bit, 01:51:47.340 |
of what it takes to create an intelligent reasoning being? 01:51:55.780 |
like the engineering perspective of intelligence? 01:52:00.180 |
but it doesn't on a day-to-day basis inform what we do, 01:52:08.180 |
the company name is connected to this idea of consciousness. 01:52:13.860 |
- Vicarious, so Vicarious is the company name. 01:52:20.180 |
At the first level, it is about modeling the world, 01:52:25.100 |
and it is internalizing the external actions. 01:52:33.060 |
And now, after having learned a lot about the world, 01:52:42.820 |
So you can run things vicariously, just in your brain. 01:52:47.180 |
And similarly, you can experience another person's thoughts 01:53:04.580 |
that you're using to model the external world 01:53:17.620 |
then that is what gives rise to consciousness, I think. 01:53:25.540 |
which is when the model feels like something, 01:53:44.100 |
but it feels like something to be that entity. 01:54:06.460 |
It seems like there's much greater cost of your decisions. 01:54:35.820 |
but goals aren't quite the same thing as our mortality. 01:54:40.820 |
It feels like first of all, humans don't have a goal. 01:54:47.540 |
And they just kind of create goals at different levels. 01:55:10.860 |
So it feels like that's an important part of cognition, 01:55:26.500 |
to come to the equation for an artificial system, 01:55:35.540 |
The problem with humans is that I can't clone you. 01:55:44.220 |
your experience that was stored in your brain, 01:55:49.660 |
all those will not be captured in the new clone. 01:56:03.820 |
is actually of fundamental importance for intelligence. 01:56:21.860 |
you could say that it doesn't feel like death 01:56:23.980 |
is a fundamental property of an intelligent system, 01:56:31.900 |
of an immortal intelligent being, we don't have those. 01:56:36.500 |
It's very possible that that is a fundamental property 01:56:42.540 |
of intelligence is a thing that has a deadline for itself. 01:56:49.860 |
Suppose you invent a way to freeze people for a long time. 01:57:03.300 |
- Well, no, you're still, it's not about time. 01:57:08.980 |
It's about the knowledge that it's temporary. 01:57:28.040 |
But, and that's why I'm not too worried about AI, 01:58:05.520 |
what kind of books, technical, fiction, philosophical, 01:58:19.040 |
that others read, maybe if you have three books 01:58:23.280 |
- Yeah, so I definitely liked Judea Pearl's book, 01:58:27.280 |
Probabilistic Reasoning and Intelligent Systems. 01:58:39.720 |
But throughout this book, Judea Pearl kind of sprinkles 01:58:43.160 |
his philosophical observations and he thinks about, 01:58:46.880 |
connects us to how the brain thinks and attentions 01:58:51.520 |
So that whole thing makes it more interesting to read. 01:59:00.840 |
Probabilistic Reasoning and Intelligent Systems. 01:59:12.520 |
it was the one in 2000, that one is really hard. 01:59:23.520 |
- Due calculus, yeah, it was pretty dense mathematically. 01:59:32.000 |
Probabilistic Reasoning and Intelligent Systems. 01:59:34.400 |
Another book I liked was one from Doug Hofstadter. 01:59:43.600 |
It was probably Hofstadter and Daniel Dennett together. 01:59:48.600 |
- Yeah, and I actually was, I bought that book. 01:59:55.040 |
But I couldn't get an electronic version of it, 01:59:58.440 |
which is annoying, 'cause you read everything on Kindle. 02:00:02.680 |
- So you have to actually purchase the physical. 02:00:05.080 |
It's like one of the only physical books I have, 02:00:07.200 |
'cause anyway, a lot of people recommended it highly, so. 02:00:13.800 |
recommend reading is, this is not a technical book. 02:00:28.520 |
and how it was, there are multiple books on this topic 02:00:49.040 |
People thought, oh, it is all about just powerful engines. 02:00:54.480 |
Just need to have powerful, lightweight engines. 02:01:20.160 |
- Do you draw any parallels between birds fly? 02:01:30.600 |
Do you see the same kind of thing with the brain 02:01:46.040 |
- So people in AI often use airplanes as an example of, 02:01:57.520 |
and the saying is, airplanes don't flap wings. 02:02:03.280 |
The funny thing and the ironic thing is that, 02:02:09.000 |
is something Wright Brothers found by observing birds. 02:02:30.600 |
propulsion is not the important problem to solve here. 02:02:56.240 |
I talk to undergraduate students all the time. 02:03:00.560 |
interested in understanding how the brain works. 02:03:03.720 |
Is there advice you would give them about their career, 02:03:12.520 |
should be taken with a pinch of salt, of course. 02:03:34.600 |
A better way to pursue it might be through computer science, 02:03:41.960 |
electrical engineering, machine learning, and AI. 02:03:44.360 |
And of course, you have to study up the neuroscience, 02:03:48.480 |
If you are more attracted by finding something intriguing, 02:03:54.080 |
discovering something intriguing about the brain, 02:03:56.600 |
then of course it is better to be an experimentalist. 02:04:00.320 |
So find that motivation, what are you intrigued by? 02:04:10.120 |
- And it's interesting to see which department, 02:04:13.640 |
if you're picking in terms of your education path, 02:04:17.520 |
whether to go with, at MIT it's brain and computer, 02:04:32.960 |
And actually, the brain folks, the neuroscience folks 02:04:36.400 |
are more and more now embracing of learning TensorFlow, 02:04:43.080 |
They see the power of trying to engineer ideas 02:04:56.640 |
So that might be the right department actually. 02:05:05.920 |
that Jeff Hawkins organized almost 10 years ago. 02:05:13.920 |
you should take if you want to understand the brain? 02:05:27.240 |
But I think it does have some of the right ingredients, 02:05:33.400 |
You learn about how you can construct circuits 02:05:50.960 |
to if you want to go to computer science or neuroscience, 02:05:55.080 |
- The downside, you're more likely to be forced to use MATLAB. 02:06:15.240 |
on developing good habits in software engineering. 02:06:20.840 |
And students can take that into their own hands, 02:06:26.000 |
I feel like everybody should learn to program, 02:06:34.560 |
'cause it empowers, it puts the data at your fingertips. 02:06:38.840 |
You can find all kinds of things in the data. 02:06:41.440 |
And then you can also, for the appropriate sciences, 02:07:04.520 |
one of the things about intelligence is it's goal-driven. 02:07:13.680 |
what's the goal that the brain is operating under? 02:07:15.960 |
What's the meaning of it all for us humans, in your view? 02:07:22.560 |
- The meaning of life is whatever you construct out of it. 02:07:36.800 |
Is there some useful aspect that you think about 02:07:45.240 |
and just the basic mechanisms of generating goals 02:07:48.240 |
in studying cognition in the brain that you think about? 02:07:54.400 |
Or is it just about, 'cause everything we've talked about, 02:08:14.040 |
Because it's basically being able to understand 02:08:19.360 |
such that you can pursue whatever goals you want, right? 02:08:23.080 |
- So the machinery of the world is really ultimately 02:08:27.920 |
The rest is just whatever the heck you wanna do, 02:08:41.120 |
I don't think there's a better way to end it. 02:09:21.560 |
It really is the best way to support this podcast. 02:09:24.280 |
If you enjoy this thing, subscribe on YouTube, 02:09:29.080 |
support it on Patreon, or connect with me on Twitter, 02:09:44.880 |
You have power over your mind, not outside events. 02:09:52.180 |
Thank you for listening and hope to see you next time.