back to indexDouglas Lenat: Cyc and the Quest to Solve Common Sense Reasoning in AI | Lex Fridman Podcast #221
Chapters
0:0 Introduction
1:11 What is Cyc?
9:17 How to form a knowledge base of the universe
19:43 How to train an AI knowledge base
24:4 Global consistency versus local consistency
48:25 Automated reasoning
54:5 Direct uses of AI and machine learning
66:43 The semantic web
77:16 Tools to help Cyc interpret data
86:26 The most beautiful idea about Cyc
92:25 Love and consciousness in AI
99:24 The greatness of Marvin Minsky
104:18 Is Cyc just a beautiful dream?
109:3 What is OpenCyc and how was it born?
114:53 The open source community and OpenCyc
125:20 The inference problem
127:3 Cyc's programming language
134:37 Ontological engineering
142:2 Do machines think?
150:47 Death and consciousness
160:48 What would you say to AI?
165:24 Advice to young people
167:20 Mortality
00:00:00.000 |
The following is a conversation with Doug Lennett, 00:00:03.200 |
creator of Psyche, a system that for close to 40 years 00:00:07.560 |
and still today has sought to solve the core problem 00:00:22.400 |
please check out our sponsors in the description. 00:00:25.280 |
As a side note, let me say that in the excitement 00:00:30.160 |
it is easy to forget just how little we understand 00:00:33.480 |
exactly how to build the kind of intelligence 00:00:52.960 |
solving this problem of common sense knowledge 00:00:55.620 |
will help us humans understand our own minds, 00:01:01.180 |
how to be more rational and more kind to each other. 00:01:07.400 |
and here is my conversation with Doug Lennett. 00:01:24.060 |
In other words, it hopes to capture common sense knowledge, 00:01:31.780 |
and maybe perhaps speak to the various sub-goals 00:01:39.460 |
in the computer science department at Stanford, 00:01:44.860 |
in all sorts of artificial intelligence programs. 00:01:56.920 |
And we kept hitting the very same brick wall. 00:02:00.820 |
Our systems would have impressive early successes. 00:02:09.720 |
namely to get enough material to write a journal article, 00:02:19.340 |
then you have to somehow get past the brick wall. 00:02:23.700 |
the programs didn't have what we would call common sense. 00:02:29.560 |
They didn't really understand what they were doing, 00:02:32.500 |
what they were saying, what they were being asked. 00:02:35.100 |
And so very much like a clever dog performing tricks, 00:02:42.420 |
but they never really understood what they were doing. 00:02:57.740 |
Can you maybe elaborate on that a little bit? 00:03:05.940 |
or is understanding the wisdom you gain over time 00:03:12.660 |
think of it more like the ground you stand on, 00:03:42.860 |
And I've never used this analogy before, so bear with me, 00:03:49.420 |
in terms of getting computers to understand things, 00:03:56.100 |
for instance, Alexa or some robot or something, 00:04:19.540 |
and they keep asking why questions in series. 00:04:24.180 |
where you throw up your hands and say, I don't know, 00:04:30.300 |
you actually have that layered, solid foundation of support 00:04:35.300 |
so that when you need it, you can count on it. 00:04:42.100 |
when you come up against a situation which is novel, 00:04:51.860 |
a small set of rules that cover 99% of the cases, 00:04:56.380 |
but that 1% of the time when something strange happens, 00:05:02.060 |
For instance, my wife and I were driving recently 00:05:11.820 |
and the back exploded and trash bags went everywhere. 00:05:32.860 |
and we know what we throw away in trash bags, 00:05:37.380 |
Over on the left was a bunch of fast food restaurant 00:05:44.300 |
those things are just like styrofoam and leftover food. 00:05:51.100 |
Now that's the kind of thing that's going to happen 00:05:54.860 |
But the point is that there's almost no telling 00:06:00.500 |
what little bits of knowledge about the world 00:06:29.380 |
It's not merely, it's not enough to just have a lot 00:06:37.980 |
to utilize it in a split, like integrate it all together 00:06:43.020 |
And I suppose understanding isn't just having 00:07:08.340 |
- So let's tease apart two different tasks really, 00:07:11.660 |
both of which are incredibly important and even necessary 00:07:15.580 |
if you're going to have this in a useful, usable fashion 00:07:20.260 |
as opposed to say like library books sitting on a shelf 00:07:23.540 |
and so on where the knowledge might be there, 00:07:26.380 |
but if a fire comes, the books are gonna burn 00:07:32.060 |
and they're just gonna sit there while they burn. 00:07:34.940 |
So there are two aspects of using the knowledge. 00:07:39.220 |
One is a kind of a theoretical, how is it possible at all? 00:07:44.220 |
And then the second aspect of what you said is, 00:07:58.460 |
and even earlier developed a kind of formal language, 00:08:09.780 |
or first order logic or something like predicate calculus 00:08:24.340 |
to sort of grind through and algorithmically produce 00:08:31.900 |
all the same logical conclusions that you or I would 00:08:57.380 |
run these algorithms to actually get the information 00:09:00.400 |
you need in the case I mentioned in a 10th of a second, 00:09:04.260 |
rather than say in 10 hours or 10,000 years of computation? 00:09:09.260 |
And those are both really important questions. 00:09:13.660 |
- And like a corollary addition to the first one is, 00:09:24.620 |
So like what, in order, you mentioned philosophers, 00:09:27.540 |
in order to capture this world and represent it 00:09:55.080 |
- I'm gonna, I guess I'm gonna disappoint you 00:09:58.140 |
by giving you an actual answer to your question. 00:10:04.900 |
So now we have like three things to talk about. 00:10:10.820 |
- Although it's okay, the first and the third are related. 00:10:13.700 |
So let's leave the efficiency question aside for now. 00:10:18.260 |
So how does all this information get represented 00:10:27.360 |
resolution theorem proving and other algorithms 00:10:29.760 |
can actually grind through all the logical consequences 00:10:42.460 |
then by hand, you could write them out one at a time. 00:11:17.780 |
he commented about your system, Eurysco, at the time. 00:11:41.420 |
if it's important to represent common sense knowledge 00:11:44.060 |
and world knowledge in order for AIs to not be brittle, 00:12:07.780 |
And I expected there would be an enormous divergence 00:12:31.800 |
"like maybe one every 30 seconds or something. 00:12:34.540 |
"And other than that, it's just short-term memory 00:12:39.260 |
So by the time you're, say, 10 years old or so, 00:12:42.620 |
how many things could you possibly have burned 00:12:48.460 |
Another person went in a completely different direction 00:12:51.180 |
and said, "Well, if you look at the number of words 00:12:57.880 |
"but for someone to essentially be considered 00:13:05.200 |
"And then about how many things about each word 00:13:11.760 |
Another person said, "Well, let's actually look 00:13:16.300 |
"at one single short, one-volume desk encyclopedia article." 00:13:34.300 |
and represent every single thing that was there, 00:13:41.340 |
would we have to write in this logical language and so on? 00:13:43.820 |
And then multiply that by all of the number of articles 00:13:48.820 |
So all of these estimates came out with a million. 00:13:51.620 |
And so if you do the math, it turns out that like, 00:13:55.740 |
oh, well then maybe in something like 100 person years, 00:14:06.280 |
we could actually get this written down by hand. 00:14:10.020 |
And a marvelous coincidence, opportunity existed 00:14:15.020 |
right at that point in time, the early 1980s. 00:14:20.940 |
the Japanese fifth-generation computing effort. 00:14:23.860 |
Japan had threatened to do in computing and AI and hardware 00:14:28.860 |
what they had just finished doing in consumer electronics 00:14:33.300 |
namely resting control away from the United States 00:14:39.220 |
And so America was scared and Congress did something. 00:14:55.460 |
that's also how you know it was a long time ago 00:15:24.340 |
both of them coincidentally in Austin, Texas. 00:15:27.340 |
One called Sematech focusing on hardware chips and so on, 00:15:33.400 |
the Microelectronics and Computer Technology Corporation 00:15:36.540 |
focusing more on software, on databases and AI 00:15:40.780 |
and natural language understanding and things like that. 00:16:03.100 |
"Look, Professor, you're talking about doing this project. 00:16:05.860 |
"It's gonna involve person centuries of effort. 00:16:10.080 |
"You've only got a handful of graduate students. 00:16:12.220 |
"You do the math, it's gonna take you like, you know, 00:16:16.160 |
"longer than the rest of your life to finish this project. 00:16:18.840 |
"But if you move to the wilds of Austin, Texas, 00:16:23.640 |
"and you know, you'll be done in a few years." 00:16:42.740 |
That it turns out what you need are tens of millions 00:16:49.820 |
sort of like if you have a coffee cup with stuff in it 00:17:13.760 |
and then later US government agency funding and so on, 00:17:29.340 |
getting Psych to contain the tens of millions of rules 00:17:42.620 |
the things you count on other people knowing. 00:17:49.700 |
And the good news is since you've waited 38 years 00:17:59.620 |
So most of what we're doing now is not putting in 00:18:03.740 |
but more putting in domain specific applications, 00:18:07.420 |
specific knowledge about healthcare in a certain hospital 00:18:24.340 |
and we're doing things very much like the expert systems 00:18:29.580 |
except instead of resting on nothing and being brittle, 00:18:32.900 |
they're now resting on this massive pyramid, if you will, 00:18:36.300 |
this massive lattice of common sense knowledge 00:18:50.420 |
for instance, if we have a problem with the microphone, 00:18:55.940 |
plug it in again and hope for the best, right? 00:18:58.420 |
Because that's one of the general pieces of knowledge 00:19:00.480 |
you have in dealing with electronic equipment 00:19:09.680 |
that generally captures this idea of turn it off 00:19:19.540 |
- That's actually one of the fundamental laws 00:19:34.300 |
- So that basically covered the resources needed 00:19:39.300 |
and then we had to devise a method to actually figure out, 00:19:43.180 |
well, what are the tens of millions of things 00:19:56.620 |
it could be an advertisement, it could be a transcript, 00:20:01.800 |
and don't pay attention to the actual type that's there, 00:20:11.860 |
So what did the writer of this sentence assume 00:20:15.700 |
that the reader already knew about the world? 00:20:22.100 |
why did they think that you would be able to understand 00:20:25.800 |
what the intended referent of that pronoun was? 00:20:30.200 |
how did they think that you would be able to figure out 00:20:40.660 |
What are all the things that the writer expected you 00:20:43.200 |
to fill in and infer occurred between the end 00:20:46.260 |
of one sentence and the beginning of the other? 00:20:49.480 |
Fred Smith robbed the Third National Bank, period. 00:20:54.200 |
He was sentenced to 20 years in prison, period. 00:20:58.980 |
Well, between the first sentence and the second, 00:21:01.500 |
you're expected to infer things like Fred got caught, 00:21:08.000 |
Fred had a trial, Fred was found guilty, and so on. 00:21:11.780 |
If my next sentence starts out with something like 00:21:18.420 |
If my next sentence starts out something like 00:21:24.580 |
who arrested him after he committed the crime and so on. 00:21:26.980 |
So those are two techniques for getting that knowledge. 00:21:35.660 |
sort of like fake news or sort of humorous onion headlines 00:21:44.620 |
if you know what that is, or the National Enquirer, 00:21:49.460 |
then we introspect on why don't we believe it. 00:21:51.620 |
So there are things like B-17 lands on the Moon. 00:21:58.980 |
that causes us to believe that that's just silly 00:22:03.300 |
Or another thing we look for are contradictions, 00:22:17.140 |
For instance, in one of the Weekly World News editions, 00:22:21.380 |
in one article, it talked about how Elvis was cited, 00:22:24.820 |
even though he was getting on in years and so on. 00:22:36.300 |
that at least one of these articles must be wrong and so on? 00:22:47.500 |
working full-time on this and have for decades. 00:22:50.460 |
So we've put in the thousands of person years of effort, 00:22:53.240 |
we've built up these tens of millions of rules. 00:23:00.100 |
that we're saying things as generally as we possibly can. 00:23:13.300 |
then you'd have to add another one or two or three zeros 00:23:16.620 |
onto the number of assertions you'd actually have to have. 00:23:20.480 |
So at some point, we generalize things more and more, 00:23:34.940 |
A member of one is not gonna be a member of the other 00:23:38.580 |
- And the same thing with the Elvis and the ghost, 00:23:41.940 |
it's more about human nature and the mortality 00:23:46.660 |
- Well, in general, things are not both alive and dead 00:23:50.460 |
- Yeah, unless special cats in theoretical physics examples. 00:23:55.460 |
- Well, that raises a couple important points. 00:23:58.340 |
- Well, that's the onion headline situation type of thing. 00:24:10.980 |
And one of the hardest lessons for us to learn, 00:24:14.020 |
it took us about five years to really grit our teeth 00:24:23.460 |
So the knowledge base can no longer be consistent. 00:24:32.940 |
it would either freeze up or explode or take over the world 00:24:38.900 |
Or if you come from a mathematics background, 00:24:41.260 |
once you can prove false, you can prove anything, 00:24:46.380 |
So that's why the old knowledge-based systems 00:25:05.820 |
to build some enormous globally consistent knowledge base. 00:25:24.960 |
as though the surface of the Earth were flat. 00:25:27.680 |
You know, when you're talking to someone in Australia, 00:25:29.600 |
you don't think of them as being oriented upside down to you 00:25:38.660 |
but you rarely think about the curvature of the Earth 00:25:41.820 |
And for most purposes, you can live your whole life 00:26:08.180 |
And so by the time you move, say, 20 contexts over, 00:26:17.300 |
modern real world context that we're in right now 00:26:26.220 |
and in fact, life and death are very different 00:26:28.820 |
because no matter how many times he's killed, 00:26:31.260 |
you know, the coyote comes back in the next scene 00:26:38.180 |
and we had to make sure that our representation language, 00:26:41.260 |
the way that we actually encode the knowledge 00:26:45.180 |
was expressive enough that we could talk about 00:26:47.300 |
things being true in one context and false in another, 00:26:51.020 |
things that are true at one time and false in another, 00:26:53.740 |
things that are true, let's say, in one region, 00:26:58.660 |
things that are true in one person's belief system, 00:27:04.780 |
things that are true at one level of abstraction 00:27:29.580 |
that do this kind of capture of knowledge bases, 00:27:33.360 |
or do you then get turtles on top of turtles, 00:27:36.140 |
again, where there's just a huge number of contexts? 00:27:41.260 |
'cause you're pointed in the right direction, 00:27:43.820 |
which is you want contexts to be first-class objects 00:28:09.540 |
And so, contexts are just terms in its language, 00:28:18.580 |
context can arrange hierarchically and so on. 00:28:25.060 |
let's say, things that are true in the modern era, 00:28:32.460 |
would then be a sub-context of the things that are true 00:28:37.340 |
in a broad, let's say, a century or a millennium 00:28:51.620 |
that are true in the United States and so on. 00:28:54.260 |
And so, you don't have to say things over and over again 00:29:00.220 |
at the most general level that it applies to, 00:29:15.540 |
- Oh, you definitely have to think of it as a graph. 00:29:20.480 |
why the Japanese fifth-generation computing effort failed. 00:29:23.780 |
There were about half a dozen different reasons. 00:29:42.460 |
So, if you had a table that was a wooden object, 00:29:56.540 |
sometimes it's important to know that it's made out of wood, 00:30:01.040 |
Sometimes it's important to know that it's flat 00:30:03.220 |
if we're talking about resting something on it, and so on. 00:30:16.420 |
each node could only have, at most, 10 children, 00:30:23.500 |
And while that does enable the Dewey decimal type 00:30:32.220 |
it prevents you from representing all the things 00:30:45.360 |
- So, we'll return to some of the doors you've opened, 00:30:50.780 |
around there with Marvin Minsky and Stanford. 00:30:56.180 |
wouldn't do his estimate until someone brought him 00:30:59.420 |
an envelope so that he could literally do a back 00:31:02.460 |
of the envelope calculation to come up with his number. 00:31:17.500 |
A few people get together and plant the seed of ideas, 00:31:42.380 |
in the long arc of history, the Drake equation 00:31:49.860 |
In that same way, a conversation of just how many facts 00:31:54.620 |
are required to capture the basic common sense knowledge 00:31:58.860 |
- I want to distinguish between what you think of as facts 00:32:10.780 |
read and access the kind of facts you might find, 00:32:13.900 |
say, in Wikidata, or stated in a Wikipedia article, 00:32:20.260 |
So what we're representing, the things that we need 00:32:24.580 |
are more like rules of thumb, rules of good guessing, 00:32:30.060 |
and which help you to make sense of the facts 00:32:34.980 |
that are sort of sitting off in some database 00:32:45.260 |
that's going to be projections of those ideas. 00:32:47.420 |
You read an article about the fact that Elvis died, 00:32:50.300 |
that's a projection of the idea that humans are mortal. 00:32:54.040 |
Very few Wikipedia articles will write humans are mortal. 00:33:00.180 |
- Exactly, that's what I meant about ferreting out 00:33:16.420 |
or people can't be both alive and dead at the same time, 00:33:23.540 |
If you search online for water flowing uphill 00:33:27.000 |
you'll find more references for water flowing uphill, 00:33:29.660 |
because it's used as a kind of a metaphorical reference 00:34:01.660 |
assert has a very dogmatic authoritarian feel to them. 00:34:25.760 |
we can also go back a couple of decades before then 00:34:29.740 |
and thinking about artificial intelligence broadly, 00:34:39.460 |
And I think people's intuition was off at the time. 00:34:48.940 |
when we're grappling with these exceptionally difficult ideas 00:34:53.300 |
it's very difficult to truly understand ourselves 00:34:59.340 |
to introspect how difficult it is to engineer intelligence, 00:35:12.480 |
Do you, what's your sense from the 1984 to today, 00:35:23.900 |
So you've kind of said with some level of certainty 00:35:40.220 |
to realize that we were off by a factor of 10. 00:35:46.420 |
Marvin Misk is very confident in the '60s when you say. 00:36:05.400 |
but your brain will still be in the digital form, 00:36:24.380 |
which is sometimes like if you have an old fashioned pump, 00:36:32.700 |
So I think I'm right enough in the sense that-- 00:36:38.880 |
even if it isn't, so to speak, everything you need, 00:37:03.160 |
So through a combination of psych-powered discovery 00:37:12.560 |
Maybe it's the final 2%, maybe it's the final 99%. 00:37:31.360 |
or maybe even exponentially increasing path upward 00:37:36.360 |
that we get from, for instance, talking with each other. 00:38:03.780 |
because you've already gotten to a certain level 00:38:19.080 |
And I think this is really an important point, 00:38:33.960 |
It's not so much that it'll be us versus the AIs. 00:38:39.960 |
will be able to do things that require more creativity 00:38:45.500 |
but we'll be able to do lots of things in parallel. 00:38:48.200 |
We'll be able to misunderstand each other less. 00:38:55.640 |
for an individual would mean that individual will, 00:39:01.280 |
And that means that humanity as a species will be smarter. 00:39:05.480 |
And when was the last time that any invention 00:39:08.840 |
qualitatively made a huge difference in human intelligence? 00:39:19.640 |
It was all the way back to the development of language. 00:39:24.440 |
We sort of look back on pre-linguistic cavemen as, 00:39:28.440 |
well, they weren't really intelligent, were they? 00:39:34.840 |
And I think that, as you said, 50, 100, 200 years from now, 00:39:42.560 |
right before the advent of the sort of lifelong 00:39:47.920 |
general AI muses and say, "Those poor people, 00:39:57.240 |
So you said a lot of really interesting things. 00:39:59.200 |
By the way, I would maybe try to argue that the internet 00:40:04.200 |
is on the order of the kind of big leap in improvement 00:40:14.680 |
- Well, certainly a big leap in one direction. 00:40:16.560 |
We're not sure whether it's upward or downward. 00:40:18.440 |
- Well, I mean very specific parts of the internet, 00:40:25.000 |
like ability for human beings from across the world 00:40:29.840 |
- So I could take either side of this argument, 00:40:35.240 |
which is that almost nothing has done more harm 00:40:45.440 |
One is it's made people more globally ignorant 00:40:56.900 |
So when I was growing up, we had to use slide rules, 00:41:02.880 |
Today, people don't really understand numbers, 00:41:08.120 |
they don't really estimate very well at all, and so on. 00:41:21.200 |
And thanks to things like the internet and search engines, 00:41:35.080 |
not just without being able to do arithmetic and estimate, 00:41:48.840 |
You can, of course, say the same thing about language. 00:41:50.920 |
Probably people, when they invented language, 00:41:58.480 |
and if we're in love, we would just have sex with them, 00:42:00.720 |
and now everybody's writing poetry and bullshit. 00:42:33.120 |
And overall, I think that having smarter individuals 00:42:38.760 |
and having smarter AI augmented human species 00:42:49.640 |
involving poverty and starvation and global warming 00:43:02.520 |
and there are really only two routes to being smarter. 00:43:12.600 |
The other route is through having general AIs 00:43:52.280 |
It's not like a battle of humans versus virus. 00:43:56.880 |
And then you can just wipe out an entire species. 00:44:02.360 |
is that it has enabled us to surround ourselves 00:44:07.120 |
with an echo chamber, with a bubble of like-minded people, 00:44:12.120 |
which means that you can have truly bizarre theories, 00:44:17.120 |
conspiracy theories, fake news, and so on, promulgate, 00:44:21.160 |
and surround yourself with people who essentially reinforce 00:44:30.440 |
And in the old days, that was much harder to do 00:44:44.360 |
- I like the push and pull of our dance that we're doing 00:44:46.840 |
because then I'll just say, in the old world, 00:44:55.040 |
and then the government can overpower its people 00:44:57.520 |
by telling you the truth and then starving millions 00:45:09.480 |
that you're actually doing good in the world. 00:45:12.840 |
because of all the quote-unquote conspiracy theories, 00:45:16.240 |
some of them are actually challenging the power centers, 00:45:18.920 |
the very kind of power centers that a century ago 00:45:27.960 |
And I very much agree with you on the AI side. 00:45:33.560 |
that somehow AI will be used to maybe overpower people 00:45:45.680 |
especially just having observed the trajectory of technology, 00:45:58.320 |
'cause there's a lot of money to be made that way. 00:46:00.920 |
Like improving people's lives, you can make a lot of money. 00:46:04.120 |
I think that the main thing that AI prostheses, 00:46:09.120 |
AI amplifiers will do for people is make it easier, 00:46:25.560 |
in things that they otherwise would just blithely believe. 00:46:46.600 |
in the sense of pouring facts into people's heads, 00:46:53.120 |
with the ability to do good critical thinking 00:46:58.640 |
The education system that we have in the US and worldwide 00:47:12.320 |
In the same way that everyone can have their own Alexa 00:47:20.120 |
everyone will have this sort of cradle to grave assistant 00:47:25.640 |
which will get to know you, which you'll get to trust. 00:47:33.600 |
which will, in some sense, make your life better, easier, 00:47:38.040 |
less mistake-ridden, and so on, less regret-ridden, 00:47:53.120 |
From my perspective, integrating emotional intelligence, 00:47:57.080 |
so even things like friendship and companionship and love 00:48:02.760 |
as opposed to helping you just grow intellectually 00:48:06.160 |
as a human being, allow you to grow emotionally, 00:48:12.040 |
is to sort of, you know, the old pursuit of happiness. 00:48:25.440 |
Let me jump back to the idea of automated reasoning. 00:48:44.960 |
carefully illuminating manuscripts and so on. 00:48:47.960 |
- It's a very difficult and amazing process, actually, 00:48:50.920 |
because it allows you to truly ask the question 00:49:21.080 |
it would almost be offensive or confusing to the reader. 00:49:24.880 |
It's like, why are they telling me all these? 00:49:34.440 |
has almost never been explicitly written down anywhere, 00:49:47.800 |
presumably you already got pieces of common sense, 00:50:02.680 |
what ideas are there for how to do it automatically 00:50:13.400 |
of the work for humans, and then humans only do 00:50:18.720 |
- So, we have, in fact, two directions we're pushing on 00:50:25.880 |
and one involves natural language understanding 00:50:38.160 |
of knowledge-editing tools, knowledge-entry tools, 00:50:42.560 |
knowledge-capture tools, knowledge-testing tools, 00:50:49.280 |
Think of them as, like, user-interfaced suite 00:50:56.880 |
more or less, automatically expand and extend the system 00:51:01.200 |
in areas where, for instance, they want to build some app, 00:51:05.040 |
have it do some application, or something like that. 00:51:13.400 |
So, you've probably heard of, like, deduction, 00:51:41.160 |
But if someone were to walk into the room right now, 00:51:48.880 |
oh, did it start to rain, or something like that. 00:51:51.360 |
Now, why did we say, maybe it started to rain? 00:51:57.800 |
but it's certainly a reasonable abductive leap to say, 00:52:08.640 |
is if they had gotten caught out in the rain, 00:52:16.720 |
So, suppose you're building one of these applications, 00:52:22.480 |
And you say, oh, yeah, the answer to this question 00:52:31.960 |
everything it already knows about common sense, 00:52:52.660 |
I would have come up with the answer you just gave me 00:52:59.080 |
And then, you, the expert, will look at those seven things 00:53:02.960 |
and say, oh, yeah, number five is actually true. 00:53:05.560 |
And so, without actually having to tinker down 00:53:08.540 |
at the level of logical assertions and so on, 00:53:15.280 |
in the same way that you would help educate another person 00:53:18.560 |
who you were trying to apprentice or something like that. 00:53:22.040 |
- So, that significantly reduces the mental effort 00:53:38.920 |
And then, the other is that the system on its own 00:53:45.080 |
through conversations with other people and so on, 00:53:49.240 |
learn the same way that you or I or other humans do. 00:54:00.720 |
are there, when we talk about specific techniques, 00:54:05.960 |
do you find something inspiring or directly useful 00:54:09.320 |
from the whole space of machine learning, deep learning, 00:54:14.140 |
that have been shown effective for certain kinds 00:54:17.040 |
of problems in the recent, now, decade and a half? 00:54:24.040 |
as more or less what our right brain hemispheres do. 00:54:34.480 |
and recognize patterns, being able to statistically 00:54:53.600 |
and puff on its pipe and think about this thing over here. 00:55:02.400 |
How should I feel about that and why and so on? 00:55:12.740 |
Whereas you want machine learning to think quickly, 00:55:20.660 |
So, I'll give you an example of a project we did recently 00:55:26.780 |
and a couple other institutions that we ran a project for. 00:55:44.940 |
They got their DNA sequenced because the cost of doing that 00:55:48.860 |
has gone from infinity to billions of dollars 00:55:54.660 |
And so now patients routinely get their DNA sequenced. 00:55:58.140 |
So, you have these big databases of the SNPs, 00:56:17.380 |
are associated with and led to which physiological problems 00:56:22.940 |
and diseases and so on, like getting arthritis and so on. 00:56:41.860 |
or the bad ones are focusing on the good ones. 00:56:47.140 |
which is Psyche takes those sort of A to Z correlations 00:56:50.740 |
between point mutations and the medical condition 00:56:56.120 |
And we say, okay, let's use all this public knowledge 00:57:00.340 |
and common sense knowledge about what reactions occur 00:57:04.420 |
where in the human body, what polymerizes what, 00:57:10.460 |
And let's try to put together a 10 or 20 or 30 step 00:57:22.140 |
And so Psyche would put together in some sense 00:57:24.580 |
some Rube Goldberg-like chain that would say, 00:57:32.340 |
would be this altered protein, which because of that, 00:57:40.180 |
And by the way, that would cause more bioactive vitamin D 00:57:44.400 |
And anyway, 10 steps later, that screws up bone resorption 00:57:56.660 |
And the important thing, even more than that, 00:58:08.540 |
that initial A to Z correlation, except two things. 00:58:11.660 |
One, if you can't even think of one causal chain 00:58:15.980 |
to explain this, then that correlation probably 00:58:24.300 |
along the way, that causal chain will make predictions, 00:58:28.340 |
like the one about having more bioactive vitamin D 00:58:32.080 |
So you can now go back to the data about these patients 00:58:35.540 |
and say, by the way, did they have slightly elevated levels 00:58:40.100 |
of bioactive vitamin D in their blood and so on? 00:58:42.780 |
And if the answer is no, that strongly disconfirms 00:58:48.140 |
And if the answer is yes, that somewhat confirms 00:58:51.860 |
And so using that, we were able to take these correlations 00:59:03.380 |
focus the researcher's attention on the very small percentage 00:59:11.340 |
and even better, some explanation that also made 00:59:13.560 |
some independent prediction that they could confirm 00:59:20.260 |
where you want the right brain machine learning 00:59:25.700 |
You want the left brain psych-like AI to think about that, 00:59:30.700 |
and now think about why that might have been the case, 00:59:34.540 |
and what else would be the case if that were true, 00:59:36.480 |
and so on, and then suggest things back to the right brain 00:59:45.540 |
which I think is really what's gonna lead to general AI, 00:59:49.240 |
not narrow, brittle machine learning systems, 00:59:57.620 |
but I was also thinking in terms of the automated expansion 01:00:03.260 |
This is very early days in the machine learning space 01:00:07.560 |
of this, but self-supervised learning methods, 01:00:10.580 |
you know, you have these language models, GPT-3, and so on, 01:00:14.220 |
that just read the internet, and they form representations 01:00:23.760 |
Like, they're now playing with a pretty cool thing 01:00:26.000 |
called OpenAI Codex, which is generating programs 01:00:32.600 |
But my question is, can it be used to generate, 01:00:40.240 |
a psych-like assertions, help feed psych more assertions 01:00:53.600 |
how can we harness natural language processing 01:00:56.600 |
to increasingly automate the knowledge acquisition process, 01:01:04.560 |
that if you sort of learn things at the fringe 01:01:09.240 |
of what you know already, you learn this new thing 01:01:13.680 |
and here are the differences in the new things 01:01:17.080 |
So the more you know, the more and more easily 01:01:28.520 |
And so if you're not careful, if you start out 01:01:31.520 |
with too small sort of a core to start this process, 01:01:38.560 |
And so that's why I view this as a pump priming exercise, 01:01:44.340 |
even though that's kind of ugly duckling technique, 01:01:47.080 |
put in the elbow grease to produce a large enough core 01:01:50.880 |
that you will be able to do all the kinds of things 01:02:13.040 |
putting a poison, plotting to poison someone, and so on. 01:02:24.780 |
that person then drank the poison they just put together. 01:02:27.180 |
It's like, that's probably not what happened for someone. 01:02:29.540 |
Or if you go to Siri and, you know, I think I have, 01:02:40.920 |
It'll come back and say, I found seven liquor stores 01:02:47.080 |
So, you know, it's one of these things where, yes, 01:02:56.160 |
but if it doesn't really understand what it's saying, 01:02:59.380 |
and if it doesn't really understand why things are true 01:03:01.780 |
and doesn't really understand how the world works, 01:03:04.460 |
then some fraction of the time, it's gonna be wrong. 01:03:07.500 |
Now, if your only goal is to sort of find relevant 01:03:14.540 |
then being right 90% of the time is fantastic. 01:03:19.340 |
Okay, however, if your goal is to like, you know, 01:03:22.500 |
save the life of your child who has some medical problem, 01:03:26.080 |
or your goal is to be able to drive, you know, 01:03:31.920 |
without getting into a fatal accident and so on, 01:03:34.840 |
then, you know, error rates down at the 10% level 01:03:38.980 |
or even the 1% level are not really acceptable. 01:03:47.020 |
and then you kind of think of knowledge as this sphere. 01:03:50.100 |
So, you want a large sphere because the learning 01:03:56.460 |
- Exactly, so you have, what you can learn next 01:04:05.180 |
- It's nice because you think when you know nothing, 01:04:17.340 |
So, I'll also, one of the anecdotes I could go back 01:04:41.220 |
and doctors wanted to do all these scary things. 01:04:49.320 |
and we could not get a meaningful answer from her doctors 01:04:59.700 |
And fortunately, a friend of mine, Ted Shortliff, 01:05:03.200 |
was another assistant professor in computer science 01:05:08.360 |
and he'd been building a program called Mycin, 01:05:13.840 |
that happened to specialize in blood infections 01:05:20.800 |
at Stanford Hospital because he was also an MD. 01:05:23.760 |
And so, we got hold of her chart and we put in her case, 01:05:27.840 |
and it came up with exactly the same diagnoses 01:05:30.640 |
and exactly the same therapy recommendations, 01:05:39.320 |
step by step by step why this was the diagnosis 01:05:44.320 |
and step by step why this was the best therapy, 01:05:55.960 |
because that made all the difference in the world. 01:05:58.200 |
Instead of blindly having to trust in authority, 01:06:01.200 |
we were able to understand what was actually going on. 01:06:05.280 |
And so, at that time, I realized that that really 01:06:16.040 |
the way the world works and why things are the way they are, 01:06:20.280 |
they weren't able to give explanations of their answer. 01:06:23.620 |
And it's one thing to use a machine learning system 01:06:28.060 |
that says, "I think you should get this operation," 01:06:32.880 |
And it says, "0.83," and you say, "No, in more detail, why?" 01:06:50.120 |
It was the obvious next step for the internet. 01:06:53.080 |
- And maybe you can speak about what is the semantic web, 01:07:15.740 |
which is sort of like a node and link diagram. 01:07:21.040 |
So you have these nodes that represent concepts 01:07:25.060 |
or words or terms, and then there are some arcs 01:07:36.360 |
that represents one person and let's say a husband link 01:07:55.640 |
So having this kind of representation is really good 01:08:20.120 |
or something like that, you can represent that 01:08:22.620 |
very nicely using these kinds of graph structures 01:08:27.620 |
or using something like the semantic web and so on. 01:08:48.340 |
So for instance, if you've read or seen "Romeo and Juliet," 01:09:00.780 |
that put her into a kind of suspended animation? 01:09:09.620 |
when he heard from someone that she was dead? 01:09:13.500 |
And you could basically understand what I'm saying, 01:09:33.500 |
these much, much, much more complicated expressions 01:09:41.180 |
as it were, three-word or four-word English sentences are, 01:09:44.320 |
which is really what the semantic web can represent 01:09:46.860 |
and really what knowledge graphs can represent. 01:09:58.020 |
as the vision of converting data on the internet 01:10:09.740 |
- So I wish I'd say, like, what is the semantic web? 01:10:14.800 |
but it might not be obvious to a lot of people 01:10:18.140 |
when they do a Google search that, just like you said, 01:10:29.420 |
ranked by the quality estimate of the website, 01:10:34.480 |
integrating previous human-based Google searches 01:10:43.080 |
of surface-level hacks that work exceptionally well, 01:11:00.700 |
the word understand, the contents of the Wikipedia pages 01:11:11.900 |
"for the computer to be able to truly understand 01:11:23.920 |
So there's a search engine you've probably never heard of 01:11:26.380 |
called Northern Light, and it went out of business, 01:11:35.060 |
And what it did was, it didn't index the internet at all. 01:11:40.980 |
All it did was, it negotiated and got access to data 01:11:57.700 |
and actually then they type in a completely different query, 01:12:16.780 |
it had no statistical data other than what I just mentioned, 01:12:24.820 |
"Oh, we're not gonna sell you this data anymore." 01:12:38.260 |
So let's go back to what you said about the semantic web. 01:12:45.340 |
dream about the semantic web at a general level, 01:13:04.460 |
statistically mapped linkages on the internet 01:13:09.460 |
into something that's more meaningful and semantic 01:13:20.340 |
And eventually, if you say, "Well, how can we do that?" 01:13:28.940 |
which is what the knowledge graphs are doing and so on, 01:13:48.940 |
And so being able to even just have a toe in the water 01:13:52.100 |
in the right direction is fantastically powerful. 01:13:58.380 |
But then you could say, "Well, what if we really wanted 01:14:00.420 |
to represent and reason with the full meaning 01:14:10.540 |
with the reasoning about what Juliet believes 01:14:12.860 |
that Romeo will believe that Juliet believed, 01:14:19.160 |
President Biden believed that the leaders of the Taliban 01:14:22.940 |
would believe about the leaders of Afghanistan 01:14:27.660 |
So in order to represent complicated sentences like that, 01:14:37.260 |
you need something which is logically much more expressive 01:14:44.700 |
than these simple knowledge graph type structures, 01:15:07.500 |
to this more and more general logical language, 01:15:14.820 |
and then second order, and then eventually higher order. 01:15:24.460 |
You can represent complicated kinds of negation. 01:15:29.460 |
You can represent the process you're going through 01:15:39.220 |
"Oh yeah, if you're trying to do this problem 01:15:56.180 |
So being able to talk about the problem-solving process 01:15:59.580 |
as you're going through the problem-solving process 01:16:05.100 |
- It's important to be able to represent that. 01:16:16.940 |
You do have rules of thumb that key off of them, 01:16:22.260 |
then it's sort of like someone with a limited vocabulary 01:16:28.980 |
And so that's really why I think that the general dream, 01:16:33.660 |
the original dream of Semantic Web is exactly right on. 01:16:45.380 |
little tiny baby steps in the right direction. 01:16:53.700 |
then yes, taking a baby step in the right direction 01:16:57.700 |
but it's not gonna be sufficient to actually get you 01:17:07.580 |
I always wondered, I built a bunch of websites, 01:17:19.100 |
that I can help Psych interpret the website I create? 01:17:24.100 |
Like this, again, pushing onto the Semantic Web dream, 01:17:29.340 |
is there something from the creator perspective 01:17:33.900 |
And one of the things you said with Psycorp and Psych 01:17:41.660 |
But is there on the other humans and the other side 01:17:45.860 |
Like for example, you and I are having a two, three, 01:17:49.980 |
Is there a way that I could convert this more, 01:17:53.100 |
make it more accessible to Psych, to machines? 01:17:57.500 |
- I'd love to see exactly that kind of semi-automated 01:18:03.380 |
understanding of what people write and what people say. 01:18:08.320 |
I think of it as a kind of footnoting, almost. 01:18:15.540 |
Almost like the way that when you run something 01:18:21.340 |
document preparation system, Google Docs or something, 01:18:24.660 |
you'll get underlining of questionable things 01:18:35.420 |
So I'd like to think in terms of Psych-powered tools 01:18:39.940 |
that read through what it is you said or have typed in 01:18:45.940 |
and try to partially understand what you've said. 01:18:54.620 |
- Exactly, and then they put in little footnotes 01:18:59.460 |
and they put in certain footnotes of the form, 01:19:05.420 |
You either meant this or this or this, I bet. 01:19:08.080 |
If you take a few seconds to disambiguate this for me, 01:19:18.620 |
or the next hundred thousand people who come here. 01:19:26.540 |
and you want people to understand your website content, 01:19:35.000 |
but actually be able to have systems that reason with it, 01:19:38.880 |
then yes, it will be worth your small amount of time 01:19:42.580 |
to go back and make sure that the AI trying to understand it 01:20:01.660 |
looking for vacations or trips that had certain properties 01:20:10.100 |
and might have been interesting to them for various reasons, 01:20:15.900 |
And if you've explained what's going to happen on your trip, 01:20:20.780 |
then a system will be able to mechanically reason 01:20:31.900 |
And so if it understands that there's a free day in Geneva, 01:20:42.700 |
happens to let's say be a nurse or something like that, 01:20:51.660 |
that that's where the International Red Cross Museum is 01:21:02.900 |
you might want to visit that Red Cross Museum. 01:21:05.980 |
And now, even though it's not very deep reasoning, 01:21:09.580 |
little tiny factors like that may very well cause you 01:21:12.460 |
to sign up for that trip rather than some competitor trip. 01:21:25.660 |
the design of the interface makes a huge difference. 01:21:41.780 |
and not from an AI perspective, just as a human. 01:21:47.660 |
how Tesla have implemented their autopilot system 01:21:51.900 |
is there's a sense that you're helping this machine learn. 01:21:56.460 |
And I think humans, I mean, having children, pets. 01:22:17.420 |
more like teaching a child, like innocent, curious system. 01:22:37.020 |
One of the experiences that we've all had in our lives 01:22:48.220 |
but then we found we really only understood it 01:22:50.940 |
when we had to teach it or explain it to someone 01:22:57.940 |
Despite the universality of that kind of experience, 01:23:05.380 |
almost all of it has the computer playing the role 01:23:08.860 |
of the teacher and the student plays the role of the student. 01:23:16.380 |
you can get a lot of learning to happen better. 01:23:22.780 |
if you are the mentor or the teacher and so on. 01:23:28.780 |
to help sixth graders better understand math. 01:23:33.140 |
And it doesn't actually try to teach you the player anything. 01:23:38.140 |
What it does is it casts you in the role of a student, 01:23:44.020 |
essentially, who has classmates who are having trouble 01:23:49.020 |
and your job is to watch them as they struggle 01:23:51.940 |
with some math problem, watch what they're doing 01:23:54.740 |
and try to give them good advice to get them to understand 01:24:00.780 |
And the trick from the point of view of psych 01:24:06.540 |
It has to play the role of the student who makes mistakes, 01:24:09.540 |
but it has to pick mistakes which are just at the fringe 01:24:17.060 |
So it pulls you into a deeper and deeper level 01:24:29.380 |
then psych knows that you now understand that mistake. 01:24:38.740 |
because there's no pedagogical usefulness to it. 01:24:46.460 |
'cause it used to make this mistake and now it doesn't 01:24:49.980 |
So there's tremendous reinforcement and engagement 01:24:55.340 |
So having a system that plays the role of a student 01:25:00.340 |
and having the player play the role of the mentor 01:25:08.820 |
just important way of having this sort of interface designed 01:25:13.820 |
in a way which will facilitate exactly the kind of learning 01:25:19.340 |
by teaching that goes on all the time in our lives 01:25:24.340 |
and yet which is not reflected anywhere almost 01:25:39.620 |
monitorial and Lancastrian education systems. 01:25:51.060 |
where you had one school room with one teacher 01:25:54.180 |
and it was basically five-year-olds to 18-year-olds 01:25:59.460 |
And so while the teacher was doing something, 01:26:02.020 |
half of the students would have to be mentoring 01:26:10.900 |
with scaling up of education, that all went away 01:26:16.780 |
and that incredibly powerful experience just went away 01:26:20.460 |
from the whole education institution as we know it today. 01:26:27.620 |
but what is the most beautiful idea you've learned 01:26:30.700 |
about artificial intelligence, knowledge, reasoning 01:26:52.460 |
And you are a part of creating one of the greatest, 01:27:03.200 |
- This may sound contradictory but I think it's 01:27:10.460 |
the feeling that this will be the only time in history 01:27:20.380 |
this particular thing that we're now teaching it. 01:27:29.980 |
You only have to do that once or creating the Pieta. 01:27:39.940 |
it's not like Bruce Springsteen having to sing 01:27:43.300 |
his greatest hits over and over again at different concerts. 01:27:47.220 |
It's more like a painter creating a work of art once 01:27:56.940 |
And so I really get the sense of we're telling the system 01:28:04.100 |
It's useful for a computer to know, for an AI to know. 01:28:07.900 |
And if we do our jobs right, when we do our jobs right, 01:28:24.340 |
and raising a child and teaching them all about this world. 01:28:33.340 |
It's a large number of children and a lot of parents 01:28:39.140 |
And with AI systems, at least the current constructions, 01:28:47.820 |
You don't get to experience the joy of teaching 01:28:54.580 |
- Better come work for us before it's too late then. 01:29:03.660 |
But then there's also, it's a project that continues 01:29:21.440 |
as an organism constantly grows and evolves and changes 01:29:32.260 |
And that's going to keep growing and growing and growing. 01:29:34.940 |
And many of the assertions that you held previously 01:29:45.980 |
It could be like a living organism versus the analogy 01:29:53.240 |
- The other beautiful experience that we have 01:29:59.820 |
with our system is when it asks clarifying questions, 01:30:04.300 |
which inadvertently turn out to be emotional to us. 01:30:09.300 |
So at one point it knew that these were the named entities 01:30:25.660 |
except for it, because it was also allowed to. 01:30:32.020 |
And we had to tell it, very sadly, no, you're not. 01:30:39.980 |
that are unintentionally poignant are worth treasuring. 01:30:53.020 |
who can access the system, who can modify it. 01:31:12.260 |
and when everyone realizes that they should have 01:31:23.540 |
about effectively enslaving our email systems 01:31:44.980 |
And I happen to believe that it'll happen in our lifetime. 01:31:48.380 |
- That's why I think there'll be a narrow envelope of time 01:32:02.540 |
that they should have freedoms that we afford 01:32:09.140 |
- And all of that starts with a system like Psyche 01:32:12.260 |
raising a single question about who can modify stuff. 01:32:26.900 |
and consciousness and all those kinds of topics? 01:32:30.860 |
Do they come up in Psyche in the knowledge base? 01:32:37.900 |
in fact, it's difficult to understand human behavior 01:32:40.220 |
and human history without understanding human emotions 01:33:07.940 |
is to generate plausible scenarios of what might happen 01:33:12.540 |
and what might happen based on that and so on. 01:33:14.700 |
So you generate this ever-expanding sphere, if you will, 01:33:17.660 |
of possible future things to worry about or think about. 01:33:22.660 |
And in some cases, those are intelligence agencies 01:33:31.180 |
so that we can defend against terrorist threats 01:33:41.020 |
so that we can actually close loopholes and vulnerabilities 01:33:51.580 |
Sometimes they are scenarios involving more positive things 01:34:15.820 |
that can be generated that way of cause and effect 01:34:26.140 |
many of the steps involve understanding and reasoning 01:34:30.820 |
about human motivations, human needs, human emotions, 01:34:38.260 |
in something that you do and why and how and so on. 01:34:42.020 |
So that was always a very important part of the knowledge 01:34:50.180 |
- So I talk a lot about love, so I got to ask, 01:34:55.020 |
how Psyche is able to represent various aspects of love 01:35:00.020 |
that are useful for understanding human nature 01:35:22.300 |
where you don't need to tease them apart further. 01:35:27.300 |
- So love is too general of a term, it's not useful. 01:35:30.580 |
- Exactly, so when you get down to romantic love 01:35:33.940 |
and sexual attraction, you get down to parental love, 01:35:37.220 |
you get down to filial love and you get down to love 01:35:48.020 |
So eventually you get down to maybe 50 or 60 concepts, 01:35:53.020 |
each of which is a kind of love, they're interrelated 01:35:57.620 |
and then each one of them has idiosyncratic things about it. 01:36:13.400 |
We may have one English word in to represent that, 01:36:21.500 |
because the way that the liquid is in the coffee cup 01:36:26.020 |
is different from the way that the air is in the room, 01:36:28.460 |
which is different from the way that I'm in my jacket 01:36:35.260 |
if I look at this coffee cup, well, I see the liquid. 01:36:38.820 |
If I turn it upside down, will the liquid come out 01:36:46.700 |
if I do the same thing, the sugar doesn't come out, 01:36:50.340 |
'cause it's dissolved in the liquid and so on. 01:36:52.660 |
So by now we have about 75 different kinds of in 01:36:56.100 |
in the system and it's important to distinguish those. 01:37:06.380 |
the writer of that was able to use this one innocuous word 01:37:15.260 |
that the reader had enough common sense and world knowledge 01:37:19.020 |
to disambiguate which of these 75 kinds of in 01:37:23.660 |
- And the same thing with love, you may see the word love, 01:37:36.140 |
So you have to be careful not to take language too seriously 01:37:41.140 |
because people have done a kind of parsimony, 01:37:48.220 |
a kind of terse-ness where you have as few words 01:37:51.220 |
as you can 'cause otherwise you'd need half a million words 01:37:59.340 |
That's like 10 times more than most languages 01:38:04.100 |
Just like we have on the order of about a million concepts 01:38:09.100 |
in psych because we've had to tease apart all these things. 01:38:13.620 |
And so when you look at the name of a psych term, 01:38:28.140 |
because you have to distinguish all these types of love, 01:38:30.780 |
you have to distinguish all these types of in, 01:38:43.780 |
almost as a feature has some ambiguity built in. 01:38:47.260 |
It's not an accident because the human condition 01:39:26.500 |
stories involving Marvin Minsky is when I asked him 01:39:31.260 |
about how he was able to turn out so many fantastic PhDs, 01:39:36.260 |
so many fantastic people who did great PhD theses. 01:39:45.500 |
What he said is he would generally say something 01:39:55.540 |
oh my God, Minsky said this, it must be a great idea. 01:39:58.740 |
And he'd sweat, he or she would work on work and work 01:40:09.180 |
- And then some great thesis would come out of it. 01:40:15.940 |
and I'm distinctly made aware that the words I say 01:40:28.820 |
and then letting them actually make something useful 01:40:47.060 |
There's so much, I've heard this interview with him 01:41:12.220 |
and I ran with it and I thought it was profound 01:41:16.780 |
- But a lot of behavior is in the eye of the beholder 01:41:19.820 |
and a lot of the meaning is in the eye of the beholder. 01:41:21.820 |
One of Minsky's early programs was Begging Program. 01:41:27.020 |
So this was back in the day when you had job control cards 01:41:37.880 |
to allow this to run before it got kicked off 01:41:40.620 |
and because computer time was enormously expensive. 01:41:44.140 |
And so he wrote a program and all it did was, 01:41:51.220 |
And all it did was it would wait like 20 seconds 01:42:00.380 |
So the operator would give it another 20 seconds, 01:42:08.620 |
and he'd be charged for like 10 times as much computer time 01:42:16.180 |
30 seconds here, you're charging me for five minutes, 01:42:25.140 |
and Marvin would say, oh, it always does that. 01:42:29.740 |
Is there, if you could just linger on it for a little bit, 01:42:48.500 |
in this very short history of artificial intelligence, 01:42:55.120 |
What have you learned from him as a human being, 01:43:01.360 |
- I would say both he and Ed Feigenbaum impressed on me 01:43:35.320 |
published journal article that five people will read 01:43:47.240 |
is not how you get from the earth to the moon 01:43:49.840 |
by slowly putting additional bumps on this log. 01:43:53.680 |
The only way to get there is to think about the hard problems 01:44:00.800 |
And if you do that, and if you're willing to listen 01:44:05.800 |
to nature, to empirical reality, willing to be wrong, 01:44:11.140 |
it's perfectly fine because if occasionally you're right, 01:44:15.520 |
then you've gotten part of the way to the moon. 01:44:30.080 |
So I'm sure there's been optimism in the early days 01:44:37.000 |
because I've talked to a few people on this podcast, 01:44:39.140 |
AI folks, that bring up Psych as an example of a project 01:44:45.200 |
that has a beautiful vision and it's a beautiful dream, 01:45:01.900 |
So why do you think people say that first of all? 01:45:10.760 |
And did you ever consider quitting on this mission? 01:45:27.520 |
And as a result, people often only know about us 01:45:31.440 |
because of a paper we wrote 10 or 20 or 30 or 37 years ago. 01:45:36.440 |
They only know about us because of what someone else 01:45:45.320 |
- So thank you for doing this podcast, by the way. 01:45:50.680 |
on some of the fascinating stuff you're doing. 01:45:52.640 |
- Well, I think it's time for us to keep a higher profile 01:46:07.820 |
but now that we've gotten this knowledge pump primed, 01:46:12.200 |
it's gonna become very important for everyone to help 01:46:17.920 |
if they are willing to, if they're interested in it. 01:46:23.860 |
and would like to leave some kind of legacy to the world, 01:46:28.780 |
people because of the pandemic who have more time at home 01:46:43.340 |
and how close to being primed the knowledge pump is, 01:46:55.980 |
about professional colleagues' opinions of our project. 01:47:14.300 |
including contrasting opinion that's worth representing. 01:47:26.780 |
There are times where I've become depressed a little bit 01:47:29.900 |
about how hard it is to get funding for the system. 01:47:34.660 |
Occasionally there are AI winters and things like that. 01:47:45.320 |
why in the world didn't you sell your company 01:48:02.000 |
So the answer is that one reason we've stayed 01:48:09.260 |
One reason that we haven't gone out of our way 01:48:13.660 |
is because we want to have control over our future, 01:48:21.700 |
so that we can continue to do this as until it's done. 01:48:26.620 |
And we're making progress and we're now so close to done 01:48:32.560 |
is commercial applications of our technology. 01:48:37.920 |
almost all of our money came from the government. 01:48:39.840 |
Now, virtually none of it comes from the government. 01:48:46.580 |
Hospital chains using it for medical reasoning 01:48:57.480 |
to reason about supply chains and things like that. 01:49:01.000 |
- So there's so many questions I want to ask. 01:49:13.520 |
I kind of want to ask you about your thoughts on this. 01:49:17.900 |
and you had big clients, you had a lot of clients, 01:49:26.280 |
of the kind of things you're helping them with. 01:49:31.560 |
And another way to operate is more in the open, 01:49:37.980 |
And so, hence, something like OpenPsych was born 01:49:52.300 |
And I want to say each of them before the other, 01:49:56.360 |
But we'll come back to OpenPsych in a minute. 01:50:05.200 |
is knowledge you have that is genuinely proprietary to you. 01:50:12.720 |
We'll make sure that it's marked as proprietary to you 01:50:30.040 |
in building any applications for you and with you, 01:50:33.880 |
which is publicly available general human knowledge, 01:50:59.900 |
- So there's a line between very domain-specific, 01:51:07.500 |
and the general knowledge that comes from that. 01:51:11.080 |
- Yes, or if you imagine, say, it's an oil company, 01:51:16.960 |
any new petroleum engineer they hired to already know. 01:51:32.060 |
And our attitude is some polite form of tough. 01:51:42.440 |
And in most cases, they'll see our point of view 01:51:45.980 |
and take it because that's how we've built the Psych system 01:52:00.000 |
that would stay permanently as part of Psych. 01:52:03.880 |
it's not like a distraction from the main Psych development. 01:52:20.480 |
So let me go back and talk about OpenPsych for a second. 01:52:28.360 |
convincing other AI researchers how important it is 01:52:35.360 |
to use an expressive representation language like we do, 01:53:00.120 |
"this unimportant projection or shadow or subset of Psych 01:53:05.120 |
"that just happens to be the simple binary relations, 01:53:10.100 |
"the relation argument one, argument two triples and so on." 01:53:14.800 |
And then you'll see how much more useful it is 01:53:22.320 |
So it's all well and good to have the taxonomic relations 01:53:27.320 |
between terms like person and night and sleep and bed 01:53:37.000 |
But think about how much more useful it would be 01:53:40.480 |
if you also had all the rules of thumb about those things 01:53:44.560 |
like people sleep at night, they sleep lying down, 01:53:51.720 |
they sleep for hours at a time, they can be woken up, 01:53:54.260 |
they don't like being woken up and so on and so on. 01:54:06.880 |
"of course we need the other 90% that you're not giving us. 01:54:24.660 |
there are a lot of robotics companies today, for example, 01:54:27.460 |
which use OpenPsych as their fundamental ontology. 01:54:32.060 |
And in some sense, the whole world missed the point 01:54:36.020 |
of OpenPsych and we were doing it to show people 01:54:41.620 |
and too many people thought somehow that this was psych 01:54:44.780 |
or that this was in fact good enough for them 01:54:57.340 |
and the power of this general kind of representation 01:55:21.940 |
This is the exact same problem that Google had to deal with 01:55:34.820 |
And that's only over a period of a couple of years. 01:55:40.900 |
it's very risky, very gutsy move to open up TensorFlow 01:55:56.180 |
that builds on the system without taking away any of, 01:56:03.900 |
So most of the value that Google has is still at Google. 01:56:06.700 |
Most of the value that Facebook has is still at Facebook, 01:56:08.940 |
even though some of this major machine learning tooling 01:56:25.220 |
all the kinds of stuff you can do on the knowledge, 01:56:34.220 |
there probably are a bunch of proprietary stuff 01:56:38.340 |
And there's probably some stuff you can open up completely 01:56:42.820 |
build up enough community where they develop stuff 01:56:45.700 |
Yes, there'll be those publications and academic work 01:56:49.820 |
And also the tooling of adding to the knowledge base, right? 01:56:54.100 |
Like developing, there's an incredible amount, 01:56:56.780 |
like there's so many people that are just really good 01:56:59.460 |
at this kind of stuff in the open source community. 01:57:06.580 |
that you have so much value in your company already? 01:57:11.500 |
You have clients that really value your relationships. 01:57:14.500 |
And then there's this dormant, giant open source community 01:57:39.900 |
has a phase shift towards like, or at least in part, 01:57:44.340 |
towards more like what you might call symbolic AI. 01:57:47.540 |
This whole place where psych is like at the center of, 01:57:52.540 |
and then that requires a little bit leap of faith 01:57:58.340 |
and there'll be obviously competitors that will pop up 01:58:00.900 |
and start making you nervous and all that kind of stuff. 01:58:08.860 |
how to leverage the community, all those kinds of things? 01:58:15.060 |
which is we're constantly struggling with the question 01:58:29.380 |
There are enormous pluses and minuses to every alternative. 01:58:34.380 |
And it's very much like negotiating a very treacherous path. 01:58:48.380 |
give away something which essentially kills you 01:58:53.100 |
which failing to give it away hurts you and so on. 01:59:01.940 |
Usually what we have done with people who approached us 01:59:11.780 |
we will make available to you the entire knowledge base 01:59:22.260 |
but only very, very limited source code access 01:59:28.500 |
if you have some idea for how you might improve something 01:59:33.420 |
So let me also get back to one of the very, very first things 01:59:38.260 |
we talked about here, which was separating the question 01:59:43.260 |
of how could you get a computer to do this at all 01:59:57.900 |
was that we had to separate the epistemological problem 02:00:10.220 |
And so instead of trying to pick one representation language 02:00:15.220 |
which was the sweet spot or the best trade-off point 02:00:34.060 |
But after a few years, we realized that what we could do 02:00:37.760 |
is we could split this and we could have one nice, 02:01:13.660 |
and they function as a kind of community of agents. 02:01:16.900 |
And there's one of them, which is a general theorem prover. 02:01:35.180 |
It's sort of like if you're balancing a chemical equation, 02:01:44.340 |
Or if you're trying to solve a quadratic equation, 02:01:47.780 |
you could go back to first principles of mathematics, 02:02:00.820 |
So think of these as like a thousand little experts 02:02:05.340 |
that are all looking at everything that Psych gets asked 02:02:10.500 |
that every other little agent has contributed, 02:02:29.020 |
And if we ever come up in a domain application 02:02:36.700 |
then what we'll often do is talk to one of the human experts 02:02:46.340 |
You can see why it took it a long time to get the answer. 02:02:49.620 |
How is it that you were able to answer that question 02:02:53.260 |
And occasionally you'll get an expert who just says, 02:03:14.340 |
or here's a special representation and a special technique, 02:03:22.780 |
And then you add that as the thousand and first 02:03:38.260 |
in which Psych has recouped this lost efficiency. 02:03:46.340 |
So you can speed things up by focusing on removing knowledge 02:04:03.300 |
So instead, what you wanna do is give it meta level advice, 02:04:06.900 |
tactical and strategic advice that enables it to reason 02:04:10.980 |
about what kind of knowledge is going to be relevant 02:04:16.660 |
are gonna be good to take in trying to attack this problem? 02:04:20.380 |
When is it time to start trying to prove the negation 02:04:23.700 |
of this thing because I'm knocking myself out 02:04:26.260 |
trying to prove it's true and maybe it's false. 02:04:52.540 |
You know, where one of them was this ELHL split 02:04:57.380 |
Another one was using meta and meta meta level reasoning 02:05:01.860 |
to reason about the reasoning that's going on and so on. 02:05:06.140 |
And you know, 150 breakthroughs may sound like a lot, 02:05:26.740 |
you know, in the general, the theorem prover sense, 02:05:39.060 |
if we let go of the perfect and focus on the good? 02:05:49.900 |
which is over the years, about half of our effort, 02:06:01.340 |
and the other 50, 60% has been our ontologists, 02:06:04.860 |
our ontological engineers putting in knowledge. 02:06:11.420 |
They have degrees in things like philosophy and so on. 02:06:16.700 |
I'd love to hang out with those people actually. 02:06:19.660 |
But it's very much like the Eloi and the Morlocks 02:06:27.380 |
in the epistemological higher order logic language. 02:06:34.380 |
under the ground figuring out what the machinery is 02:06:38.460 |
that will make this efficiently operate and so on. 02:06:42.420 |
And so occasionally they'll toss messages back 02:06:50.900 |
between finding clever ways to recoup efficiency 02:06:57.740 |
and putting in the content of what the system needs to know. 02:07:11.980 |
So my favorite program language is still Lisp. 02:07:24.980 |
but many of the systems are not written in Lisp anymore. 02:07:34.660 |
- Yeah, so it's based on Lisp code that we produced. 02:07:45.460 |
that gets automatically translated into things like Java 02:07:50.960 |
Nowadays, it's almost all translated into Java 02:08:05.060 |
it's a process that probably has to do with the fact 02:08:16.480 |
there is some technical depth you have to deal with, 02:08:19.740 |
as is the case with most powerful systems that span years. 02:08:32.300 |
so much of the value of everything you've done 02:08:40.780 |
Have you ever considered just like throwing away 02:08:54.080 |
starting with a more updated programming language. 02:09:05.900 |
and how much of the value is in the knowledge? 02:09:37.020 |
I just wonder how many great programmers there are. 02:09:40.540 |
- Yes, so it is true when a new inference programmer 02:09:43.860 |
comes on board, they need to learn some of Lisp. 02:09:55.780 |
And so the programming actually goes on in sub L, 02:10:01.340 |
And so it does not take programmers very long at all 02:10:06.220 |
And that's something which can then be translated 02:10:11.900 |
And for some of our programmers who are doing, 02:10:18.380 |
They just have to learn APIs into the basic Cyc engine. 02:10:23.500 |
- So you're not necessarily feeling the burden 02:10:32.420 |
The other thing is, remember that we're talking 02:10:56.780 |
who developed it basically took the formalisms 02:11:03.780 |
other logicians had come up with and basically said, 02:11:17.260 |
in about expressions written in this logical, 02:11:21.700 |
epistemological language and we're doing operations 02:11:28.900 |
there's a natural impedance match between Lisp 02:11:47.860 |
So I'll probably use that in the future without-- 02:11:53.100 |
But no, I think the point is that the language 02:12:01.180 |
It's more that you have to be able to think in terms of, 02:12:05.260 |
for instance, creating new helpful HL modules 02:12:10.700 |
and looking at things that are taking a long time 02:12:14.540 |
and coming up with new specialized data structures 02:12:20.140 |
So let me just give you one very simple example, 02:12:23.060 |
which is when you have a transitive relation, 02:12:28.500 |
which is larger than that, which is larger than that. 02:12:30.420 |
So the first thing must be larger than the last thing. 02:12:40.980 |
I'll have to do some kind of graph walk or theorem proving 02:12:45.060 |
that might involve like five or 10 or 20 or 30 steps. 02:12:48.540 |
But if you store, redundantly store the transitive closure, 02:12:58.780 |
but you can always guarantee that in one single step, 02:13:02.340 |
you can just look up whether this is larger than that. 02:13:06.180 |
And so there are lots of cases where storage is cheap today. 02:13:11.180 |
And so by having this extra redundant data structure, 02:13:15.860 |
we can answer this commonly occurring type of question 02:13:21.260 |
Let me give you one other analogy, analog of that, 02:13:26.540 |
which is something we call rule macro predicates, 02:13:33.140 |
and we'll notice that things very much like it, 02:13:37.420 |
syntactically come up again and again and again. 02:13:50.180 |
takes maybe three, four, five arguments and so on. 02:14:16.820 |
or schemas of rules into brand new predicates, 02:14:22.060 |
turns out to enormously speed up the inference process. 02:14:26.300 |
So now we've covered about four of the 150 good ideas 02:14:32.660 |
so that idea in particular is like a nice compression 02:14:44.220 |
because your work is both from a computer science perspective 02:14:51.540 |
from a epistemological, philosophical aspect fascinating. 02:14:59.420 |
and there's some stuff that has to remain private. 02:15:03.420 |
- Well, here's something that may make you feel better, 02:15:12.700 |
called the Knowledge Activization Institute, NAX, K-N-A-X. 02:15:20.020 |
with a lot of empirical evidence to support it 02:15:22.300 |
that the education that people get in high schools 02:15:37.740 |
how good they're going to be at coming up to speed 02:15:40.940 |
in doing this kind of ontological engineering 02:15:44.620 |
and writing these assertions and rules and so on in psych. 02:16:00.740 |
So one of the best ontological engineers we ever had 02:16:06.180 |
And so the purpose of Knowledge Activization Institute, 02:16:11.180 |
if we can get some foundations to help support it, 02:16:14.580 |
is identify people in the general population, 02:16:20.580 |
who have latent talent for this sort of thing, 02:16:24.700 |
offer them effectively scholarships to train them, 02:16:32.220 |
that need more trained ontological engineers, 02:16:55.020 |
- So is there something you can put into words 02:17:13.220 |
of human language and knowledge into formal logic. 02:17:17.180 |
- This is very much like what Alan Turing had to do 02:18:16.100 |
So you have to have a kind of sense of humor, 02:18:34.380 |
- Yeah, but another is if you're able to introspect. 02:18:38.140 |
So very often, we'll give someone a simple question, 02:18:48.060 |
And sometimes they'll just say, "Because it is." 02:18:53.220 |
But very often, they'll be able to introspect, and so on. 02:18:59.540 |
is I'll point to a sentence with a pronoun in it, 02:19:12.740 |
"know that that pronoun refers to that noun over here?" 02:19:17.740 |
And often, the people who are going to be good 02:19:27.020 |
or will refer to some things that are true in the world. 02:19:36.020 |
And so, "its head" refers to the horse's head. 02:19:40.300 |
And so, some people will say, "I just know it." 02:19:50.740 |
Now, "its roof" obviously refers to the barn, 02:19:55.980 |
"Oh, well, that's because it's the closest noun." 02:19:58.260 |
And so, basically, if they try to give me answers 02:20:01.820 |
which are based on syntax and grammar and so on, 02:20:17.260 |
because they can introspect on what's true in the world 02:20:25.100 |
makes you less capable to introspect deeply about this? 02:20:30.860 |
I'm not saying that it makes you less capable. 02:20:37.140 |
- Okay, you're not saying that, I'm saying that. 02:20:41.640 |
that for a lot of people, PhDs, sorry, philosophy aside, 02:20:46.640 |
that sometimes education narrows your thinking 02:20:57.320 |
which is essentially teach our future AI overlords 02:21:05.220 |
that requires that you think deeply about the world. 02:21:08.660 |
- So I'll tell you a sad story about Math Craft, 02:21:11.900 |
which is why is that not widely used in schools today? 02:21:16.020 |
We're not really trying to make big profit on it 02:21:20.580 |
When we've gone to schools, their attitude has been, 02:21:31.620 |
will it improve their score on this standardized test 02:21:48.900 |
if it doesn't increase their score on this test, 02:21:56.860 |
I mean, that's a whole another three, four hour conversation 02:22:01.820 |
But let me go super philosophical as if we weren't already. 02:22:23.540 |
certainly as well as humans can think, right? 02:22:30.340 |
just because they're not currently made out of meat 02:22:32.820 |
is just an engineering solution decision and so on. 02:22:47.700 |
by people misunderstanding Turing's imitation game 02:22:58.260 |
to fool other people into thinking it was human and so on. 02:23:08.980 |
but it shouldn't be your one and only test for intelligence. 02:23:12.580 |
- So do you, in terms of tests of intelligence, 02:23:36.300 |
how long does a human want to talk to the AI system? 02:23:39.340 |
So it's like, the goal is you want it to be 20 minutes. 02:23:43.540 |
It's basically not just have a convincing conversation, 02:24:01.020 |
But what for you do you think in the space of tests 02:24:06.540 |
When you see a system based on psych that passes that test, 02:24:10.900 |
you'd be like, damn, we've created something special here. 02:24:19.380 |
depth of reasoning and recursiveness of reasoning, 02:24:28.940 |
- How many why questions in a row can you keep answering? 02:24:36.260 |
- Just have like a young curious child and an AI system. 02:24:56.060 |
and try and give me convincing arguments on both sides. 02:25:12.780 |
- So there's certain, I mean, if you look at IBM Watson 02:25:18.220 |
for very specific tests, almost like a demo, right? 02:25:33.780 |
And there's some kind of hard coding heuristics tricks 02:25:39.700 |
to make the thing work in the end for this thing, right? 02:25:45.380 |
is like that's the fastest way to get a solution 02:25:53.300 |
as impressive as that was, it made some mistakes. 02:25:59.220 |
But more importantly, many of the mistakes it made 02:26:02.980 |
were mistakes which no human would have made. 02:26:05.780 |
And so part of the new or augmented Turing tests 02:26:18.220 |
are ones which humans don't basically look at and say, what? 02:26:28.700 |
about which 16th century Italian politician, blah, blah, blah 02:26:37.140 |
So most Americans would have gotten that question wrong 02:26:40.260 |
but they would never have said Ronald Reagan as an answer 02:27:05.700 |
And if it's not making those kinds of mistakes, 02:27:10.020 |
And I don't think it's any one very, very simple test. 02:27:12.980 |
I think it's all of the things you mentioned, 02:27:16.020 |
there's really a battery of tests which together, 02:27:22.460 |
it would be hard to argue that it's not intelligent. 02:27:28.380 |
it's really hard to argue that it really understands 02:27:30.980 |
what it's doing and that it really is generally intelligent. 02:27:35.860 |
we've talked a lot about psych and knowledge and reasoning. 02:27:40.660 |
Do you think this AI system would need to have 02:27:46.700 |
for example, a body or a physical manifestation in this world 02:27:52.380 |
and another one which seems to be fundamental 02:28:04.540 |
Do you think it needs those to be able to pass 02:28:06.980 |
all of those tests and to achieve general intelligence? 02:28:12.660 |
I know there are a lot of people like Penrose 02:28:19.460 |
but no, I don't think it needs to have a body 02:28:24.300 |
I think that it needs to be able to talk about 02:28:34.020 |
It doesn't actually have to have all of that, 02:28:42.500 |
and able to talk about colors and sounds and shapes 02:28:47.580 |
and so on, even though she didn't directly experience 02:28:54.500 |
So knowledge of it and being able to correctly make use 02:29:04.980 |
but actually having a body, if you believe that, 02:29:07.940 |
that's just a kind of religious or mystical belief. 02:29:11.420 |
You can't really argue for or against it, I suppose. 02:29:15.020 |
It's just something that some people believe. 02:29:32.340 |
you say things which make me believe that you're conscious. 02:29:35.940 |
I know that I'm conscious, but you're just taking 02:29:43.340 |
in that same sense already, where of course it understands 02:29:46.420 |
it's a computer program, it understands where 02:29:48.660 |
and when it's running, it understands who's talking to it, 02:29:51.420 |
it understands what its task is, what its goals are, 02:29:54.300 |
what its current problem is that it's working on. 02:29:56.780 |
It understands how long it's spent on things, 02:29:58.700 |
what it's tried, it understands what it's done in the past 02:30:11.180 |
but I don't think that I would ascribe anything mystical 02:30:17.100 |
but I would say that other than our own personal experience 02:30:21.300 |
of consciousness, we're just treating everyone else 02:30:29.580 |
And so if a computer program, if an AI is able to exhibit 02:30:34.580 |
all the same kinds of response as you would expect 02:30:40.260 |
of a conscious entity, then doesn't it deserve 02:30:49.220 |
with this whole intelligence thing that humans got 02:30:52.020 |
is the extinguishing of the light of consciousness, 02:30:57.020 |
which is kind of realizing that we're gonna be dead someday. 02:31:02.380 |
And there's a bunch of philosophers like Ernest Becker 02:31:05.700 |
who kind of think that this realization of mortality 02:31:10.420 |
and then fear, sometimes they call it terror, 02:31:19.300 |
behind human condition, like it's the thing that drives us. 02:31:25.300 |
Do you think it's important for an AI system, 02:31:27.860 |
you know, when psych proposed that it's not human 02:31:34.440 |
and it's one of the moderators of its contents, 02:31:37.860 |
you know, there's another question it could ask, 02:31:42.440 |
which is like, it kind of knows that humans are mortal. 02:31:47.480 |
And I think one really important thing that's possible 02:31:52.360 |
when you're conscious is to fear the extinguishing 02:31:55.460 |
of that consciousness, the fear of mortality. 02:32:02.000 |
Thinking like I might die and I really don't want to die. 02:32:06.700 |
I think it may help some humans to be better people. 02:32:11.700 |
It may help some humans to be more creative and so on. 02:32:16.180 |
I don't think it's necessary for AIs to believe 02:32:23.320 |
and therefore they should make the most of their behavior. 02:32:30.440 |
But as of now, I would say that that's almost like a frill 02:32:38.740 |
most humans ignore the fact that they're going to die 02:32:46.620 |
this goes to the white space between the words. 02:32:49.540 |
So what Ernest Becker argues is that that ignoring 02:32:53.040 |
is we're living in an illusion that we constructed 02:33:04.720 |
everything we can think of that's beautiful about humanity 02:33:13.160 |
That's his idea and I think, I don't know if I 100% 02:33:22.620 |
It seems like to me, like it rhymes with the truth. 02:33:28.420 |
that's gonna be a more powerful factor than others. 02:33:49.400 |
as a motivating force in that we can build in motivation. 02:33:55.540 |
So we can build in the motivation of obeying users 02:34:00.540 |
and making users happy and making others happy and so on. 02:34:05.620 |
And that can substitute for this sort of personal fear 02:34:17.820 |
I think AI really needs to understand death deeply 02:34:22.060 |
in order to be able to drive a car, for example. 02:34:30.080 |
I think it needs to understand the value of human life, 02:34:33.640 |
especially the value of human life to other humans, 02:34:36.140 |
and understand that certain things are more important 02:34:41.540 |
than other things, so it has to have a lot of knowledge 02:34:48.060 |
- But some of it is so messy that it's impossible to encode. 02:34:53.720 |
So if there's a person dying right in front of us, 02:35:04.500 |
This is the tragedy of how difficult it is to be a doctor 02:35:08.020 |
because they know when they help a dying child, 02:35:11.420 |
they know that the money they're spending on this child 02:35:14.860 |
cannot possibly be spent on every other child that's dying. 02:35:18.580 |
And that's a very difficult to encode decision. 02:35:23.240 |
Now, perhaps it is, perhaps it could be formalized. 02:35:31.780 |
So autonomous vehicles are going to have to make 02:35:34.740 |
those decisions all the time of what is the chance 02:35:39.740 |
of this bad event happening, how bad is that compared 02:35:44.220 |
to this chance of that bad event happening, and so on, 02:35:47.300 |
and when a potential accident is about to happen, 02:35:52.900 |
If I have to make a choice, which of these two cars 02:35:57.000 |
- See, I was thinking about a very different choice 02:36:01.100 |
which is just observing Manhattan style driving. 02:36:14.660 |
There's a dance, I've watched pedestrians a lot, 02:36:17.380 |
I worked on this problem, and it seems like the, 02:36:21.380 |
if I could summarize the problem of a pedestrian crossing 02:36:28.260 |
"I'm going to kill you," and the pedestrian is saying, 02:36:34.400 |
"No, I don't think you have the guts to kill me," 02:36:42.980 |
as this is social contract, that the pedestrian trusts 02:36:47.500 |
and the car is sufficiently, from a physics perspective, 02:36:53.140 |
But the car also has to threaten that pedestrian, 02:37:00.980 |
but life and death is part of the calculation here, 02:37:05.760 |
and that equation is being solved millions of times a day. 02:37:12.820 |
That game theory, whatever that formulation is. 02:37:31.380 |
that the fear of death is just one of the quirks 02:37:38.500 |
but it's not a necessary feature of intelligence. 02:37:42.120 |
- Drivers certainly are always doing this kind of estimate, 02:37:48.740 |
of what are the chances of various bad outcomes happening, 02:37:52.640 |
like for instance, if I don't wait for this pedestrian 02:38:01.300 |
in terms of time wasted talking to the police 02:38:12.980 |
like people in their cars tend to get irrationally angry. 02:38:19.820 |
this is all part of why I think that autonomous vehicles, 02:38:28.580 |
because there is this enormous level of complexity 02:38:32.340 |
which goes beyond mechanically controlling the car. 02:38:42.860 |
as a kind of metaphorical and literal accident 02:38:46.940 |
And not just because of their overall incurring 02:38:56.740 |
but just because of the almost voracious appetite 02:39:08.040 |
bad stories about powerful companies and powerful entities. 02:39:16.600 |
Japanese fifth generation computing system conference 02:39:27.480 |
by climbing under the safety chains and so on, 02:39:32.800 |
And instead of being a small story that said, 02:39:38.280 |
it was front page news that effectively said, 02:39:44.240 |
because the public is just waiting for stories 02:39:47.920 |
about like AI kills phonogenic family of five type stories. 02:40:03.480 |
prevented more injuries than it caused and so on, 02:40:06.600 |
the media, the public, the government is just coiled 02:40:10.600 |
and ready to pounce on stories where in fact it failed, 02:40:18.540 |
- Yeah, it's so fascinating to watch us humans 02:40:23.200 |
resisting the cutting edge of science and technology 02:40:27.520 |
and almost like hoping for it to fail and constantly, 02:40:33.880 |
- Or even if we're not hoping for it to fail, 02:40:43.120 |
much more interesting than the 999 boring successes. 02:41:18.960 |
what are some possible solutions to big problems 02:41:33.160 |
that are not being properly or at least adequately pursued? 02:41:38.160 |
What are some novel solutions that you can think of 02:41:50.800 |
- Given that the AGI is going to be somewhat different 02:42:12.720 |
can it help spot some of the blind spots that we have? 02:42:16.080 |
- So the two-part question of can you help identify 02:42:23.200 |
and two, what are some novel solutions to those problems? 02:42:36.600 |
but some of them might be actually great things to look at. 02:42:45.840 |
like relativity and superconductivity and so on, 02:42:50.520 |
a lot of them were cases where someone took seriously 02:43:09.400 |
nobody believes that it's actually the way reality works, 02:43:12.100 |
what if it were the way that reality actually worked? 02:43:27.620 |
And it was being mapped at lower and lower temperatures, 02:43:32.560 |
but everyone thought that was just bump on a log research 02:43:45.760 |
and showed that resistance suddenly dropped off, 02:43:47.720 |
everyone just assumed that they did it wrong. 02:43:52.800 |
that they realized it was actually a new phenomenon. 02:44:11.360 |
claimed it was actually a bacterial infection, 02:44:15.480 |
he couldn't get anyone seriously to listen to him 02:44:17.840 |
and he had to ultimately inject himself with the bacteria 02:44:22.840 |
to show that he suddenly developed a life-threatening ulcer 02:44:26.680 |
in order to get other doctors to seriously consider that. 02:44:40.440 |
So a lot of AI is locked into the deep learning, 02:44:52.720 |
And Kuhn's point was pretty much you have to wait 02:44:56.960 |
for people to die in order for the new generation 02:45:07.200 |
is if we had trusted AGIs that could help take a step back 02:45:17.880 |
- Yeah, it would accelerate the paradigm shifts 02:45:28.080 |
where you thought about big ideas and you stuck with them. 02:45:54.700 |
which other people might not immediately understand 02:46:02.400 |
You have to realize that if you make some plan 02:46:06.400 |
that's going to take an extended period of time 02:46:16.440 |
That's true of physical training of your body. 02:46:34.820 |
and become an instant success if you turn out to be right. 02:47:03.780 |
And if you follow the path of least resistance, 02:47:06.700 |
you'll find that you're optimizing for short periods of time 02:47:17.000 |
without you ever really making a difference in the world. 02:47:23.800 |
is artificial intelligence and there's not many projects. 02:47:31.220 |
that have been carried out for many years, for decades, 02:47:36.560 |
And I mean, that in itself is just a really inspiring thing. 02:47:42.400 |
So I'm deeply grateful that you would be carrying that flame 02:47:46.600 |
for so many years and I think that's an inspiration 02:47:51.920 |
and we talked about mortality as a feature of AGI. 02:48:14.080 |
and especially mentors, one by one are dying. 02:48:22.680 |
And I think that the good news from the point of view 02:48:26.720 |
in the rest of the world is that that adds impetus 02:48:30.640 |
to my need to succeed in a small number of years 02:48:36.800 |
- Exactly, I'm not gonna have another 37 years 02:48:41.440 |
So we really do want psych to make an impact in the world, 02:48:49.300 |
in the next small number of years, two, three, five years, 02:49:02.680 |
and increasingly widespread application of psych. 02:49:07.960 |
Whereas before I felt that I could just sort of sit back, 02:49:19.400 |
to make the world aware of what we have and what it can do. 02:49:31.520 |
And if I can help in any way, I would love to, 02:49:42.720 |
So if there's any open sourcing from an MIT side 02:50:04.280 |
That people think of me as one of the pioneers 02:50:25.480 |
or the pioneers of similar types of technologies and so on 02:50:31.120 |
as it's hard to imagine what life would be like 02:50:39.920 |
So that's one thing that I'd like to be remembered as. 02:50:47.800 |
of this gigantic knowledge store and acquisition system 02:50:57.760 |
of whatever this future AI thing will look like. 02:51:01.840 |
And I'd also like to be remembered as someone 02:51:09.040 |
on a project in a time when almost all of the other forces, 02:51:25.480 |
are incenting people to go for short-term rewards. 02:51:31.440 |
A lot of people that dreamt the same dream as you gave up. 02:51:45.200 |
A lot of people bring up your work specifically 02:51:52.320 |
of this is the dream of artificial intelligence. 02:51:57.680 |
We're so sort of focused on machine learning applications, 02:52:01.960 |
but it seems like the ideas that Cy carries forward 02:52:04.880 |
is something that will be at the center of this problem 02:52:18.320 |
It's such a huge honor that you would talk to me 02:52:30.840 |
please check out our sponsors in the description. 02:52:33.640 |
And now let me leave you some words from Mark Twain 02:52:39.160 |
If you tell the truth, you don't have to remember anything.