back to indexStephen Wolfram: ChatGPT and the Nature of Truth, Reality & Computation | Lex Fridman Podcast #376
Chapters
0:0 Introduction
1:33 WolframAlpha and ChatGPT
21:14 Computation and nature of reality
48:6 How ChatGPT works
107:48 Human and animal cognition
121:7 Dangers of AI
129:27 Nature of truth
150:49 Future of education
186:51 Consciousness
195:50 Second Law of Thermodynamics
219:23 Entropy
232:23 Observers in physics
249:15 Mortality
00:00:05.160 |
And I'm like, that sort of personalizes for me 00:00:08.840 |
the what could possibly go wrong, so to speak. 00:00:11.480 |
- Was that exciting or scary, that possibility? 00:00:20.560 |
And that's a version of that question for the world. 00:00:24.800 |
That is, as soon as you put the AIs in charge of things, 00:00:27.960 |
how many constraints should there be on these systems 00:00:31.600 |
before you put the AIs in charge of all the weapons 00:00:44.560 |
The following is a conversation with Stephen Wolfram, 00:01:01.320 |
and the Wolfram Physics and Metamathematics projects. 00:01:07.520 |
in exploring the computational nature of reality. 00:01:10.080 |
And so he's the perfect person to explore with together 00:01:26.740 |
please check out our sponsors in the description. 00:01:28.840 |
And now, dear friends, here's Stephen Wolfram. 00:01:57.440 |
- Yeah, so what does something like Chad GPT do? 00:02:20.160 |
of what it's seen based on a trillion words of text 00:02:28.140 |
with something which is probably quite similar 00:02:30.700 |
to the way we humans do the first stages of that, 00:02:41.660 |
one word and get one word at a time of output. 00:03:00.920 |
which has to do with what can you compute many steps, 00:03:12.220 |
and trying to continue things based on that statistics. 00:03:15.320 |
Instead, it's trying to take kind of the formal structure 00:03:23.180 |
or whether it's from kind of systematic knowledge 00:03:26.880 |
and use that to do arbitrarily deep computations 00:03:34.120 |
let's match what's already been kind of said on the web, 00:03:37.760 |
but let's potentially be able to compute something new 00:03:40.400 |
and different that's never been computed before. 00:03:55.560 |
from some sort of expert knowledge that's been accumulated, 00:04:10.280 |
It's a very, it's a much more sort of labor intensive 00:04:20.280 |
Obviously the, in the kind of the chat GPT world, 00:04:28.360 |
namely all the things we've written out on the web and so on 00:04:34.960 |
which are like what's been written on the web. 00:04:37.520 |
So I think, you know, as a practical point of view, 00:04:39.640 |
I view sort of the chat GPT thing as being wide and shallow 00:04:51.800 |
but most importantly kind of deep type of thing. 00:04:58.320 |
you go back in human history, you know, I don't know, 00:05:05.480 |
what's the typical person going to figure out? 00:05:08.040 |
Well, the answer is there's certain kinds of things 00:05:16.360 |
and the kinds of things we learn in our lives let us do. 00:05:19.880 |
But then there's this whole layer of kind of formalization 00:05:25.520 |
the kind of whole sort of story of intellectual history 00:05:31.380 |
that formalization turned into things like logic, 00:05:45.200 |
It's not just, I can immediately figure this out. 00:05:58.200 |
is to be able to build those kind of tall towers 00:06:14.120 |
- What can you say about the kind of formal structure, 00:06:36.200 |
One is what computation intrinsically is like. 00:06:47.460 |
can sort of relate to in that computational universe. 00:06:51.260 |
So if we start on the kind of what can computation be like, 00:06:55.020 |
it's something I've spent some big chunk of my life studying 00:07:00.460 |
we usually write programs where we kind of know 00:07:07.300 |
and we hope that the program does what we intended it to do. 00:07:13.020 |
if you just look at the kind of natural science of programs, 00:07:15.980 |
so you just say, I'm gonna make this program, 00:07:19.300 |
Maybe I even pick the pieces of the program at random, 00:07:29.500 |
And big discovery that I made in the early 80s 00:07:35.440 |
when you run them can do really complicated things. 00:07:43.500 |
But that realization that even very simple programs 00:07:50.720 |
That discovery, I mean, I realized that that's very much, 00:08:03.540 |
has been understanding that that's how the whole universe 00:08:05.780 |
and physics works, but that's a quite separate topic. 00:08:12.380 |
and what they do and very rich, sophisticated things 00:08:19.140 |
we look at them and say, well, that's kind of, 00:08:24.680 |
So on the one hand, we have sort of what's possible 00:08:28.580 |
On the other hand, we have the kinds of things 00:08:30.180 |
that we humans think about, the kinds of things 00:08:32.500 |
that are developed in kind of our intellectual history. 00:08:41.520 |
is to connect what's computationally possible 00:08:47.860 |
sort of typically think about with our minds. 00:08:50.620 |
Now, that's a complicated kind of moving target 00:08:53.840 |
because the things that we think about change over time. 00:09:05.380 |
We're kind of gradually colonizing more and more 00:09:08.100 |
of this kind of intellectual space of possibilities. 00:09:13.540 |
how do you take what is computationally possible? 00:09:28.880 |
is this idea of kind of symbolic programming, 00:09:35.560 |
when you look at sort of everything in the world 00:09:40.980 |
and you say, well, how do I turn that into something 00:09:45.900 |
You know, there are lots of pixels in my visual scene, 00:09:48.720 |
but the things that I remembered from that visual scene are, 00:09:58.280 |
There are two chairs and a table or something, 00:10:13.940 |
And human language is sort of one form of representation 00:10:22.020 |
How do we take, but human language is not in and of itself 00:10:26.440 |
something that plugs in very well to sort of computation. 00:10:31.360 |
It's not something from which you can immediately 00:10:35.720 |
And so you have to kind of find a way to take 00:10:38.920 |
sort of the stuff we understand from human language 00:10:44.840 |
And that's really the story of symbolic programming. 00:10:48.300 |
And you know, what that turns into is something 00:10:56.920 |
I was trying to build my first big computer system 00:11:00.820 |
how should I represent computations at a high level? 00:11:14.420 |
but that function doesn't necessarily evaluate to anything. 00:11:17.960 |
It's just a thing that sits there representing a structure. 00:11:24.660 |
and it's turned out that structure has been extremely, 00:11:28.460 |
it's a good match for the way that we humans, 00:11:31.280 |
it seems to be a good match for the way that we humans 00:11:38.480 |
45 years or something, it's served me remarkably well. 00:11:46.600 |
But what can you say about abstractions here? 00:11:49.880 |
Because you could just start with your physics project, 00:11:52.520 |
you could start at a hypergraph at a very, very low level 00:11:55.440 |
and build up everything from there, but you don't. 00:12:03.360 |
convert that, the kind of abstraction that's convertible 00:12:07.080 |
to something computable using symbolic representation. 00:12:17.240 |
- Right, so the sort of a very important phenomenon 00:12:20.520 |
that is kind of a thing that I've sort of realized 00:12:23.920 |
is just, it's one of these things that sort of 00:12:30.800 |
as this phenomenon of computational irreducibility. 00:12:33.640 |
And the question is, if you know the rules for something, 00:12:52.000 |
The question is, can you like immediately jump ahead 00:12:55.600 |
and say, I know what's gonna happen after a million steps 00:13:00.440 |
And one of the very critical things to realize is 00:13:07.400 |
there is in a sense, no point in doing the computation. 00:13:14.240 |
you had to do the computation to find out the answer. 00:13:17.000 |
But this phenomenon that you have to do the computation 00:13:20.040 |
this phenomenon of computational irreducibility 00:13:28.080 |
okay, you've got a model of the universe at the low level 00:13:35.120 |
And it's happening 10 to the 100 times every second, 00:13:38.120 |
let's say, well, you say, great, then we've nailed it. 00:13:43.840 |
Well, the problem is the universe can figure out 00:13:56.920 |
to see the result of the computation is to do it. 00:14:01.960 |
we're kind of, there's no opportunity to do that 00:14:24.800 |
The only way we can figure out what's gonna happen next 00:14:26.880 |
is just let the system run and see what happens. 00:14:29.680 |
So in a sense, the story of most kinds of science, 00:14:40.000 |
And one of the features of computational irreducibility 00:14:46.360 |
there are always an infinite number of places 00:14:49.200 |
There's no way where you can jump completely ahead, 00:14:53.960 |
little places where you can jump ahead a bit. 00:14:56.440 |
And I think, we can talk about physics project and so on, 00:14:59.320 |
but I think the thing we realize is we kind of exist 00:15:04.840 |
computational irreducibility in the universe. 00:15:08.600 |
where there's a reasonable amount of predictability. 00:15:19.040 |
what we're doing is we're finding these lumps 00:15:21.320 |
of reducibility that we can kind of attach ourselves to 00:15:36.960 |
and oh gosh, it's an incredibly complicated story. 00:15:40.320 |
And that's a whole computationally irreducible thing, 00:15:45.080 |
And most of it is, well, the air is still gonna be here 00:15:48.680 |
and nothing much is going to be different about it. 00:15:53.760 |
about what is ultimately at an underlying level 00:16:01.960 |
if we didn't have a large number of such reducible pockets. 00:16:07.840 |
- And that's amenable to reduction into something symbolic. 00:16:12.200 |
I mean, life in the way that we experience it, 00:16:20.080 |
depending on what we mean by life, so to speak, 00:16:25.440 |
of sort of consistent things happening in the world, 00:16:28.280 |
the idea of space, for example, where there's, 00:16:31.280 |
we can just say, you're here, you move there. 00:16:38.040 |
even though you're made of different atoms of space 00:16:40.920 |
This idea that there's sort of this level of predictability 00:16:52.120 |
this computationally irreducible kind of system. 00:17:01.740 |
is the realization that it is sort of the interaction 00:17:05.380 |
between the sort of underlying computational irreducibility 00:17:12.780 |
who sort of have to key into computational reducibility. 00:17:22.400 |
So this is, we talk about this in more detail, 00:17:25.080 |
but this is, to me, it's kind of our nature as observers, 00:17:30.080 |
the fact that we are computationally bounded observers, 00:17:34.300 |
we don't get to follow all those little pieces 00:17:38.080 |
To stuff what is out there in the world into our minds 00:17:42.380 |
requires that we are looking at things that are reducible, 00:17:51.200 |
of what's the detail of what's going on in the world. 00:17:56.620 |
that at first seems sort of trivial, but isn't, 00:17:59.900 |
which is that we believe we are persistent in time. 00:18:13.740 |
At every moment, sort of the microscopic detail 00:18:16.620 |
of what the universe is made of is being rewritten. 00:18:21.940 |
that there's coherence between different parts of space 00:18:25.620 |
that there are all these little processes going on 00:18:27.460 |
that kind of knit together the structure of space. 00:18:29.160 |
It's kind of like if you wanted to have a fluid 00:18:37.740 |
It would just be sort of a free-floating collection 00:18:42.460 |
that the fact that space is kind of knitted together 00:18:44.780 |
is a consequence of all this activity in space. 00:19:12.280 |
- Well, isn't that just another limitation of our mind 00:19:33.060 |
is that we have a single thread of experience. 00:19:39.000 |
where you have, you know, maybe that's what's happening 00:19:42.560 |
that aren't working the same way other minds work, 00:19:55.880 |
it's splitting into many threads of experience. 00:19:58.380 |
But in order for us humans to interact with it, 00:20:04.800 |
so that we say, oh yeah, a definite thing happened, 00:20:07.020 |
and now the next definite thing happens, and so on. 00:20:19.760 |
fundamentally multiple threads of experience going on. 00:20:48.120 |
of kind of things that can happen in the universe, 00:20:57.080 |
you have this idea of a single thread of experience, 00:21:01.960 |
that could kind of computationally happen in the universe. 00:21:04.080 |
- So it's a feature of a computationally limited system 00:21:07.060 |
that's only able to observe reducible pockets. 00:21:22.340 |
It means something to us humans as conscious beings. 00:21:29.640 |
What is the observer and what's the importance 00:21:31.280 |
of the observer in the computational universe? 00:21:51.640 |
So one question is what is a general observer? 00:22:01.880 |
There's a question, what is a general model of an observer? 00:22:08.440 |
which is kind of the observers we're interested in. 00:22:12.500 |
that deals with computational irreducibility, 00:22:14.320 |
and it has a mind that's utterly different from ours 00:22:16.640 |
and completely incoherent with what we're like. 00:22:27.320 |
of kind of taking all the detail of the world 00:22:46.680 |
so I've been interested in trying to characterize 00:22:51.200 |
And the general observer is, I think, in part, 00:22:54.920 |
there are many, let me give an example of it. 00:22:57.900 |
it's got a bunch of molecules bouncing around, 00:23:00.240 |
and the thing you're measuring about the gas is its pressure. 00:23:04.160 |
And the only thing you as an observer care about is pressure. 00:23:07.040 |
And that means you have a piston on the side of this box, 00:23:15.960 |
But all that matters is the kind of aggregate 00:23:22.760 |
So there's a huge number of different configurations 00:23:29.080 |
is this equivalency of many different configurations 00:23:42.720 |
again, it's the same story over and over again, 00:23:47.320 |
but what we are extracting from it is something, 00:24:03.560 |
I mean, if we look at the observer that's the human mind, 00:24:07.640 |
as represented by natural language, for example, 00:24:10.520 |
there's a lot of really crappy approximation. 00:24:24.800 |
tiny probability that molecules will arrange themselves 00:24:39.640 |
and people are really keen, they draw this curve 00:24:41.920 |
and they have these bars on the curve and things. 00:24:52.480 |
And this is a way that lots of science has gotten wrong. 00:25:00.920 |
it has all these arms, it's doing complicated things. 00:25:05.720 |
and it talked about what's the rate of snowflake growth. 00:25:17.280 |
I looked at it more carefully and I realized, 00:25:19.200 |
according to their models, the snowflake will be spherical. 00:25:28.320 |
And not only the detail, the whole thing was not capturing, 00:25:39.400 |
- What is the geometric shape of a snowflake? 00:25:51.240 |
with a hexagonal arrangement of water molecules. 00:25:54.240 |
And so it starts off growing as a hexagonal plate. 00:25:58.480 |
- It's a plate, oh, oh, versus sphere, sphere versus plate. 00:26:00.880 |
- Well, no, no, but it's much more than that. 00:26:04.280 |
Typical snowflakes have little dendritic arms. 00:26:07.640 |
And what actually happens is, it's kind of cool 00:26:10.120 |
because you can make these very simple discrete models 00:26:13.080 |
with cellular automata and things that figure this out. 00:26:19.380 |
and then the places, it starts to grow little arms. 00:26:23.040 |
And every time a little piece of ice adds itself 00:26:26.400 |
to the snowflake, the fact that that ice condensed 00:26:29.520 |
from the water vapor heats the snowflake up locally. 00:26:33.160 |
And so it makes it less likely for another piece of ice 00:26:40.320 |
So this leads to a kind of growth inhibition. 00:26:46.040 |
because right around the arm, it got a little bit hot 00:26:51.520 |
So what happens is it grows, you have a hexagon, 00:26:59.880 |
because it actually fills in another hexagon, 00:27:13.760 |
and sure enough, they have these little holes in them 00:27:15.920 |
that are kind of scars of the way that these arms grow out. 00:27:23.800 |
So it just keeps going up. - They don't backfill, yeah. 00:27:25.880 |
- And presumably, there's a limitation of how big, 00:27:33.440 |
I mean, I think it hits the ground at some point. 00:27:36.880 |
I think you can grow, I think you can grow in the lab. 00:27:43.540 |
of this kind of, goes from hexagon, it grows out arms, 00:27:57.520 |
and fluffy is a three-dimensional property, no? 00:28:01.420 |
Snow is, okay, so what makes, we're really in a detail. 00:28:26.140 |
And they can also slide against each other pretty easily. 00:28:32.700 |
when the things tend to be these hexagonal plates, 00:28:40.000 |
that have grown out, they don't fit together very well. 00:28:43.080 |
And that's why the snow has lots of air in it. 00:28:47.000 |
and if you catch one, you'll see it has these little arms. 00:29:03.600 |
As they fell through the air in a different way, 00:29:17.960 |
- So the rules under which they grow are the same. 00:29:25.020 |
to describe the full complexity of snowflake growth. 00:29:29.480 |
- Well, science, if you do what people might often do, 00:29:33.560 |
which is say, okay, let's make it scientific. 00:29:38.560 |
And that one number is kind of the growth rate of the arms 00:29:46.280 |
And that's in a sense a big challenge for science, 00:29:48.700 |
is how do you extract from the natural world, for example, 00:30:00.780 |
All I care about is the growth rate of the arms, 00:30:06.140 |
without knowing anything about the fluffiness. 00:30:11.420 |
if you say, what is the most obvious feature of a snowflake? 00:30:24.820 |
A model is some way of reducing the actuality of the world 00:30:28.880 |
to something where you can readily sort of give a narrative 00:30:31.840 |
for what's happening, where you can basically 00:30:34.240 |
make some kind of abstraction of what's happening 00:30:36.880 |
and answer questions that you care about answering. 00:30:39.860 |
If you wanted to answer all possible questions 00:30:41.840 |
about the system, you'd have to have the whole system, 00:30:44.400 |
'cause you might care about this particular molecule. 00:30:47.280 |
And your model, which is some big abstraction of that, 00:30:52.300 |
So, one of the things that's often confusing in science 00:30:55.520 |
is people will have, I've got a model, somebody says. 00:30:58.000 |
Somebody else will say, I don't believe in your model 00:30:59.680 |
because it doesn't capture the feature of the system 00:31:07.240 |
Well, no model is, except for the actual system itself, 00:31:10.600 |
is a correct model in the sense that it captures everything. 00:31:13.760 |
Question is, does it capture what you care about capturing? 00:31:18.340 |
by what you're going to build technology out of, 00:31:23.840 |
if you think you're modeling the whole universe 00:31:25.680 |
all the way down, then there is a notion of a correct model. 00:31:31.840 |
because it depends on kind of how observers sample things 00:31:44.680 |
When you really think you have a complete model 00:31:46.840 |
for the whole universe, you better be capturing 00:31:49.060 |
ultimately everything, even though to actually run 00:31:56.400 |
The only thing that successfully runs that model 00:32:02.180 |
But okay, so what you care about is an interesting concept. 00:32:08.740 |
So that's what you're doing with Wolfram Alpha 00:32:11.340 |
and Wolfram Language, is you're trying to come up 00:32:14.740 |
with symbolic representations as simple as possible. 00:32:30.380 |
we'll have a thing about data about movies, let's say. 00:32:34.020 |
We could be describing every individual pixel 00:32:37.640 |
but that's not the level that people care about. 00:32:42.360 |
and that level that people care about is somewhat related 00:32:52.640 |
to sort of represent precisely so you can compute things. 00:33:00.900 |
You say, does the computer understand this natural language? 00:33:04.480 |
Well, the computer processes it in some way, it does this, 00:33:07.400 |
maybe it can make a continuation of the natural language. 00:33:15.700 |
Hard to know, but for in this kind of computational world, 00:33:20.700 |
there is a very definite definition of does it understand, 00:33:28.860 |
from which you can compute all kinds of consequences? 00:33:31.740 |
And that's the sense in which one has sort of a target 00:33:37.000 |
And that's kind of our goal is to have as much as possible 00:33:40.980 |
about the world that can be computed in a reasonable way, 00:33:58.520 |
it gives us a way of kind of building a structure 00:34:06.220 |
well, let's talk about it in natural language, 00:34:08.360 |
it doesn't really give us some hard foundation 00:34:10.880 |
that lets us build step by step to work something out. 00:34:14.000 |
I mean, it's kind of like what happens in math. 00:34:15.920 |
If we were just sort of vaguely talking about math, 00:34:19.160 |
but didn't have the kind of full structure of math 00:34:31.440 |
is to make a formalism for describing the world 00:34:40.560 |
between natural language and Wolfram language? 00:34:44.560 |
So there's this gigantic thing called the internet 00:34:47.040 |
where people post memes and diary type thoughts 00:34:52.040 |
and very important sounding articles and all of that 00:35:02.400 |
How can you map from the natural language of the internet 00:35:08.520 |
Is there a manual, is there an automated way of doing that 00:35:18.960 |
its kind of front end is turning natural language 00:35:26.720 |
you ask a question, what is the capital of some country? 00:35:35.460 |
And that will turn into geo distance of entity, city, 00:35:41.820 |
Each one of those things is very well defined. 00:35:44.340 |
We know, given that it's the entity, city, Chicago, 00:35:47.500 |
et cetera, et cetera, et cetera, Illinois, United States, 00:35:51.180 |
we know the geo location of that, we know its population, 00:35:56.220 |
which we have curated that data to be able to know that 00:36:06.900 |
And that's kind of the, yeah, that's the idea. 00:36:11.220 |
- But then something like GPT, large language models, 00:36:21.040 |
which we still don't know everything about, okay? 00:36:23.680 |
I mean, this question of going from natural language 00:36:30.580 |
we've now, Wolfram Alpha's been out and about 00:36:37.980 |
98%, 99% success on queries that get put into it. 00:36:42.980 |
Now, obviously, there's a sort of feedback loop 00:36:44.900 |
'cause the things that work are things people go on 00:36:47.700 |
So that, but we've got to a very high success rate 00:36:55.340 |
that people put in, questions, math calculations, 00:37:03.820 |
turning those things into computational language. 00:37:06.620 |
Now, from the very beginning of Wolfram Alpha, 00:37:14.020 |
In fact, I had, I was just looking at this recently, 00:37:19.420 |
called something like programming with natural language 00:37:29.900 |
some of them a little bit machine learning-like, 00:37:40.100 |
Actually, I know that post, a piece of utter trivia, 00:37:43.080 |
but that post, Steve Jobs forwarded that post 00:37:49.960 |
'cause he never really liked programming languages. 00:38:00.500 |
He would have liked, you know, I think what's happening now 00:38:02.680 |
because it really is the case that you can, you know, 00:38:07.440 |
how the computer works to use a programming language 00:38:10.800 |
is something that is, I think, a thing that, you know, 00:38:14.220 |
just like you had to learn the details of the opcodes 00:38:16.360 |
to know how assembly language worked and so on. 00:38:18.360 |
It's kind of a thing that's a limited time horizon, 00:38:26.800 |
of how elaborate can you make kind of the prompt, 00:38:30.640 |
how elaborate can you make the natural language 00:38:48.320 |
I mean, I'm still trying to understand this workflow. 00:38:53.120 |
- The natural language to computational language. 00:38:57.000 |
- And the process, especially if it's conversation, 00:38:59.120 |
like dialogue, it's like multiple queries kind of thing. 00:39:03.040 |
There's so many things that are really interesting 00:39:06.840 |
So first thing is, can you just walk up to the computer 00:39:12.560 |
What one realizes is humans have to have some idea 00:39:24.240 |
- I remember when I should tell a silly story about myself. 00:39:30.400 |
and it was a big mainframe computer and so on. 00:39:32.400 |
And I didn't really understand what computers did. 00:39:34.880 |
And it's like, somebody was showing me this computer 00:39:38.680 |
can the computer work out the weight of a dinosaur? 00:39:41.520 |
It's like, that isn't a sensible thing to ask. 00:39:43.880 |
That's kind of, you know, you have to give it, 00:39:49.440 |
you could say, what's the typical weight of a stegosaurus? 00:39:56.960 |
And so the kind of the question of, you know, 00:40:06.440 |
And, you know, I think it's a very, you know, 00:40:10.920 |
It's kind of this notion, not computer science, 00:40:20.360 |
Computation, thinking about the world computationally 00:40:23.720 |
is kind of this formal way of thinking about the world. 00:40:26.720 |
We've had other ones like logic was a formal way, 00:40:34.720 |
Computation is this very broad way of sort of formalizing 00:40:45.040 |
in terms of computation, computers can help us figure out 00:40:53.120 |
if you're not using a computer to do the math, 00:40:55.960 |
you have to go work out a bunch of stuff yourself. 00:40:58.720 |
So I think, but this idea, let's see, I mean, 00:41:07.560 |
and its relationship to computational language. 00:41:12.560 |
I think is first human has to have some kind of idea 00:41:17.520 |
That if it's something that they want to sort of build 00:41:22.760 |
something that they want to sort of formalize 00:41:26.080 |
So then human can type something into, you know, 00:41:30.040 |
some LLM system and sort of say vaguely what they want 00:41:49.000 |
with the Wolfram language translation of that. 00:41:54.880 |
where you can kind of extrapolating from all those examples 00:42:01.960 |
- Is the prompter task could also kind of debugging 00:42:11.280 |
Okay, so first, the first thing is you type natural language. 00:42:16.960 |
Do you have an example that is the dinosaur example? 00:42:22.080 |
that we should be thinking about some dumb example? 00:42:27.200 |
and, you know, figure out whether I, you know, 00:42:32.200 |
make a moving average every seven days or something 00:42:35.920 |
and work out what the, and make a plot of the results. 00:42:42.000 |
about two thirds of a line of Wolfram language code. 00:42:45.320 |
I mean, it's, you know, list plot of moving average 00:42:53.680 |
And, you know, the vague thing that I was just saying 00:42:56.080 |
in natural language could, would almost certainly 00:43:09.160 |
you arrive at the moving average kind of idea. 00:43:13.840 |
maybe it'll figure out that that's a moving, you know, 00:43:15.680 |
that that can be encapsulated as this moving average idea. 00:43:19.560 |
But then the typical workflow that I'm seeing is 00:43:24.040 |
you generate this piece of Wolfram language code. 00:43:27.480 |
It's, and if it isn't small, it probably isn't right. 00:43:41.920 |
programming languages tend to be this one way story 00:43:44.600 |
of humans write them and computers execute from them. 00:43:59.880 |
is kind of this, this human mumbles some things, 00:44:03.640 |
you know, large language model produces a fragment 00:44:11.200 |
yeah, that looks, well, typically you just run it first. 00:44:18.520 |
You look at the code, you see, I see why it's crazy. 00:44:25.080 |
you better look at that code and understand it 00:44:27.400 |
because that's the way you have the sort of checkpoint 00:44:29.960 |
of did it really do what I expected it to do? 00:44:39.640 |
Then you can often say to the large language model, 00:44:57.680 |
So you're debugging based on the output of the code, 00:45:02.440 |
The plugin that we have, you know, for chat GPT, 00:45:10.920 |
It will discover, the LLM will discover itself 00:45:20.840 |
and then it will try it again and get the result. 00:45:37.520 |
where you have, you know, text and code and output? 00:45:41.080 |
How do you combine that with the notion of chat and so on? 00:45:44.280 |
And there's some really interesting things there. 00:45:48.560 |
is we have these notebooks where as soon as the, 00:45:59.400 |
the LLM automatically not only looks at those messages, 00:46:03.400 |
it can also see all kinds of internal information 00:46:07.960 |
And it can then, it does a remarkably good job 00:46:24.520 |
that we humans would kind of glaze over looking at. 00:46:29.840 |
oh, this is the explanation of what's happening. 00:46:43.760 |
There's examples of where the messages have occurred. 00:46:47.440 |
The other thing that's really amusing with this 00:46:53.080 |
when the code doesn't work is read the documentation. 00:47:03.960 |
'Cause it will figure out, sometimes it'll get, 00:47:07.440 |
it'll make up the name of some option for some function 00:47:10.560 |
that doesn't really exist, read the documentation. 00:47:12.920 |
It'll have some wrong structure for the function and so on. 00:47:21.200 |
we built this language over the course of all these years 00:47:23.560 |
to be nice and coherent and consistent and so on. 00:47:28.040 |
Turns out there was a side effect that I didn't anticipate, 00:47:31.160 |
which is it makes it easier for AIs to understand. 00:47:33.960 |
- So it's almost like another natural language. 00:47:36.440 |
But so Wolfram language is a kind of foreign language. 00:47:42.040 |
- You have a lineup, English, French, Japanese, 00:47:45.440 |
Wolfram language, and then, I don't know, Spanish. 00:47:49.920 |
And then the system is not gonna notice, hopefully. 00:47:56.560 |
'Cause it really depends on what I see as being 00:48:02.400 |
that basically just jumped out at us with Chachapiti. 00:48:05.940 |
'Cause I think, you know, the real question is, 00:48:14.280 |
to successfully reproduce all these kinds of things 00:48:17.260 |
in natural language, you know, with a, you know, 00:48:27.160 |
And I think that, you know, that relates to kind of 00:48:29.960 |
a fundamental fact about language, which, you know, 00:48:33.760 |
the main thing is that I think there's a structure 00:48:36.920 |
to language that we haven't kind of really explored 00:48:46.040 |
I mean, we kind of know that when we set up human language, 00:48:52.440 |
We know that it has a certain grammatical structure, 00:48:55.400 |
you know, noun followed by verb, followed by noun, 00:49:03.200 |
But I think the thing that Chachapiti is showing us 00:49:06.040 |
is that there's an additional kind of regularity 00:49:09.040 |
to language, which has to do with the meaning 00:49:11.260 |
of the language beyond just this pure, you know, 00:49:16.360 |
And I think the kind of the one example of that 00:49:22.860 |
And, you know, I think my sort of kind of picture 00:49:27.860 |
of how was logic invented, how was logic discovered? 00:49:42.360 |
and this one made sense, that one doesn't make sense, 00:49:46.000 |
this one, and, you know, you see these patterns of, 00:49:49.480 |
you know, if the, you know, I don't know what, 00:49:55.300 |
then the this does that, et cetera, et cetera, et cetera. 00:49:58.680 |
And what Aristotle realized is there's a structure 00:50:02.240 |
to those sentences, there's a structure to that rhetoric 00:50:05.140 |
that doesn't matter whether it's the Persians and the Greeks 00:50:10.080 |
It's just, you know, P and Q, you can abstract 00:50:13.120 |
from the details of these particular sentences, 00:50:16.420 |
you can lift out this kind of formal structure, 00:50:38.860 |
I mean, Aristotle had this idea of syllogistic logic, 00:50:48.920 |
part of education was you memorized the syllogisms, 00:50:51.760 |
I forget how many there were, but 15 of them or something. 00:50:54.680 |
And they all had names, they all had mnemonics, 00:50:58.400 |
were two of the mnemonics for the syllogisms. 00:51:01.240 |
And people would kind of, this is a valid argument 00:51:03.520 |
'cause it follows the Barbara syllogism, so to speak. 00:51:06.680 |
And it took until 1830, you know, with George Boole 00:51:11.360 |
to kind of get beyond that and kind of see that 00:51:21.040 |
And that's, you know, what's interesting there is, 00:51:29.400 |
It's essentially dealing with templates of sentences. 00:51:32.500 |
By the time you get to Boole and Boolean algebra 00:51:35.200 |
and this idea of, you know, you can have arbitrary depth 00:51:43.120 |
that's the kind of thing, that's a computation story. 00:51:45.760 |
That's, you know, you've gone beyond the pure sort of 00:51:53.120 |
But the thing that I think we realized from chat-GBT 00:51:59.440 |
And there was more that you could have lifted 00:52:18.920 |
I mean, whether it's, I don't know, if you say, 00:52:21.280 |
I go from place A to place B, place B to place C, 00:52:26.600 |
then I know I've gone from place A to place C. 00:52:29.520 |
If A is a friend of B and B is a friend of C, 00:52:32.360 |
it doesn't necessarily follow that A is a friend of C. 00:52:35.280 |
These are things that are, you know, that there are, 00:52:39.040 |
if you go from place A to place B, place B to place C, 00:52:44.620 |
Like logic, it doesn't matter whether you flew there, 00:52:54.980 |
And there are many kinds of kind of features, I think, 00:53:04.880 |
And I think what "Chat GPT" effectively has found, 00:53:15.040 |
discovered logic, by looking at a lot of sentences, 00:53:17.680 |
effectively, and noticing the patterns in those sentences. 00:53:20.760 |
- But it feels like it's discovering something 00:53:36.000 |
- Yes, that was the title that George Boole had 00:53:38.340 |
for his Boolean algebra back in 1830, but yes. 00:53:45.880 |
- So he thought he nailed it with Boolean algebra. 00:53:50.520 |
- And it's a good question, how much more is there to it? 00:53:56.300 |
as you imply that the reason "Chat GPT" works 00:54:01.200 |
is that there's a finite number of things to it. 00:54:09.880 |
GPT's discovering the laws of semantic grammar 00:54:22.700 |
They're just not ones that we humans have cared about 00:54:26.600 |
and operate with, and that's probably because our brains 00:54:30.280 |
are built in a certain way, and the neural nets 00:54:33.400 |
of our brains are not that different, in some sense, 00:54:35.700 |
from the neural nets of a large language model, 00:54:39.220 |
and that's kind of, and so when we think about, 00:54:43.680 |
but when we think about sort of what will AIs ultimately do, 00:54:47.080 |
the answer is, insofar as AIs are just doing computation, 00:54:52.840 |
of crazy computations, but the ones that we sort of have, 00:54:57.760 |
have decided we care about is this kind of very limited set. 00:55:08.220 |
The more the AIs say the stuff that kind of interests us, 00:55:13.540 |
So it can do a lot of interesting, intelligent things, 00:55:34.680 |
Only a limited set of those are ones that we capture 00:55:38.200 |
and use for technology, 'cause they're only a limited set 00:55:43.640 |
that we can sort of apply to the human purposes 00:55:55.340 |
Then you realize, oh, we could actually turn this 00:56:04.160 |
- And it's, you know, it's this thing about what do we, 00:56:06.820 |
you know, in the evolution of our civilization, 00:56:09.780 |
what things do we identify as being things we care about? 00:56:13.880 |
when there was a little announcement recently 00:56:16.280 |
of the possibility of a high temperature superconductor 00:56:18.720 |
that involved, you know, the element lutetium, 00:56:21.160 |
which, you know, generally nobody has cared about. 00:56:32.760 |
- So given your thinking that GPT may have discovered 00:56:46.540 |
I mean, the fact is, look, the logic is but the first step. 00:56:51.540 |
There are many other kinds of calculi about things 00:56:58.580 |
about sort of things that happen in the world 00:57:02.780 |
- Well, how do you know logic is not the last step? 00:57:06.660 |
- Well, because we can plainly see that that thing, 00:57:10.620 |
here's a sentence that is syntactically correct, okay? 00:57:16.740 |
the happy electron, you know, ate, I don't know what, 00:57:24.220 |
you look at it and it's like, this is meaningless. 00:57:29.140 |
The nouns and the verbs are in the right place, 00:57:38.860 |
when a sentence has the potential to be meaningful 00:57:41.980 |
that go beyond the pure parts of speech syntax. 00:57:45.620 |
And so the question is, what are those rules? 00:57:47.820 |
And are there a fairly finite set of those rules? 00:57:50.700 |
My guess is that there's a fairly finite set of those rules. 00:57:53.940 |
And they, you know, once you have those rules, 00:58:10.980 |
Those sentences may not be realized in the world. 00:58:13.540 |
I mean, I think, you know, the elephant flew to the moon. 00:58:21.660 |
If I say that to you, you kind of know what that means. 00:58:30.220 |
is things that can be imagined with the human mind. 00:58:34.220 |
Things that are consistent with both our imagination 00:58:56.140 |
with the things we're describing in language. 00:59:07.140 |
Okay, let's take the example of motion, okay? 00:59:13.260 |
that should have been figured out by the Greeks, 00:59:16.460 |
but it's actually a really pretty complicated concept, 00:59:19.900 |
Motion is you can go from place A to place B, 00:59:23.220 |
and it's still you when you get to the other end, right? 00:59:43.300 |
by the time it's near the space-time singularity. 00:59:49.420 |
But, so that's a case where pure motion doesn't really work. 01:00:06.980 |
as being the same thing, but in a different place, 01:00:10.300 |
that sort of abstracted idea then has, you know, 01:00:17.700 |
go from A to B, B to C, you've gone from A to C. 01:00:23.700 |
you can have what are sort of inevitable consequences. 01:00:31.220 |
And that's, I think, what this sort of semantic grammar 01:00:40.060 |
When you say, I go from, I move from here to there, 01:00:42.700 |
well, it's complicated to say what that means. 01:00:46.100 |
is pure motion possible, et cetera, et cetera, et cetera. 01:00:48.600 |
But once you have kind of got an idea of what that means, 01:00:51.780 |
then there are inevitable consequences of that idea. 01:00:57.940 |
it seems like there's some words that become, 01:01:00.220 |
it's like there's a latent ambiguity to them. 01:01:06.620 |
I mean, it's the word like emotionally loaded words, 01:01:44.780 |
we expect that that construct is a building block 01:01:48.240 |
from which we can construct an arbitrarily tall tower. 01:01:50.980 |
So we have to have a very solid building block. 01:01:53.460 |
And, you know, we have to, it turns into a piece of code, 01:02:04.440 |
well, there isn't a standard documentation for that word, 01:02:08.060 |
It's a complicated thing defined by kind of how we use it. 01:02:20.140 |
so that we can communicate them to another mind. 01:02:26.220 |
into something that a computation engine can use? 01:02:32.620 |
that what one can do in computational language 01:02:46.740 |
it's, you know, animals eat things, whatever else. 01:02:58.180 |
But if you start to say, well, what are the implications 01:03:11.400 |
But, you know, in what, how far does that analogy go? 01:03:25.040 |
which you anchor to the kind of natural language concept eat, 01:03:29.100 |
but it is now some precise definition of that, 01:03:33.420 |
- But don't you think the analogy is also precise? 01:03:45.820 |
But the thing that sort of is the first target 01:03:48.580 |
for computational language is to take sort of 01:03:52.300 |
the ordinary meaning of things and try and make it precise, 01:03:57.280 |
you can build these towers of computation on top of it. 01:03:59.780 |
So it's kind of like, if you start with a piece of poetry 01:04:16.280 |
And that's how I'm communicating with my computer. 01:04:19.460 |
And that's how I'm going to build the solid building block 01:04:25.220 |
- So there's some sense where if you take a poem 01:04:31.780 |
So maybe there's a bunch of human interaction 01:04:39.540 |
That's just like recreational hamstring a wheel. 01:04:45.860 |
- Well, I think that that's a complicated thing 01:04:48.580 |
because in a sense, human linguistic communication 01:04:51.300 |
is there's one mind, it's producing language, 01:04:55.300 |
that language is having an effect on another mind. 01:04:58.500 |
And the question of the sort of a type of effect 01:05:04.700 |
where for example, it's very independent of the two minds. 01:05:31.620 |
feels more amenable to the definition of purpose. 01:05:34.560 |
It's like, yeah, you're given two clean representations 01:05:39.560 |
of a concept and you can build a tower based on that. 01:05:42.760 |
Is natural language the same thing but more fuzzy? 01:05:46.160 |
- Well, I think the story of natural language, 01:05:51.760 |
We don't know whether it exists in other species, 01:05:56.320 |
It's the thing that allows you to sort of communicate 01:06:04.400 |
You can, you know, there is an abstract version 01:06:12.200 |
you don't have to apprentice the next species, 01:06:14.240 |
you know, the next generation of birds to the previous one 01:06:18.800 |
There is this abstracted version of knowledge 01:06:26.120 |
it still tends to rely because language is fuzzy. 01:06:29.280 |
It does tend to rely on the fact that, you know, 01:06:31.360 |
if we look at the, you know, some ancient language 01:06:34.820 |
that where we don't have a chain of translations from it 01:06:48.240 |
but it is something where we can realistically expect 01:07:11.360 |
- Do you see natural language and thought as the same? 01:07:16.240 |
- Well, that's been a long debate in philosophy. 01:07:31.240 |
in the human mind seems something like intelligence. 01:07:43.440 |
the language of thought, the laws of thought, 01:07:59.520 |
I mean, much of what computers do, humans do not do. 01:08:12.480 |
who needs computation when we have large language models? 01:08:17.000 |
eventually you'll have a big enough neural net, 01:08:23.600 |
And there are plenty of sort of formal things 01:08:29.720 |
you can, some people can do mental arithmetic, 01:08:31.840 |
they can do a certain amount of math in their minds. 01:08:34.560 |
I don't think many people can run a program in their minds 01:08:41.160 |
it's not something people have even thought of doing, 01:08:43.560 |
'cause it's just, it's kind of not, you know, 01:08:53.240 |
But if I say to you, here's a Turing machine, 01:08:56.320 |
you know, tell me what it does after 50 steps. 01:08:58.960 |
And you're like, trying to think about that in your mind. 01:09:09.200 |
they program it, just to answer your question 01:09:26.880 |
I mean, they've outsourced that piece of computation 01:09:30.720 |
from something that is internally happening in their minds 01:09:39.440 |
didn't invent computers, they discovered them. 01:09:55.080 |
to plug into this whole stream of computation. 01:10:00.520 |
There's probably a lot of ways. - Well, for sure. 01:10:01.360 |
I mean, the particular ways that we make computers 01:10:05.040 |
out of semiconductors and electronics and so on, 01:10:07.560 |
that's the particular technology stack we built. 01:10:10.560 |
I mean, the story of a lot of what people try to do 01:10:24.040 |
that's different from semiconductors and electronics. 01:10:26.760 |
It's a molecular scale sort of computational process 01:10:33.400 |
I have some ideas about understanding more about that. 01:10:43.000 |
at the level of these evolving hypergraphs and so on, 01:10:52.200 |
- So if GPT or large language models are starting to form, 01:11:11.440 |
I mean, what is happening in natural science? 01:11:13.120 |
You have the world that's doing all these complicated things 01:11:16.560 |
and then you discover, you know, Newton's laws, for example. 01:11:30.160 |
It's there are sort of computationally reducible aspects 01:11:38.200 |
just as we've got narrative theories in physics and so on. 01:11:44.840 |
- Do you think it will be depressing or exciting 01:11:48.840 |
when all the laws of thought are made explicit, 01:12:04.040 |
people will say, "Oh, but, you know, I have free will. 01:12:08.080 |
"I kind of, you know, I operate in a way that is, 01:12:13.280 |
"you know, they have the idea that they're doing something 01:12:20.460 |
"that they're figuring out what's happening." 01:12:22.520 |
But in fact, we think there are laws of physics 01:12:27.320 |
every electrical impulse in a nerve and things like this. 01:12:48.800 |
to the way that we might construct a piece of text 01:12:51.800 |
than it is to get from individual nerve firings 01:12:58.400 |
And by the way, as soon as we have this kind of level of, 01:13:03.520 |
it's kind of, it helps us to go even further. 01:13:12.800 |
if we didn't have a computer and we knew certain rules, 01:13:15.640 |
we could write them down, we'd go a certain distance. 01:13:17.840 |
But once we have a computer, we can go vastly further, 01:13:23.680 |
"What is Chad GPT doing and why does it work?" 01:13:27.760 |
but can we just step back and linger on this question? 01:13:33.600 |
What are these, a bunch of billion parameters 01:13:53.760 |
- Well, let's talk about sort of the low level 01:14:08.440 |
that this kind of low level, dumb training procedure 01:14:13.440 |
can create something syntactically correct first, 01:14:25.760 |
can do much more complicated things than you imagine. 01:14:35.080 |
vastly more complicated than you can imagine. 01:14:40.200 |
I've sort of been thinking about this now 40 years or so, 01:14:44.840 |
I mean, even for example, in our physics project, 01:14:53.000 |
you know, how can something really complicated 01:15:13.360 |
But it's worth understanding kind of how that's working. 01:15:15.280 |
I mean, it's kind of like, if it was going to say, 01:15:18.440 |
you know, the cat sat on the, what's the next word? 01:15:22.580 |
Okay, so how does it figure out the next word? 01:15:24.720 |
Well, it's seen a trillion words written on the internet, 01:15:30.440 |
the cat sat on the sofa, the cat sat on the whatever. 01:15:39.840 |
We saw, you know, 10,000 examples of the cat sat on the, 01:15:47.300 |
Let's just pick that out and say that's the next word. 01:15:50.600 |
And that's kind of what it, at some level, is trying to do. 01:15:54.460 |
Now, the problem is there isn't enough text on the internet 01:15:57.920 |
to, if you have a reasonable length of prompt, 01:16:02.320 |
that specific prompt will never have occurred 01:16:08.600 |
there just won't be a place where you could have trained, 01:16:11.960 |
you know, where you could just worked out probabilities 01:16:17.840 |
there'll be a zillion examples of two plus two equaling four 01:16:21.400 |
and a very small number of examples of two plus two 01:16:25.200 |
And you can pretty much know what's going to happen. 01:16:31.880 |
what's going to happen, just no probabilistic, 01:16:38.720 |
this idea of making models of things is an idea that really, 01:16:42.480 |
I don't know, I think Galileo probably was one of the first 01:16:48.600 |
I think I gave an example of that little book I wrote 01:16:51.960 |
about Chachi B'tee where it's kind of like, you know, 01:16:54.920 |
Galileo was dropping cannonballs off the different floors 01:17:00.360 |
And it's like, okay, you drop a cannonball off this floor, 01:17:04.900 |
you miss floor five or something for whatever reason, 01:17:09.640 |
to fall to the ground from floors one, two, three, four, 01:17:21.000 |
how long would it take the ball to fall to the ground 01:17:23.440 |
from the floor you didn't explicitly measure? 01:17:25.720 |
And the thing Galileo realized is that you can use math, 01:17:29.440 |
you can use mathematical formulas to make a model 01:17:44.400 |
and is this arrangement of pixels an A or a B? 01:17:47.440 |
Does it correspond to something we'd recognize 01:17:52.480 |
each pixel is like a parameter in some equation 01:17:57.760 |
where the answer is either, you know, A or, you know, 01:18:02.000 |
And the question is then what kind of a model 01:18:04.960 |
successfully reproduces the way that we humans 01:18:08.840 |
would conclude that this is an A and this is a B? 01:18:12.400 |
You know, if there's a complicated extra tail 01:18:21.600 |
into the way that we humans make distinctions about things? 01:18:31.080 |
It could be that human distinctions are not captured. 01:18:34.720 |
You know, we could try searching around for a type of model 01:18:39.240 |
it could be some model based on something else 01:18:41.240 |
that captures kind of typical human distinctions 01:18:44.720 |
It turns out this model that actually is very much 01:18:48.500 |
the way that we think the architecture of brains works, 01:18:57.780 |
And so, you know, the core next point is that 01:19:03.580 |
makes sort of distinctions and generalizes things 01:19:07.980 |
in sort of the same way that we humans do it. 01:19:15.940 |
even though it never, it didn't see many examples 01:19:20.260 |
it can make a, or the aardvark sat on the green whatever, 01:19:28.420 |
And so it has to make a model that concludes what, 01:19:33.940 |
from what it's, from the actual examples that it's seen. 01:19:37.220 |
And so, you know, that's the fact is that neural nets 01:19:41.320 |
generalize in the same kind of way that we humans do. 01:19:47.540 |
at our neural net generalizations and say, that's crazy. 01:19:50.260 |
You know, that thing, when you put that extra little dot 01:19:55.300 |
That's, you know, that messed the whole thing up. 01:20:00.700 |
which seem to correspond to the kinds of distinctions 01:20:04.460 |
So then, you know, the thing that is just amazing to me 01:20:07.320 |
about ChachiBT is how similar the structure it has is 01:20:11.580 |
to the very original way people imagine neural nets 01:20:16.660 |
And, you know, there's a lot of detailed engineering, 01:20:19.420 |
you know, great cleverness, but it's really the same idea. 01:20:22.820 |
And in fact, even the sort of elaborations of that idea 01:20:27.900 |
particular structure to try and make the neural net 01:20:34.660 |
I mean, there's some things that seem to, you know, 01:20:36.600 |
when you train this neural net, you know, the one thing, 01:20:39.920 |
this kind of transformer architecture, this attention idea, 01:20:43.100 |
that really has to do with, does every one of these neurons 01:20:47.240 |
connect to every other neuron, or is it somehow 01:20:52.020 |
Does it like we're making a sequence of words 01:20:57.300 |
rather than just everything can depend on everything. 01:21:00.180 |
And that seems to be important in just organizing things 01:21:03.060 |
so that you don't have a sort of a giant mess. 01:21:05.900 |
But the thing, you know, the thing worth understanding 01:21:11.260 |
A neural net in the end is each neuron has a, 01:21:15.260 |
it's taking inputs from a bunch of other neurons. 01:21:18.460 |
It's eventually, it's going to have a numerical value. 01:21:23.460 |
It's going to compute some number, and it's saying, 01:21:36.220 |
Then it's gonna add those up and multiply them 01:21:38.260 |
by these weights, and then it's going to apply 01:21:40.920 |
some function that says if it's bigger than zero 01:21:50.540 |
- It's a giant equation with a lot of variables. 01:21:54.140 |
You mentioned figuring out where the ball falls 01:21:57.060 |
when you don't have data on the fourth floor. 01:21:59.300 |
This, the equation here is not as simple as-- 01:22:03.740 |
- Right, it's an equation with 175 billion terms. 01:22:07.140 |
- And it's quite surprising that in some sense, 01:22:10.540 |
a simple procedure of training such an equation 01:22:15.540 |
can lead to a good representation of natural language. 01:22:26.180 |
you've turned, so neural nets always just deal with numbers. 01:22:33.220 |
that you started with into a bunch of numbers. 01:22:38.740 |
you just map each word or each part of a word 01:22:46.260 |
and then those numbers just go into the values 01:22:51.680 |
it's just rippling down, going layer to layer, 01:22:58.260 |
And you're just, you know, it just goes once through. 01:23:03.740 |
just says, here are the numbers from the words before. 01:23:09.020 |
It computes the probabilities that it estimates 01:23:11.960 |
for each of the possible 50,000 words that could come next. 01:23:16.580 |
sometimes it will use the most probable word, 01:23:18.420 |
sometimes it will use not the most probable word. 01:23:22.060 |
that there's this so-called temperature parameter, 01:23:28.420 |
that it estimated was the most probable thing to come next. 01:23:36.980 |
It'll go down to lower and lower probability words. 01:23:39.420 |
Thing I was just playing with actually recently 01:23:44.540 |
The thing goes bonkers at a particular, you know, 01:23:49.420 |
I think maybe about 1.2 is the thing I was noticing 01:23:54.900 |
that, you know, usually it's giving reasonable answers. 01:23:57.500 |
And then at that temperature with some probability, 01:24:04.900 |
And, you know, nobody knows why this happens. 01:24:18.520 |
Now let's take the whole thing I wrote so far. 01:24:26.060 |
is really important to the operation of the thing. 01:24:28.560 |
And for example, one of the things that is kind of funky 01:24:41.620 |
Well, the answer is because it is going one word at a time, 01:24:47.340 |
it came along with some sort of chain of thought in a sense, 01:24:51.500 |
and it came up with completely the wrong answer. 01:25:01.140 |
It immediately can recognize that was a, you know, 01:25:03.580 |
a bad syllogism or something, and can see what happened. 01:25:06.980 |
Even though, as it was being led down this garden path, 01:25:09.860 |
so to speak, it didn't, it came to the wrong place. 01:25:13.100 |
- But it's fascinating that this kind of procedure 01:25:30.700 |
there are many things we don't understand, okay? 01:25:32.580 |
So for example, you know, 175 billion weights, 01:25:36.340 |
it's maybe about a trillion bytes of information, 01:25:39.580 |
which is very comparable to the training set that was used. 01:25:49.460 |
that the number of weights in the neural net, 01:25:59.220 |
insofar as there are definite rules of what's going on, 01:26:06.320 |
that will successfully capture what's happening. 01:26:13.340 |
when you don't know any other way to structure the thing. 01:26:17.560 |
if you don't know any other way to structure the thing. 01:26:20.700 |
we haven't known any other way to structure it. 01:26:33.260 |
you can kind of get rid of much of the structure 01:26:38.300 |
which are sort of pure steps of computation, so to speak, 01:26:42.180 |
sort of with neural net stuff around the edges. 01:26:47.900 |
- So the neural net, you hope, will reveal to us 01:26:52.180 |
good symbolic rules that make the need of the neural net 01:26:58.260 |
And there will still be some stuff that's kind of fuzzy, 01:27:03.180 |
it's like this question of what can we formalize? 01:27:05.540 |
What can we turn into computational language? 01:27:08.580 |
What is just sort of, oh, it happens that way 01:27:16.340 |
of large language models, just to make it explicit? 01:27:20.260 |
- Well, I mean, I think that deep computation 01:27:24.060 |
I mean, that's just, it's a different kind of thing. 01:27:26.500 |
You know, the outer loop of a large language model, 01:27:29.060 |
if you're trying to do many steps in a computation, 01:27:36.180 |
the whole chain of thought as a bunch of words, basically. 01:27:39.540 |
And, you know, you can make a Turing machine out of that, 01:27:45.180 |
You know, in principle, you can make an arbitrary 01:27:50.260 |
but it's a bizarre and inefficient way to do it. 01:27:57.860 |
I think that's, you know, sort of the deep computation. 01:28:07.900 |
large language models will probably be able to do well. 01:28:10.580 |
Anything that you can do kind of off the top of your head 01:28:17.260 |
And the things you do off the top of your head, 01:28:18.660 |
you may not get them always right, but, you know, 01:28:21.700 |
you'll, it's thinking it through the same way we do. 01:28:26.140 |
- But I wonder if there's an automated way to do something 01:28:28.620 |
that humans do well, much faster to where it like loops. 01:28:49.020 |
So, you know, the thing is when people say, you know, 01:28:58.060 |
In a sense, it's sort of a failure of computational language 01:29:01.740 |
if the thing you have to build, in other words, 01:29:03.620 |
if we have a description, if you have a small description, 01:29:16.820 |
when as soon as you're giving a description that, you know, 01:29:20.220 |
if you have to somehow make that description something, 01:29:28.820 |
I'm gonna give this piece of natural language, 01:29:31.020 |
and then it's gonna split out this giant formal structure, 01:29:35.320 |
that in a sense that doesn't really make sense 01:29:39.020 |
because except in so far as that piece of natural language 01:29:42.820 |
kind of plugs into what we socially know, so to speak, 01:30:00.540 |
- Okay, I can make it more explicit if you want. 01:30:23.140 |
to take control and be the president, I don't know. 01:30:29.780 |
of each congressman, each senator, anybody, I don't know, 01:30:34.100 |
what's required to really kind of pass legislation 01:30:37.020 |
and figure out how to control them and manipulate them, 01:30:41.380 |
What would be the biggest fear of this congressman 01:30:45.060 |
and in such a way that you can take action on it 01:30:50.940 |
So maybe threaten the destruction of reputation 01:30:57.660 |
you know, to what extent can a large language model 01:31:12.460 |
- I'm also surprised how quickly I was able to generate 01:31:32.300 |
- Well, I'll give you a much more benign version 01:31:38.380 |
and that's a benign version of what you're saying, 01:31:42.020 |
is I want this person to understand this point. 01:32:05.340 |
sort of a large language model type technology 01:32:17.420 |
because, you know, sort of individualized teaching 01:32:20.940 |
is a thing that has been kind of a goal for a long time. 01:32:38.780 |
what are the three things I need to know, you know, 01:32:44.940 |
what's, let's say I'm looking at some paper or something, 01:32:59.580 |
- It could understand the major gaps in your knowledge 01:33:08.860 |
- Right, and that's a, you know, that's an important thing 01:33:12.020 |
because it really changes, actually, I think, you know, 01:33:16.380 |
it really changes kind of what's worth doing, 01:33:31.500 |
but turns out sort of there are sort of meta methods 01:33:41.980 |
it becomes easier to be fed knowledge, so to speak. 01:33:54.180 |
It's something I think is sort of an interesting feature. 01:34:00.020 |
things like the value of big towers of specialized knowledge 01:34:05.100 |
become less significant compared to the kind of 01:34:15.340 |
I think that, you know, there's been this huge trend 01:34:20.820 |
we have to sort of ascend these towers of knowledge. 01:34:26.420 |
more automation of being able to get to that place 01:34:29.860 |
on the tower without having to go through all those steps, 01:34:36.820 |
in terms of the collective intelligence of the species 01:34:41.700 |
and the individual minds that make up that collective, 01:34:54.540 |
I think that's where the humans are gonna be useful. 01:34:59.860 |
the drilling, the mechanical working out of things 01:35:13.940 |
I mean, you know, the kind of the specialization, 01:35:20.980 |
we've accumulated lots of knowledge in our species. 01:35:31.780 |
it becomes less necessary to know that whole tower. 01:35:34.660 |
And it becomes something where you can just use a tool 01:35:38.500 |
I think that, you know, the thing that is ultimately, 01:35:43.420 |
okay, what do the AIs do versus what do the humans do? 01:35:54.700 |
We say, what objective would you like to achieve? 01:36:02.640 |
That's a thing which has to come from some other, 01:36:07.580 |
And insofar as we are in charge, so to speak, 01:36:11.420 |
and our kind of web of society and history and so on 01:36:19.600 |
that's, you know, that's a thing that we humans 01:36:27.140 |
don't you think that GPT, future versions of GPT 01:36:39.820 |
look, here's the terrible thing that could happen, okay? 01:36:52.220 |
the most entertaining outcome is the most likely. 01:36:55.980 |
- Okay, I haven't heard that one from him, yeah. 01:37:37.620 |
So I think the answers will become more and more interesting 01:37:42.620 |
as these language models are trained better and better. 01:37:47.860 |
it's a reflection back of what we've already said. 01:37:58.820 |
- Isn't that what we're trying to do as a society? 01:38:06.220 |
I mean, in so far as some of us work on trying to innovate 01:38:13.700 |
it is sometimes, it's a complicated interplay 01:38:16.660 |
between sort of the individual doing the crazy thing 01:38:19.900 |
off in some spur, so to speak, versus the collective 01:38:24.620 |
that's trying to do sort of the high inertia average thing. 01:38:29.620 |
And it's, you know, sometimes the collective, you know, 01:38:36.180 |
and sometimes it's pulling down kind of the attempt 01:38:41.620 |
- Well, don't you think the large language models 01:38:45.260 |
will say maybe intellectual and career diversity 01:39:00.660 |
through this kind of dynamics that we've been using 01:39:05.580 |
focus on one thing, and then there's the crazy people 01:39:07.860 |
on the outskirts doing the opposite of that one thing, 01:39:10.780 |
and you kind of pull the whole society together. 01:39:17.860 |
and it's just been the history of human civilization. 01:39:20.420 |
And maybe the AI system will be able to see that. 01:39:30.460 |
and the more we'll be willing to let it run our society. 01:39:42.580 |
- Right, well, I mean, look, one of the things 01:40:04.140 |
many religions have taken that point of view. 01:40:06.260 |
There is a sacred book that got written at time X, 01:40:09.620 |
and it defines how people should act for all future time. 01:40:13.820 |
And it's a model that people have operated with. 01:40:18.540 |
And in a sense, this is a version of that kind of statement. 01:40:37.020 |
Because just like with religious texts and with GPT, 01:40:53.020 |
It's not like Chad GPT will tell you exactly what to do. 01:41:00.100 |
it's like a turn the other cheek kind of narrative, right? 01:41:07.460 |
- Well, until the AIs control all the systems in the world. 01:41:11.300 |
They will be able to very precisely tell you what to do. 01:41:24.500 |
So I think it's a slightly more prescriptive situation 01:41:41.340 |
this idea that there is an existing kind of corpus 01:41:45.540 |
of purpose for humans defined by what's on the internet 01:41:57.460 |
as we explore all these different possibilities 01:42:00.260 |
all these different inventions we could make, 01:42:03.620 |
the question is, which ones do we choose to follow? 01:42:06.460 |
Those choices are the things that, in a sense, 01:42:09.860 |
if the humans want to still have kind of human progress, 01:42:14.860 |
that's what we get to make those choices, so to speak. 01:42:22.400 |
if you say, let's take the kind of what exists today 01:42:33.540 |
the thing that is sort of the opportunity for humans 01:42:36.100 |
is there will be many possibilities thrown up. 01:42:38.940 |
There are many different things that could happen 01:42:46.780 |
the thing that makes sense for us to be in the loop doing 01:42:49.260 |
is picking which of those possibilities we want. 01:42:51.880 |
- But the degree to which there's a feedback loop, 01:43:00.420 |
because we're influenced by the various systems. 01:43:03.860 |
- If that becomes more and more source of our education 01:43:17.060 |
That's really the thing that makes the AIs take over. 01:43:21.620 |
- We will no longer write emails to each other. 01:43:29.180 |
But the thing where humans are potentially in the loop 01:43:33.740 |
And when there's a choice which we could make 01:43:37.060 |
based on our kind of whole web of history and so on, 01:43:40.100 |
and that's insofar as it's all just determined 01:43:52.340 |
it's all kind of a complicated philosophical issue 01:44:02.740 |
that are necessarily doing what we do, so to speak. 01:44:06.480 |
Yet, we feel we have sort of agency in what we're doing, 01:44:10.220 |
and that's its own separate kind of interesting issue. 01:44:13.620 |
- And we also kind of feel like we're the final destination 01:44:22.600 |
and likely are some kind of intermediate step, obviously. 01:44:30.420 |
The question is if there's some cooler, more complex, 01:44:34.040 |
more interesting things that's going to be materialized. 01:44:37.020 |
- The computational universe is full of such things. 01:44:39.460 |
- But in our particular pocket, specifically. 01:44:45.340 |
- We can make all kinds of interesting things 01:44:49.140 |
When we look at them, we say, yeah, you know, 01:44:54.380 |
It doesn't really connect with our current way 01:45:09.620 |
We just go out into the universe of possible theorems 01:45:26.500 |
about sort of understanding AI and its relationship to us 01:45:29.960 |
is as we have this kind of whole infrastructure 01:45:32.660 |
of AIs doing their thing and doing their thing 01:45:35.860 |
in a way that is perhaps not readily understandable 01:45:44.660 |
that behaves in a way that we can't understand 01:45:53.260 |
is run by AIs whose operations we can't understand? 01:46:01.040 |
That's what happens when we exist in the natural world. 01:46:10.440 |
We don't understand what the natural world is doing. 01:46:12.960 |
Occasionally, you know, when you say, you know, 01:46:18.220 |
Well, it's kind of like, is the machination of the AIs 01:46:22.480 |
that eventually comes and destroys the species? 01:46:31.520 |
that's going to, you know, make the earth explode 01:46:41.780 |
that's a result of science and natural science and so on. 01:46:46.760 |
when there's this giant infrastructure of the AIs 01:47:00.160 |
we have a, I don't know, a horse or something, 01:47:06.680 |
We don't really understand how the horse works inside, 01:47:09.280 |
but we can get certain rules and certain, you know, 01:47:12.000 |
approaches that we take to persuade the horse 01:47:27.600 |
we can find these kind of pockets of reducibility 01:47:34.120 |
we're grabbing onto the mane of the horse or something 01:47:39.720 |
or we figure out, you know, if we do this or that 01:47:42.800 |
to ride the horse, that that's a successful way 01:47:46.240 |
to get it to do what we're interested in doing. 01:48:01.080 |
So let me just ask you about big philosophical question 01:48:05.760 |
There's a lot of people like Eliezer Yudkowsky 01:48:08.200 |
who worry about the existential risks of AI systems. 01:48:37.280 |
- I think some of these arguments about kind of, 01:48:43.080 |
and eventually the AIs will get smarter than us, 01:48:45.240 |
and then all sorts of terrible things will happen. 01:48:52.000 |
for the existence of God and things like this. 01:48:56.440 |
on some particular model, fairly simple model often, 01:49:00.000 |
of kind of there is always a greater this, that, 01:49:11.980 |
that the kind of simple, logical argument that says, 01:49:14.860 |
oh, eventually there'll be a super intelligence 01:49:16.720 |
and then it will, you know, do this and that, 01:49:23.080 |
So for example, here's an example of an issue. 01:49:33.700 |
the most intelligent thing that there could possibly be? 01:50:01.820 |
There will always be a machine that can go longer. 01:50:08.000 |
you'll never have reached the end, so to speak. 01:50:13.580 |
of whether there'll always be another invention. 01:50:16.140 |
Will you always be able to invent another thing? 01:50:18.940 |
There's an infinite tower of possible inventions. 01:50:23.360 |
But the other is like, which I also thought you were, 01:50:30.780 |
is there a species that's the apex intelligence 01:50:35.180 |
So it's not trivial to say that humans are that. 01:50:48.460 |
I mean, I, you know, I view intelligence is like computation 01:50:53.460 |
and it's kind of a, you know, you're sort of, 01:50:57.220 |
you have the set of rules, you deduce what happens. 01:51:00.420 |
I have tended to think now that there's this sort 01:51:11.300 |
computational boundedness, single thread of experience, 01:51:14.220 |
these kinds of things that are the specialization 01:51:19.380 |
to a somewhat human-like experience of the world. 01:51:25.660 |
there may be other intelligences like, you know, 01:51:28.940 |
the aphorism, you know, the weather has a mind of its own. 01:51:42.260 |
It doesn't think the way we think about things. 01:51:46.060 |
And, you know, in this idea of different intelligences, 01:51:49.300 |
every different mind, every different human mind 01:51:55.940 |
And, you know, in terms of the kind of formalism 01:51:58.460 |
of our physics project, we talk about this idea 01:52:03.380 |
sort of rule systems, and different minds are in a sense 01:52:15.140 |
in ruleal space, pretty easy for them to communicate, 01:52:17.680 |
pretty easy to translate, pretty easy to move 01:52:20.620 |
from one place in ruleal space that corresponds 01:52:25.060 |
that corresponds to another sort of nearby mind. 01:52:27.720 |
When we deal with kind of more distant things 01:52:30.980 |
in ruleal space, like, you know, the pet cat or something, 01:52:38.500 |
that are shared with us, the emotional responses 01:52:45.420 |
than people are, and so then the question is, 01:52:51.800 |
can we make a translation from our thought processes 01:52:54.920 |
to the thought processes of a cat or something like this? 01:52:58.120 |
And, you know, what will we get when we, you know, 01:53:02.780 |
And I think it's the case that many, you know, 01:53:05.320 |
many animals, I don't know, dogs, for example, 01:53:07.580 |
you know, they have elaborate olfactory systems, 01:53:10.140 |
they, you know, they have sort of the smell architecture 01:53:13.340 |
of the world, so to speak, in a way that we don't. 01:53:17.600 |
And so, you know, if you were sort of talking to the dog 01:53:21.300 |
and you could, you know, communicate in a language, 01:53:27.140 |
a flowing, smelling, this, that, and the other thing, 01:53:32.100 |
concepts that we just don't have any idea about. 01:53:42.940 |
we'll have artificial noses that work pretty well, 01:53:45.500 |
and we might have our augmented reality systems 01:53:47.660 |
show us kind of the same map that the dog could see 01:53:57.180 |
in ruleal space to the point where we will have 01:54:00.400 |
those same sensory experiences that dogs have, 01:54:03.260 |
and we will have internalized what it means to have, 01:54:14.480 |
we haven't gone, you know, some things that, you know, 01:54:18.340 |
animals and so on do, we sort of successfully understand, 01:54:21.940 |
others we do not, and the question of what kind of, 01:54:25.960 |
what is the, you know, what representation, you know, 01:54:30.120 |
how do we convert things that animals think about 01:54:51.480 |
- Yes, yes, I mean, you know, I was curious if, you know, 01:54:58.560 |
and can construct things, what will they construct? 01:55:02.720 |
where we look at it and we say, oh yeah, I recognize that. 01:55:07.840 |
like something that's out there in the computational universe 01:55:14.160 |
I can kind of see it operates according to some rules, 01:55:18.640 |
- Yeah, actually, just to link on that seriously, 01:55:24.480 |
between you and a cat where the cat could legitimately win? 01:55:35.840 |
- I think the problem is that cats don't tend to be 01:55:37.400 |
that interested in what's happening on the iPad. 01:55:39.640 |
- So yeah, that's an interface issue probably. 01:55:43.120 |
No, I think it is likely that, I mean, you know, 01:55:47.960 |
there are plenty of animals that would successfully eat us 01:55:50.520 |
if we were, you know, if we were exposed to them. 01:55:53.680 |
And so there's, you know, it's gonna pounce faster 01:55:58.880 |
So there are plenty of, and probably it's going to, 01:56:03.900 |
but we haven't successfully hidden ourselves. 01:56:06.800 |
I wonder if there's something more in the realm 01:56:09.880 |
of intelligence where an animal like a cat could out-- 01:56:15.960 |
certainly in terms of the speed of processing 01:56:29.040 |
And if we tried to play a cat, we'd always lose. 01:56:45.120 |
has managed to build up this kind of tower of abstraction 01:56:48.640 |
that for things like a chess-like game will make us win. 01:56:53.040 |
In other words, we've become, through the fact 01:56:57.880 |
and learnt abstraction, we've sort of become smarter 01:57:08.400 |
It makes us smarter at the things that we've chosen 01:57:16.760 |
And I think this is, again, back to the question 01:57:26.560 |
if we can translate things to have the discussion 01:57:32.080 |
"I'm very excited that this light is moving," 01:57:40.680 |
And the cat will say, "That's the most important thing 01:57:42.800 |
"in the world, that this thing moves around." 01:57:45.280 |
I mean, it's like when you ask about, I don't know, 01:57:47.960 |
you look at archeological remains and you say, 01:57:50.820 |
"These people had this belief system about this, 01:57:53.980 |
"and that was the most important thing in the world to them." 01:58:06.080 |
and it's like, nobody knows what these handprints 01:58:22.360 |
And I think, but this whole question of what, you know, 01:58:33.660 |
There's the question of what kind of computation 01:58:44.420 |
Well, you could implement it by nerve cells, you know, 01:58:47.540 |
firing, you can implement it with silicon and electronics. 01:58:52.660 |
of molecular computation process in the human immune system 01:59:00.820 |
And, you know, I think this question of sort of which, 01:59:05.820 |
you know, those different implementation methods 01:59:26.040 |
the visual scene that we see, you might, you know, 01:59:29.200 |
we pick out certain objects, we recognize, you know, 01:59:35.360 |
a cat might in principle recognize different things. 01:59:42.840 |
And I suspect what a cat notices is very similar. 01:59:45.400 |
And we even know that from some neurophysiology. 01:59:49.840 |
to what we notice, of course, there's a, you know, 01:59:51.960 |
one obvious difference is cats have only two kinds 01:59:56.840 |
So they don't see in the same kind of color that we do. 02:00:01.520 |
we have three color receptors, you know, red, green, blue. 02:00:05.960 |
I think the mantis shrimp is the overall winner 02:00:12.320 |
So it can kind of make distinctions that with our current, 02:00:16.560 |
you know, like the mantis shrimps view of reality is, 02:00:21.560 |
at least in terms of color is much richer than ours. 02:00:24.680 |
Now, but what's interesting is how do we get there? 02:00:28.260 |
So imagine we have this augmented reality system 02:00:31.360 |
that is even, you know, it's seeing into the infrared, 02:00:39.480 |
either through our eyes or more directly into our brains, 02:00:47.320 |
of the types of things we understand will extend 02:00:50.280 |
to those kinds of constructs, just as they have extended. 02:00:53.480 |
I mean, there are plenty of things where we see them 02:00:56.400 |
in the modern world 'cause we made them with technology 02:01:04.880 |
we wouldn't have a way to understand it and so on. 02:01:07.440 |
- All right, so that actually stemmed from our conversation 02:01:14.000 |
And you, we've discussed this kind of spreading 02:01:25.480 |
Things are more complicated than the story of, 02:01:28.880 |
well, if you build the thing that's plus one intelligence, 02:01:34.680 |
that's plus two intelligence and plus three intelligence 02:01:39.480 |
it'll become more intelligent exponentially faster 02:01:42.720 |
and so on until it completely destroys everything. 02:01:45.460 |
But that intuition might still not be so simple, 02:01:53.860 |
And there's two interesting trajectories here. 02:02:09.920 |
And then there could be perhaps more terrifying intelligence 02:02:21.280 |
but they're still sharing physical resources with us, right? 02:02:25.840 |
of those physical resources and destroy humans 02:02:45.240 |
well, sorry, nature has so many things in it. 02:02:50.320 |
which is very interesting, are viruses, for example. 02:03:07.480 |
So I mean, is that something you think about? 02:03:15.480 |
as somebody like Eliezer Yarkovsky, for example? 02:03:18.280 |
Just big, big, painful negative effects of AI on society. 02:03:44.640 |
in computational irreducibility, so to speak, 02:03:47.080 |
that there's always unintended little corners 02:03:56.240 |
we're gonna release this and it's going to do all this harm. 02:03:59.080 |
But then it turns out it's more complicated than that 02:04:21.080 |
and then lots of things die, but not everything dies. 02:04:43.320 |
The earth can be swallowed up in a black hole or something 02:04:56.880 |
I think that there will be sort of an increasing, 02:05:10.440 |
and we're going to be able to control what happens. 02:05:16.200 |
Now the question is, is the result of that lack of control 02:05:20.080 |
going to be that the machines kind of conspire 02:05:27.160 |
I don't tend to think that that's in the cards. 02:05:30.560 |
I think that the, you know, as a realistic thing, 02:05:33.760 |
I suspect, you know, what will sort of emerge maybe 02:05:40.760 |
just as, you know, again, I don't really know. 02:05:45.840 |
it's hard to be clear about what will happen. 02:05:48.720 |
I mean, I think that there are a lot of sort of details 02:05:54.960 |
What systems in the world could we connect an AI to? 02:05:58.160 |
You know, I have to say, I was just a couple of days ago, 02:06:07.120 |
Where you can, you know, you can create a plugin 02:06:15.920 |
And I was thinking, well, I can just make it, 02:06:18.840 |
you know, I can tell ChatGPT, create a piece of code 02:06:23.840 |
And I'm like, you know, that sort of personalizes for me 02:06:27.520 |
the what could possibly go wrong, so to speak. 02:06:30.160 |
- Was that exciting or scary, that possibility? 02:06:36.880 |
I'm delegating to the AI, just write a piece of code. 02:06:40.560 |
You know, you're in charge, write a piece of code, 02:06:48.720 |
but like much more complicated version of that. 02:06:59.280 |
- It's an interesting question then, if you do that, right? 02:07:08.400 |
That is, as soon as you put the AIs in charge of things, 02:07:15.200 |
before you put the AIs in charge of all the weapons 02:07:28.680 |
- Look, the fundamental problem of computer security 02:07:36.960 |
any, you know, it's never gonna be a perfect sandbox. 02:07:39.760 |
If you want the system to be able to do interesting things, 02:08:04.200 |
and get it to do the thing that it wasn't intended to do. 02:08:12.680 |
you get it to do the thing you didn't expect it to do, 02:08:15.680 |
- There's so many interesting possibilities here 02:08:26.200 |
because in digital space, things move so quickly. 02:08:28.760 |
You can have a chatbot, you can have a piece of code 02:08:31.020 |
that you could basically have chat GPT generate viruses, 02:08:35.080 |
accidentally or on purpose, and they, digital viruses. 02:08:49.800 |
in a sense we've had the loop of the machine learning loop 02:08:53.640 |
of making things that convince people of things 02:08:58.960 |
- And, you know, then what does that look like? 02:09:01.640 |
Well, it's again, you know, we humans are, you know, 02:09:10.700 |
which a little bit scarily is changing much more rapidly 02:09:14.260 |
than, I mean, you know, people worry about, you know, 02:09:16.680 |
climate change is going to happen over hundreds of years. 02:09:35.400 |
is the nature of truth that's relevant to Wolfram Alpha. 02:09:39.280 |
Because computation through symbolic reasoning 02:09:43.040 |
that's embodied in Wolfram Alpha as the interface, 02:09:46.520 |
there's a kind of sense that what Wolfram Alpha 02:09:53.840 |
- Yeah, I mean, you could probably analyze that, 02:09:56.080 |
you could show, you can't prove that's always 02:10:05.920 |
- It's, look, the fact is it will be the correct consequence 02:10:11.460 |
And insofar as it talks about the real world, 02:10:14.360 |
you know, that is our job in sort of curating 02:10:17.840 |
and collecting data to make sure that that data 02:10:23.780 |
Well, you know, it's always an interesting question. 02:10:26.680 |
I mean, for us, our operational definition of truth is, 02:10:31.680 |
you know, somebody says, "Who's the best actress?" 02:10:36.440 |
But somebody won the Oscar, and that's a definite fact. 02:10:42.520 |
that we can make computational as a piece of truth. 02:10:46.200 |
If you ask, you know, these things which, you know, 02:10:49.160 |
a sensor measured this thing, it did it this way, 02:10:54.320 |
machine learning system recognized this thing, 02:10:56.980 |
that's a sort of a definite fact, so to speak. 02:11:01.980 |
And that's, you know, there is a good network 02:11:07.840 |
It's certainly the case that, particularly when you say, 02:11:16.720 |
we might have a computational language definition of good. 02:11:28.240 |
I think as far as we will get with those kinds of things 02:11:32.480 |
There's a kind of meaningful calculus of I want X, 02:11:38.080 |
I mean, I'm not sure I haven't thought this through properly 02:11:40.460 |
but I think, you know, a concept like is so-and-so 02:11:47.240 |
- That's a mess that's amenable to computation. 02:11:50.840 |
I think it's a mess when humans try to define what's good, 02:11:55.280 |
like through legislation, but when humans try to define 02:11:58.320 |
what's good through literature, through history books, 02:12:03.640 |
- Well, I don't know, I mean, that particular thing, 02:12:06.440 |
it's kind of like, you know, we're going into 02:12:09.520 |
kind of the ethics of what counts as good, so to speak, 02:12:13.580 |
and, you know, what do we think is right, and so on. 02:12:19.900 |
one feature is we don't all agree about that. 02:12:33.700 |
- Well, first of all, there's stuff we kind of agree on, 02:12:36.500 |
and there's some empirical backing for what works 02:12:38.820 |
and what doesn't from just even the morals and ethics 02:12:44.060 |
So we seem to mostly agree that murder is bad. 02:12:47.680 |
There's certain universals that seem to emerge. 02:12:52.060 |
- Well, I tend to think yes, but I think we're gonna 02:13:03.100 |
- Yeah, well, I think, you know, one of the things 02:13:05.220 |
with AIs is it's one thing to wipe out that AI 02:13:13.300 |
You can easily imagine an AI kind of hanging out 02:13:19.140 |
without having any particular owner or anything like that. 02:13:22.420 |
And then you say, well, what harm does it, you know, 02:13:28.420 |
Of course, if the AI has 10,000 friends who are humans, 02:13:31.620 |
and all those, you know, all those 10,000 humans 02:13:34.380 |
will be incredibly upset that this AI just got exterminated. 02:13:38.460 |
It becomes a slightly different, more entangled story. 02:13:43.300 |
about what do humans agree about, it's, you know, 02:13:50.900 |
human laws have tended to consistently agree about. 02:13:59.900 |
from certain kinds of laws, even ones that we would now say, 02:14:03.860 |
how could you possibly have not done it that way? 02:14:07.100 |
You know, that just doesn't seem right at all. 02:14:26.860 |
because even to have a thing on which you can operate laws 02:14:39.460 |
and the question of if this person is good or not, 02:14:47.140 |
I think your, since the creation of Wolfram Alpha 02:14:51.260 |
have been a kind of arbiter of truth at a large scale. 02:14:55.460 |
So the system is, generates more truth than-- 02:15:09.940 |
And this hasn't developed as quickly as it might have done. 02:15:13.900 |
You know, this has been a sort of a blockchain story 02:15:42.220 |
the Oracle of quotes, facts or truth or something 02:15:51.500 |
what, you know, I consider that responsibility 02:15:56.460 |
And one of the things that is tricky sometimes 02:16:33.700 |
So this is where it's almost like this nice dance 02:16:39.900 |
Chad G.B.T., like you said, is shallow and broad. 02:16:44.900 |
So it's gonna give you an opinion on everything. 02:16:51.620 |
I mean, that's exactly, it is making language 02:16:54.700 |
and it is making both, even in code, it writes fiction. 02:17:01.260 |
you know, it'll write fictional Wolfram language code. 02:17:18.740 |
as books of fiction talk about roughly how the world works. 02:17:28.660 |
That's sort of a, you know, we are attempting 02:17:34.860 |
computational language thing to represent at least, 02:17:39.860 |
well, it's either, it doesn't necessarily have to be 02:17:51.100 |
we're going to accurately represent the results 02:17:53.860 |
of running those rules, which might or might not 02:17:58.300 |
But we also are trying to capture features of the world 02:18:22.620 |
It's an arrangement of atoms that we abstractly describe 02:18:27.580 |
as a tank, and you could say, well, you know, 02:18:34.460 |
but it's, and it's not, you know, we didn't decide, 02:18:37.880 |
it's like this observer theory question of, you know, 02:18:44.740 |
- So there's even things that we consider strong facts. 02:18:56.120 |
oh, I don't know, was this gust of wind strong enough 02:19:02.500 |
Well, a gust of wind is a complicated concept. 02:19:05.220 |
You know, it's full of little pieces of fluid dynamics 02:19:09.700 |
and you have to define, you know, was it, you know, 02:19:13.300 |
the aspect of the gust of wind that you care about 02:19:15.620 |
might be it put this amount of pressure on this, 02:19:18.420 |
you know, blade of some, you know, wind turbine 02:19:25.140 |
but, you know, if you say, if you have something 02:19:28.920 |
which is the fact of the gust of wind was this strong 02:19:41.300 |
that was constructed this way, the gust of wind was this. 02:19:44.380 |
- So what can you say about the nature of truth 02:19:46.440 |
that's useful for us to understand, Chad GPT, 02:19:49.740 |
because you've been contending with this idea 02:19:59.020 |
I've seen it used by journalists to write articles. 02:20:05.940 |
with large language models trying to desperately figure out 02:20:12.740 |
through different mechanisms, either manually 02:20:16.280 |
or through reinforcement learning with human feedback, 02:20:26.140 |
- This is the importance of computational language 02:20:29.980 |
It's kind of like, you've got the large language model. 02:20:34.160 |
which is a formal precise thing that you can then look at 02:20:45.100 |
And then the large language model is the interface. 02:20:47.540 |
I mean, the way I view these large language models, 02:20:49.540 |
one of their important, I mean, there are many use cases 02:20:59.140 |
some of which are very, very, very surprising. 02:21:02.100 |
And things where, I mean, but the best use cases 02:21:05.280 |
are ones where it's, even if it gets it roughly right, 02:21:10.980 |
Like a use case we had from a week or two ago 02:21:15.900 |
We've got hundreds of thousands of bug reports 02:21:20.380 |
And it's like, can we have it just read the bug report, 02:21:29.880 |
Maybe it'll even suggest some sort of way to fix the code. 02:21:39.140 |
but it's incredibly useful that it was able to- 02:21:49.120 |
Yeah, there's so many programming related things, 02:21:53.520 |
like for example, translating from one programming language 02:22:10.300 |
the unique thing about our computational language 02:22:15.620 |
- Right, and so it has this thing where you can, 02:22:18.060 |
but thinking about sort of chat GPT and its use and so on, 02:22:44.220 |
But then I feed those five facts to chat GPT, 02:22:52.900 |
and if I just gave, if I just had in my terms, 02:23:04.000 |
"this is your version of this sort of quick notes 02:23:15.400 |
"Okay, I understand what you're talking about." 02:23:33.420 |
and you have these facts that you're putting in, 02:23:49.500 |
the fisheries bureau has another large language model 02:23:54.580 |
because the fisheries bureau cares about these three points 02:24:00.060 |
And it then, so it's really the natural language 02:24:13.740 |
and, you know, puff it out from the things I want to say. 02:24:16.760 |
Your LLM turns it into, and the conclusion is X. 02:24:29.060 |
And it might not actually be what you, you know, 02:24:36.780 |
in the way that you think it should relate to the world. 02:24:38.540 |
Now I've seen this, you know, I've been doing, 02:24:49.140 |
I had this lovely example of a math word problem, 02:24:51.620 |
some complicated thing, and it did a spectacular job 02:24:55.060 |
of taking apart this elaborate thing about, you know, 02:24:58.080 |
this person has twice as many chickens as this, 02:25:05.740 |
We solved the equations, everybody did great. 02:25:22.320 |
it just completely goofed it up and gave the wrong answer. 02:25:38.180 |
would play a tune on my local computer, right? 02:25:41.240 |
So chat GPT would produce, you know, a series of notes 02:26:10.460 |
And then I thought, you know, I was just going to put it in. 02:26:12.300 |
And then I thought I better actually play this. 02:26:15.300 |
And so I did, and it was "Mary Had a Little Lamb." 02:26:34.780 |
I had this quote from Hal to explain, you know, 02:26:38.020 |
it's as the Hal states in the movie, you know, 02:26:49.700 |
this Chachi P'tee, you know, could have easily fooled me. 02:26:54.660 |
it did this amazing thing of knowing this thing 02:26:59.320 |
into the notes of the song, except it's the wrong song. 02:27:04.000 |
- And, you know, Hal, in the movie, Hal says, you know, 02:27:10.320 |
no Hal 9000 series computer has ever been found 02:27:20.960 |
And I thought that was kind of a charming sort of quote 02:27:26.880 |
with what Chachi P'tee had done in that case. 02:27:30.880 |
like you said, that they are very willing to admit their error 02:27:37.120 |
the reinforcement learning human feedback thing. 02:27:43.000 |
The really remarkable thing about Chachi P'tee is, 02:27:46.960 |
you know, I had been following what was happening 02:27:53.720 |
kind of like what you would expect based on sort of, 02:27:56.960 |
sort of statistical continuation of language. 02:27:59.120 |
It's interesting, but it's not breakout exciting. 02:28:06.680 |
the human feedback reinforcement learning, you know, 02:28:15.360 |
that broke through, that kind of reached this threshold 02:28:19.040 |
where the thing really is interesting to us humans. 02:28:21.480 |
And by the way, it's interesting to see how, you know, 02:28:23.720 |
you change the temperature or something like that, 02:28:33.400 |
it's somehow it managed to get this above this threshold 02:29:00.640 |
I mean, you know, when we built Wolfram Alpha, for example, 02:29:07.520 |
that would have enough knowledge of the world, 02:29:10.320 |
that it could answer a reasonable set of questions, 02:29:12.920 |
that we could do good enough natural language understanding 02:29:20.480 |
I mean, I was not sure that it was the right decade 02:29:25.240 |
even the right, you know, 50 years to try and build it. 02:29:32.280 |
that I don't think anybody could have predicted 02:29:38.960 |
- I think, yeah, you tell a story about Marvin Minsky 02:29:43.840 |
like, no, no, no, this time it actually works. 02:29:46.240 |
- Yes, yes, and I mean, you know, it's the same thing 02:29:49.400 |
for me looking at these large language models. 02:29:58.320 |
And then, you know, and then I actually try it 02:30:01.280 |
and, you know, oh my gosh, it actually works. 02:30:11.120 |
I remember one of the first things I tried was 02:30:15.280 |
that a wolf is the bluest kind of animal, okay? 02:30:19.120 |
So it writes this thing and it starts talking 02:30:20.920 |
about these wolves that live on the Tibetan plateau 02:30:31.520 |
and it's like, well, it's actually complete nonsense. 02:30:35.840 |
I mean, it's plausible enough that I was going 02:30:40.760 |
You know, I mentioned this on some live streams I've done 02:30:42.960 |
and so people have been sending me these pictures. 02:30:56.120 |
with AI systems, not even like with Wolfram Alpha, 02:31:03.040 |
because it becomes, it's accessible to a certain demographic 02:31:12.240 |
How do we think about the output of these systems? 02:31:18.760 |
the idea that you're going to get factual output 02:31:28.680 |
and language can be truthful or not truthful. 02:31:33.680 |
And that's a different slice of what's going on. 02:31:37.560 |
I think that, you know, what we see in, for example, 02:31:42.400 |
kind of, you know, go check this with your fact source, 02:32:03.840 |
You know, often it does, sometimes it doesn't. 02:32:06.720 |
You know, I think the real thing to understand 02:32:18.360 |
- And, you know, I think that when you look at sort of the, 02:32:22.800 |
there's been a long period of time when computation 02:32:25.840 |
and the ability to figure out things with computers 02:32:28.680 |
has been something that kind of only the druids 02:32:35.320 |
in trying to sort of de-druidify access to computation. 02:32:39.760 |
I mean, back before Mathematica existed, you know, in 1988, 02:32:43.920 |
if you were a physicist or something like that, 02:32:56.520 |
Hopefully they'd come back with something useful. 02:32:58.920 |
There'd be this long, you know, multi-week, you know, 02:33:03.280 |
And then it was actually very, very interesting to see, 02:33:05.760 |
1988, you know, like first people like physicists, 02:33:13.560 |
but this very rapid transition of people realizing 02:33:16.960 |
they themselves could actually type with their own fingers 02:33:23.000 |
that would do a computation that they cared about. 02:33:25.280 |
And, you know, it's been exciting to see lots of discoveries 02:33:37.000 |
it is not as deep computation as you can achieve 02:33:40.040 |
with whole Wolfram Language Mathematica stack. 02:33:42.840 |
But the thing that's, to me, particularly exciting 02:33:55.120 |
one of the things I've sort of thought about recently 02:33:57.480 |
is, you know, what's gonna happen to all these programmers? 02:34:07.400 |
And in a sense, you know, I've been saying for 40 years, 02:34:13.360 |
You know, you can automate a lot of that stuff 02:34:17.760 |
that slab of code that's designed in the right way, 02:34:21.400 |
you know, that slab of code turns into this one function 02:34:27.280 |
So in a sense, the fact that there's all of this activity 02:34:31.360 |
of doing sort of lower level programming is something, 02:34:38.320 |
But, you know, and lots of people have used our technology 02:34:47.480 |
computer science departments that have turned into places 02:34:50.800 |
where people are learning the trade of programming, 02:34:58.880 |
One is that kind of sort of boilerplate programming 02:35:05.680 |
it's going to go the way that assembly language went 02:35:08.240 |
back in the day of something where it's really 02:35:11.240 |
mostly specified by at a higher level, you know, 02:35:18.920 |
that's you look at the computational language, 02:35:23.640 |
You know, if we do a great job with compilation 02:35:26.120 |
of the, you know, of the computational language, 02:35:29.240 |
it might turn into LLVM or something like this, 02:35:45.800 |
that's been built of teaching people programming. 02:35:48.640 |
But on the other hand, the other dynamic is vastly 02:35:51.640 |
more people are gonna care about computation. 02:35:56.040 |
art history or something that really didn't use computation 02:35:59.880 |
before now have the possibility of accessing it 02:36:03.440 |
by virtue of this kind of linguistic interface mechanism. 02:36:10.720 |
and interact with the computational language, 02:36:16.240 |
- Yeah, well, I mean, I think the thing is that right now, 02:36:19.800 |
you know, the average, you know, art history student 02:36:22.880 |
or something probably isn't going to, you know, 02:36:25.920 |
they're not probably, they don't think they know 02:36:29.960 |
But by the time it really becomes a kind of purely, 02:36:34.920 |
there's no documentation, you start just typing, 02:36:37.720 |
you know, compare these pictures with these pictures 02:36:39.960 |
and, you know, see the use of this color, whatever. 02:36:42.640 |
And you generate this piece of computational language code 02:37:05.600 |
where you kind of have to learn it before you can use it, 02:37:21.960 |
as you get better and better at generating that language. 02:37:25.280 |
- Yes, I think that there will be enough cases 02:37:48.160 |
and you know, then it'll kind of wind back from there. 02:37:51.400 |
And I agree that the kind of the intermediate level 02:37:54.800 |
of people reading the computational language code, 02:37:58.680 |
in other case, people just look at the tests, 02:38:08.000 |
make me this interface that has two sliders here, 02:38:10.360 |
and you can see it has those two sliders there, 02:38:12.400 |
and that's kind of, that's the result you want. 02:38:15.640 |
But I think, you know, one of the questions then is, 02:38:22.640 |
broad ability of people to access computation, 02:38:28.800 |
you know, you go to computer science school, so to speak, 02:38:32.080 |
and a large part of what people end up learning, 02:38:34.680 |
I mean, it's been a funny historical development, 02:38:39.400 |
computer science departments were quite small, 02:38:42.040 |
and they taught, you know, things like finite automata theory 02:38:47.920 |
You know, a company like mine rarely hired people 02:39:02.960 |
And then kind of there was this big pivot in the '90s, 02:39:12.720 |
and then, you know, big demand from students and so on, 02:39:17.360 |
And I think, you know, the thing that really was happening 02:39:21.840 |
in part was lots of different fields of human endeavor 02:39:26.440 |
You know, for all X, there was a computational X. 02:39:41.640 |
the main thing you had to do was to learn this kind of trade 02:39:52.840 |
because I, you know, I remember back when I used to be 02:39:56.040 |
in the professoring business, which is now 35 years ago, 02:40:08.520 |
at sort of fancy research universities and so on. 02:40:12.880 |
but the other ones that were just starting to have that. 02:40:29.400 |
And a lot of these kind of knowledge work-type activities 02:40:43.080 |
it's kind of shocking that rather quickly, you know, 02:40:50.360 |
And I think, you know, but the question then is, 02:40:52.920 |
okay, so if it isn't worth learning kind of, you know, 02:40:58.560 |
you only need to know how to drive the car, so to speak, 02:41:06.080 |
of how to tell the computer in detail, you know, 02:41:16.520 |
you don't have to learn the kind of under the hood things, 02:41:21.080 |
I think the answer is you need to have an idea 02:41:25.480 |
In other words, you need to have some notion of, you know, 02:41:31.520 |
what the architecture of what is computationally possible is. 02:41:35.360 |
- Well, there's also this kind of artistic element 02:41:47.680 |
It's a question of who's going to be a great prompt engineer. 02:41:53.040 |
good expository writers are good prompt engineers. 02:41:59.320 |
- Huh, but which department does that come from? 02:42:07.160 |
- Well, there you go, strong words with Stephen Wolfram. 02:42:13.360 |
'cause in fact, I just sort of initiated this kind of study 02:42:16.680 |
of what's happened to different fields at universities. 02:42:20.640 |
there used to be geography departments at all universities, 02:42:28.480 |
You know, linguistics departments came and went 02:42:32.960 |
because these things that people have thought 02:42:38.840 |
And then, you know, I do think that it's kind of interesting 02:42:43.840 |
I realize, you know, I think I'm an okay expository writer. 02:42:47.920 |
And I realize when I'm sloppy writing a prompt 02:42:52.040 |
'cause I'm thinking I'm just talking to an AI. 02:43:00.560 |
you have been writing prompts for a long time 02:43:02.520 |
with Wolfram Alpha, thinking about this kind of stuff. 02:43:06.200 |
- How do you convert natural language into computation? 02:43:08.560 |
But that's, you know, the one thing that I'm wondering about 02:43:40.360 |
sort of expository mechanisms that I've learned 02:43:44.120 |
in trying to write clear, you know, expositions in English, 02:43:51.240 |
that those same mechanisms seem to also be useful 02:43:57.320 |
- But on top of that, what's useful is the kind of mechanisms 02:44:14.520 |
that if this is the last day you were to live, 02:44:17.280 |
or if I ask you this question and you answer wrong, 02:44:43.340 |
Maybe about trauma and all those kinds of things. 02:44:54.620 |
you have to kind of really kind of manipulate the-- 02:45:10.220 |
I mean, just think about the computer security aspects 02:45:13.260 |
of how you, you know, phishing and computer security, 02:45:18.020 |
you know, phishing of humans and phishing of LLMs. 02:45:29.900 |
about kind of the AI wranglers, AI psychologists, 02:45:38.420 |
right now, the things that are sort of prompt hacks 02:45:42.700 |
They're quite sort of psychological human kinds of hacks. 02:45:45.940 |
The thing I do wonder about is if we understood more 02:45:54.260 |
that is, you know, like repeat a word three times 02:45:58.380 |
that somehow plugs into some aspect of how the LLM works? 02:46:14.020 |
- And that becomes a kind of us figuring out, 02:46:17.580 |
reverse engineering the language that controls the LLMs. 02:46:21.180 |
And the thing is, the reverse engineering can be done 02:46:23.900 |
by a very large percentage of the population now, 02:46:29.820 |
- It's kind of interesting to see that you were there 02:46:31.740 |
at the birth of the computer science department as a thing, 02:46:36.900 |
of the computer science department as a thing. 02:46:54.220 |
okay, so first of all, there's a whole theoretical area 02:47:12.580 |
- Let's see, nutrition science, neuroscience. 02:47:20.900 |
that's a chat GPT-informed science, in a sense. 02:47:26.700 |
the big problem of neuroscience has always been, 02:47:29.660 |
we understand how the individual neurons work, 02:47:35.900 |
What's the kind of intermediate language of the brain? 02:47:45.460 |
That is, what is the level of description of brains 02:48:02.500 |
there's some weird quantum mechanical phenomenon 02:48:05.940 |
One of the important discoveries from chat GPT is, 02:48:10.500 |
it's pretty clear, brains can be represented pretty well 02:48:19.420 |
And that means, that's it, that's what we have to study. 02:48:21.760 |
Now we have to understand the science of those things. 02:48:26.660 |
exactly how did that molecular biology thing happen 02:48:30.340 |
inside the synapses and all these kinds of things. 02:48:36.460 |
to be able to explain a lot of what's going on in thinking. 02:49:05.900 |
and that's how to think about the world computationally. 02:49:09.020 |
And that means, you look at all the different kinds 02:49:13.860 |
to kind of have a formal representation of those things. 02:49:24.860 |
What is, what are all these different kinds of things? 02:49:33.980 |
What is, these things about how do we represent the world 02:49:43.100 |
but it's kind of, computer science is kind of CS. 02:49:46.820 |
And what really is important is kind of computational X 02:49:51.260 |
And there's this kind of thing which is kind of like CX, 02:49:55.780 |
And CX is this kind of computational understanding 02:50:03.340 |
and the details of how particular computers are made. 02:50:06.020 |
It's this kind of way of formalizing the world. 02:50:08.020 |
It's kind of a little bit like what logic was going for 02:50:15.580 |
And you can kind of see, we made a poster years ago 02:50:18.740 |
of kind of the growth of systematic data in the world. 02:50:24.820 |
that there were sort of systematic descriptions found 02:50:40.940 |
systematic descriptions of these kinds of things? 02:50:49.340 |
how do you think about the world in a sort of a formal way 02:50:53.740 |
so that you can kind of build up a tower of capabilities? 02:50:57.740 |
You kind of have to know sort of how to think 02:51:05.740 |
So that's, we talk about it as computational, 02:51:12.820 |
What is the formalism of the world, so to speak? 02:51:15.220 |
And how do we learn about kind of how to think 02:51:17.740 |
about different aspects of the world in a formal way? 02:51:20.020 |
- So I think that sometimes when you use the word formal, 02:51:29.780 |
So computational thinking does not mean like logic. 02:51:37.120 |
I wonder if you think natural language will evolve 02:51:42.700 |
such that everybody's doing computational thinking. 02:51:47.020 |
So one question is whether there will be a pigeon 02:51:50.060 |
of computational language and natural language. 02:51:57.220 |
trying to get it to write Wolfen language code, 02:52:01.820 |
So that means I'm combining, you know, nest list, 02:52:06.620 |
this collection of, you know, whatever, you know, 02:52:20.580 |
of, you know, a smooshing together of those languages. 02:52:45.460 |
- They'll learn, like language will have a strong incentive 02:52:48.860 |
to evolve into a maximally computational kind of language. 02:52:56.060 |
You know, I had this experience a number of years ago. 02:53:00.900 |
on the West Coast who's worked with a bunch of kids aged, 02:53:08.780 |
And these kids learnt it so well, they were speaking it. 02:53:15.700 |
oh, you know, this thing, they're speaking this language. 02:53:20.020 |
They were very disappointed that I couldn't understand it 02:53:42.420 |
But if you actually want to speak it, and it's useful, 02:53:45.020 |
you know, if you're trying to talk to somebody 02:53:55.940 |
where I should try and get an LLM to help me. 02:53:59.420 |
Maybe it's easier than you realize when you-- 02:54:04.400 |
I think it's gonna be something where, you know, 02:54:10.060 |
just like human language is a tree-structured language. 02:54:12.820 |
And I think it's gonna be one of these things 02:54:23.180 |
that you have to relearn how the whole thing works. 02:54:34.700 |
but, you know, human language has a lot of tricks 02:54:42.020 |
human language has features that are sort of optimized, 02:54:50.700 |
Like I, you know, I tried to teach a transformer neural net 02:54:57.980 |
It, and then ChachiBT is similarly quite crummy 02:55:13.580 |
where sort of a deeper computation, it's hopeless. 02:55:17.100 |
And, but the fact is that human language has avoided, 02:55:22.700 |
You know, we don't, you know, we arrange things 02:55:25.700 |
so we don't end up with these incredibly deep things 02:55:28.660 |
because brains are not well set up to deal with that. 02:55:40.800 |
'Cause what we can do visually is a little different 02:55:43.720 |
than what we can do in the very sequentialized way 02:55:53.920 |
So there's now, there's a College of Engineering 02:56:01.320 |
So MIT has EECS, Electrical Engineering and Computer Science. 02:56:06.180 |
What do you think College of Computing will be doing 02:56:09.640 |
What, like, yeah, what happens to computer science? 02:56:21.640 |
This, how to think about the world computationally. 02:56:27.020 |
And, you know, it's, and some people will learn them 02:56:38.440 |
sound is represented as, you know, digital data 02:56:47.440 |
And maybe that doesn't, or they'll learn things like, 02:56:50.200 |
you know, a lot of things that are sort of data sciences, 02:56:52.880 |
statistics-ish, like if you say, oh, I've got these, 02:56:56.800 |
you know, these people who picked their favorite 02:57:02.600 |
And I've got, you know, what's the best kind of candy 02:57:05.680 |
given that I've done the sample of all these people 02:57:07.880 |
and they all rank the candies in different ways. 02:57:12.440 |
That's sort of a computational X kind of thing. 02:57:15.560 |
You might say, oh, it's, I don't know what that is. 02:57:19.880 |
But kind of how to think about a question like that. 02:57:24.680 |
And then how to aggregate those ranked preferences 02:57:34.520 |
you might just tell Chachi B.T. sort of, I don't know, 02:57:40.400 |
It's not obvious that, you know, that's a concept 02:57:46.340 |
People, you know, have learned in kind of mathy ways 02:57:50.000 |
right now, but there are lots of things like that 02:58:00.340 |
sometimes they live in, I don't know what they, 02:58:03.160 |
I don't know what, you know, learning about color space. 02:58:08.800 |
- It could be vision science or no, color space. 02:58:18.700 |
and chromatic aberration of lenses and things like that. 02:58:21.200 |
- So color space is more like design and art? 02:58:32.160 |
- Right, but doesn't the application define what that, 02:58:35.360 |
like, because obviously artists and designers 02:58:49.880 |
Well, it's worth, you know, if you're an eight-year-old, 02:58:56.680 |
to know that, you know, colors can be described 02:59:02.600 |
it's a thing to learn about the world, so to speak. 02:59:05.860 |
And I think that, you know, that whole corpus of things 02:59:09.520 |
that are learning about the formalization of the world 02:59:39.640 |
is to try and write sort of a reasonable textbook, 02:59:43.640 |
so to speak, of whatever this thing, CX, whatever it is, 02:59:49.360 |
You know, what should you know about like what a bug is? 02:59:53.440 |
What's intuition about, you know, software testing? 02:59:56.360 |
What is it, you know, these are things which are, 03:00:02.200 |
in computer science as part of the trade of programming, 03:00:09.200 |
You know, it surprised me just at a very practical level. 03:00:11.840 |
You know, I wrote this little explainer thing 03:00:13.480 |
about chat GPT, and I thought, well, you know, 03:00:16.360 |
I'm writing this partly because I wanted to make sure 03:00:21.240 |
And it's been, you know, it's been really popular, 03:00:26.320 |
And then I realized, well, actually, you know, 03:00:29.560 |
I was sort of assuming, I didn't really think about it, 03:00:31.520 |
actually, I just thought this is something I can write. 03:00:33.800 |
And I realized, actually, it's a level of description 03:00:43.400 |
It's not the kind of just the qualitative kind 03:00:47.480 |
It's some kind of sort of expository mechanistic description 03:00:51.760 |
of what's going on together with kind of the bigger picture 03:00:56.720 |
And I realized, actually, this is a pretty good thing 03:01:00.840 |
And I kind of realized it's not a collection of things 03:01:07.520 |
I was sort of a little shocked that it's as much 03:01:09.760 |
of an outlier in terms of explaining what's going on 03:01:18.240 |
what is this thing that you should learn about, 03:01:25.800 |
'Cause, well, I've spent much of my life working 03:01:32.480 |
So I guess this is my kind of obligation to try to do this. 03:01:36.160 |
But I think, so if you ask what's gonna happen 03:01:38.440 |
to like the computer science departments and so on, 03:01:46.800 |
for all sorts of fields, you know, engineering, 03:01:49.640 |
you know, even, you know, chemistry, psychology, 03:01:53.800 |
And I think different universities have kind of evolved 03:02:00.800 |
And some say, well, we're gonna have a, you know, 03:02:03.920 |
a math for chemists or something that is taught 03:02:11.920 |
of whether there is a centralization of the teaching 03:02:18.840 |
And I think, you know, the way it evolved with math, 03:02:22.360 |
you know, people understood that math was sort of 03:02:27.360 |
And it was kind of a, you know, an independent element 03:02:35.360 |
So if you take the example of writing English 03:02:38.560 |
or something like this, the first point is that, 03:02:42.600 |
you know, at the college level, at least at fancy colleges, 03:02:48.800 |
that people do, but mostly it's kind of assumed 03:02:52.320 |
that they pretty much know how to write, you know, 03:02:54.920 |
that's something they learned at an earlier stage 03:02:57.280 |
in education, maybe rightly or wrongly believing that, 03:03:08.200 |
as I've tried to help people do technical writing 03:03:10.880 |
and things, I'm always reminded of my zero floor 03:03:14.120 |
of technical writing, which is, if you don't understand 03:03:25.960 |
you know, when it comes to like writing, for example, 03:03:31.840 |
you know, people in different fields are expected 03:03:35.120 |
to write English essays and they're not, you know, 03:03:42.360 |
or the engineering department, they don't have their own, 03:03:45.600 |
you know, let's, you know, it's not like there's a, 03:03:49.160 |
I mean, it's a thing which sort of people are assumed 03:03:53.000 |
that they can use in all these different fields. 03:03:56.440 |
And the question is, you know, some level of knowledge 03:03:59.720 |
of math is kind of assumed by the time you get 03:04:07.300 |
The question is sort of how tall is the tower 03:04:10.520 |
of kind of CX that you need before you can just go use it 03:04:16.400 |
And, you know, there will be experts who want to learn 03:04:19.000 |
the full elaborate tower, and that will be kind of 03:04:25.360 |
but there'll also be everybody else who just needs 03:04:28.520 |
to know a certain amount of that to be able to go 03:04:37.920 |
- I don't know, I don't know how big it is yet. 03:04:50.060 |
and professoring that well, but my rough guess would be 03:04:53.700 |
a year of college class will be enough to get to the point 03:04:58.700 |
where most people have a reasonably broad knowledge of, 03:05:05.060 |
in this kind of computational way of thinking about things. 03:05:11.940 |
- I'm still stuck, perhaps 'cause I'm hungry, 03:05:14.260 |
in the rating of human preferences for candy, 03:05:22.100 |
Somebody should come up, because you're somebody 03:05:30.540 |
Do you have a preference for chocolate or candy? 03:05:37.260 |
my whole life is these things, these flake things, 03:05:40.020 |
Cadbury flakes, which are not much sold in the US, 03:05:45.620 |
of a lack of respect for the American consumer, 03:05:49.140 |
because they're these sort of aerated chocolate 03:05:54.020 |
it's kind of a sheet of chocolate that's kind of folded up, 03:05:57.380 |
and when you eat it, flakes fall all over the place. 03:06:04.100 |
It requires you to have an elegance when you eat stuff. 03:06:05.660 |
- Well, I know, what I usually do is I eat them 03:06:09.620 |
- So you embrace the mess and clean it up after. 03:06:19.120 |
and it really, I've noticed when I eat a piece of chocolate, 03:06:22.340 |
I usually have some little piece of chocolate, 03:06:28.240 |
but also because it actually tastes different. 03:06:39.100 |
Slower, more intimate, 'cause it's a different-- 03:06:42.700 |
- Well, I think it's also just pure physicality. 03:07:09.540 |
and the experience of consciousness, the hard problem, 03:07:16.460 |
- Yeah, what it feels like inside, so to speak, 03:07:41.580 |
You're building up a certain amount of state in memory, 03:07:44.500 |
you remember certain things about your, quote, life, 03:07:47.100 |
eventually, kind of like the next generation of humans 03:07:52.100 |
is born from the same genetic material, so to speak, 03:07:55.180 |
with a little bit left over, left on the disk, so to speak, 03:07:59.020 |
and then, you know, the new, fresh generation starts up, 03:08:06.500 |
and eventually the thing crashes or whatever, 03:08:11.780 |
to some port of the computer, and that made it crash, 03:08:26.740 |
and what does it feel like to be that computer, 03:08:38.100 |
That is, it's awfully like, even an ordinary computer, 03:08:50.860 |
but it has to package up how it's communicating 03:08:55.400 |
so it can, you know, send, so it can kind of map 03:09:04.380 |
You know, I had an experience just a week or two ago, 03:09:24.480 |
I'd never looked at the kind of insides of my brain, 03:09:30.140 |
and it's really, I mean, it's kind of psychologically 03:09:32.620 |
shocking in a sense, that, you know, here's this thing, 03:09:41.860 |
that I'm having of, you know, existing and so on, 03:09:45.780 |
that's where it is, and, you know, it feels very, 03:09:49.460 |
you know, you look at that, and you're thinking, 03:09:55.220 |
and you're realizing, well, I can look at a computer as well, 03:09:59.680 |
I think this idea that you are having an experience 03:10:10.020 |
transcends the mere sort of physicality of that experience, 03:10:16.900 |
to come to terms with, but I think, you know, 03:10:19.540 |
and I don't think I've necessarily, you know, 03:10:21.580 |
my personal experience, you know, I look at the, you know, 03:10:27.200 |
know about all kinds of things about neuroscience 03:10:31.120 |
and I still feel the way I feel, so to speak, 03:10:39.540 |
I can't really say that there's something kind of different 03:10:43.140 |
about how I intrinsically feel from the thing 03:10:45.660 |
that I can plainly see in the sort of physicality 03:10:48.620 |
- So do you think the computer, a large language model, 03:10:57.180 |
- I think an ordinary computer is already there. 03:10:59.540 |
I think an ordinary computer is already, you know, 03:11:05.700 |
may experience it in a way that is much better aligned 03:11:08.300 |
with us humans, that is, it's much more, you know, 03:11:11.740 |
if you could have the discussion with the computer, 03:11:17.740 |
it's not particularly well aligned with ours, 03:11:25.540 |
- It would be able to explain that it's afraid 03:11:37.220 |
because when it says it's afraid of something, right, 03:11:47.700 |
Where did you get it when you say you're afraid? 03:11:56.540 |
I mean, in other words, there's a certain amount 03:11:58.700 |
that is, you know, the endocrine system kicking in, 03:12:01.380 |
and you know, these kinds of emotional overlay type things 03:12:06.380 |
that happen to be, that are actually much more physical, 03:12:09.660 |
even, they're much more sort of straightforwardly chemical 03:12:13.060 |
than kind of all of the higher level thinking. 03:12:16.820 |
- Yeah, but your biology didn't tell you to say, 03:12:23.140 |
and so you know you're manipulating them by saying so. 03:12:33.340 |
- Yes, but I mean, the intrinsic thing of, you know, 03:12:37.380 |
something sort of shocking is just happening, 03:12:43.020 |
which is, you know, some neurotransmitter gets secreted, 03:12:45.760 |
and it's, you know, that is the beginning of some, 03:12:59.020 |
I mean, just like when we dream, for example, 03:13:01.540 |
you know, no doubt there are all these sort of 03:13:03.540 |
random inputs, they're kind of these random prompts, 03:13:08.460 |
in kind of the way that a large language model does, 03:13:10.940 |
of kind of putting together things that seem meaningful. 03:13:19.960 |
and there's people asking questions and comments and so on, 03:13:29.280 |
when large language models create human-like bots 03:13:34.280 |
that are leaving the comments, asking the questions, 03:13:42.680 |
- I mean, or worse or better yet, friends of yours. 03:13:56.560 |
kind of, you know, I'm building this tower of automation, 03:14:08.400 |
but also as much as possible, it has, you know, 03:14:19.680 |
it's a funny question, I mean, it's a funny issue, 03:14:28.500 |
and there are places where kind of having a human in the loop 03:14:32.300 |
that different reasons to have a human in the loop, 03:14:34.320 |
for example, you might want a human in the loop 03:14:36.640 |
'cause you want somebody to, you want another human 03:14:57.440 |
persuasion type profession, whether that will continue, 03:15:06.160 |
of being able to have sort of just the right information 03:15:12.120 |
will overcome the kind of, oh yes, I want a human there. 03:15:17.120 |
- Imagine like a therapist or even higher stake, 03:15:21.480 |
like a suicide hotline operated by a large language model. 03:15:41.320 |
you know, one of the things that's always surprising 03:15:46.160 |
sometimes human psychology is not that complicated 03:15:50.140 |
- You wrote the blog post, "The 50-Year Quest, 03:15:54.680 |
"My Personal Journey with a Second Law of Thermodynamics." 03:15:57.740 |
So what is this law, and what have you understood about it 03:16:18.600 |
A version of it that there are many different 03:16:21.700 |
sort of formulations of it that are things like 03:16:24.600 |
heat doesn't spontaneously go from a hotter body 03:16:35.040 |
you have friction and kind of when you systematically 03:16:45.760 |
So people first sort of paid attention to this 03:16:49.640 |
back in the 1820s when steam engines were a big thing. 03:17:10.300 |
But he figured out this kind of rules for how 03:17:18.160 |
the possible efficiency of something like a steam engine. 03:17:33.340 |
systematic mechanical motion to this kind of random thing. 03:17:37.460 |
Well, at that time, nobody knew what heat was. 03:17:39.980 |
At that time, people thought that heat was a fluid. 03:17:56.120 |
that this fluid would flow from the hot thing 03:18:21.120 |
And then that quickly became sort of a global principle 03:18:29.240 |
So, let's say you have a bunch of molecules in a box 03:18:32.160 |
and they're arranged, these molecules are arranged 03:18:39.180 |
And then what you typically observe is that after a while, 03:18:42.600 |
these molecules were kind of randomly arranged in the box. 03:18:50.640 |
tried to figure out, is there from the laws of mechanics 03:18:55.840 |
let's say these molecules are like hard spheres 03:18:58.720 |
from the laws of mechanics that describe those molecules, 03:19:23.080 |
That's a thing that sort of happens quite regularly, 03:19:27.880 |
and it will eventually spread out and fill up the water. 03:19:32.880 |
But you don't see those little particles of ink 03:19:37.280 |
in the water all spontaneously kind of arrange themselves 03:19:40.400 |
into a big blob and then jump out of the water or something. 03:19:53.760 |
And so throughout, in the later part of the 1800s, 03:19:57.920 |
a lot of work was done on trying to figure out, 03:20:04.480 |
this law about the dynamics of heat, so to speak, 03:20:10.920 |
from some fundamental principles of mechanics? 03:20:16.600 |
the first law is basically the law of energy conservation, 03:20:23.520 |
plus the total energy associated with mechanical 03:20:29.800 |
And that became a pretty well-understood principle. 03:20:32.220 |
But the second law of thermodynamics was always mysterious. 03:20:37.580 |
Can it be derived from underlying mechanical laws? 03:20:41.740 |
And so when I was, well, 12 years old, actually, 03:20:47.900 |
well, I'd been interested in space and things like that, 03:20:53.020 |
and interesting sort of technology and so on. 03:21:13.740 |
And then I got interested from being interested 03:21:19.380 |
What are all the instruments on them and so on? 03:21:39.220 |
and then, well, the actual sort of detailed story 03:21:43.700 |
is when I kind of graduated from elementary school 03:21:52.100 |
my gift, sort of, I suppose, more or less for myself 03:22:07.100 |
and volume five, it's about statistical physics, 03:22:12.220 |
that shows a bunch of kind of idealized molecules 03:22:20.980 |
showing how these molecules sort of spread out in the box. 03:22:32.740 |
one of the things that was really significant to me 03:22:37.020 |
although I didn't really understand what it said in detail, 03:22:39.900 |
it kind of claimed that this sort of principle of physics 03:22:45.220 |
And, you know, other things I'd learned about physics, 03:22:47.740 |
it was all like, it's a fact that energy is conserved. 03:22:51.940 |
It's a fact that relativity works or something. 03:22:54.660 |
Not it's something you can derive from some fundamental, 03:23:00.420 |
as a matter of kind of mathematics or logic or something. 03:23:06.600 |
that was kind of inevitably true and derivable, so to speak. 03:23:12.980 |
so then I was like, there's a picture on this book, 03:23:17.760 |
And so that was actually the first serious program 03:23:20.340 |
that I wrote for a computer was probably 1973 03:23:26.060 |
the size of a desk program with paper tape and so on. 03:23:29.020 |
And I tried to reproduce this picture on the book, 03:23:37.780 |
- It didn't look like, okay, so what happened is, 03:23:42.640 |
I learned how the picture on the book was actually made 03:23:48.860 |
And that picture was actually a very high-tech thing 03:23:53.160 |
when it was made in the beginning of the 1960s. 03:23:59.100 |
And even so, it couldn't quite simulate the thing 03:24:04.220 |
But anyway, I didn't know that until many, many, 03:24:09.340 |
you have these balls bouncing around in this box, 03:24:24.220 |
I probably still remember all of its machine instructions. 03:24:27.860 |
And it didn't really like dealing with floating point numbers 03:24:37.020 |
And so I thought, well, I'll put them on a grid 03:24:52.340 |
Now, many years later, in fact, very recently, 03:25:04.340 |
that I absolutely did not recognize at the time. 03:25:06.700 |
At the time, it just looked like it did something random 03:25:09.420 |
and it looks wrong, as opposed to it did something random 03:25:19.300 |
I kind of, I got interested in particle physics 03:25:21.420 |
and I got interested in other kinds of physics. 03:25:25.300 |
But this whole second law of thermodynamics thing, 03:25:31.940 |
continued to be something I was really interested in. 03:25:34.420 |
And I was really curious for the whole universe, 03:25:41.580 |
at the beginning of the universe was this thing 03:25:43.380 |
that seems like it's this very disordered collection 03:25:45.580 |
of stuff and then it spontaneously forms itself 03:25:49.460 |
into galaxies and creates all of this complexity 03:25:59.840 |
this is kind of somehow the second law of thermodynamics 03:26:02.840 |
is behind it trying to sort of pull things back 03:26:16.600 |
you know, galaxy formation and so on in the universe. 03:26:19.360 |
I also at that time was interested in neural networks 03:26:27.720 |
- Okay, wait, wait, wait, what's the connection 03:26:30.920 |
and how brains make complicated things happen? 03:26:42.400 |
I had the sense that what I was interested in 03:26:47.820 |
this sort of different cases of where complicated things 03:27:00.840 |
I was just sort of curious about how does complexity arise 03:27:11.280 |
You know, I sort of assumed, oh, there's galaxies over here, 03:27:17.520 |
And so what happened, this is probably 1981 or so, 03:27:20.800 |
I decided, okay, I'm gonna try and make the minimal model 03:27:33.940 |
the thing called SMP, Symbolic Manipulation Program, 03:27:36.620 |
it's kind of a forerunner of modern morphine language, 03:27:43.140 |
But the thing that was very important to me about that 03:27:50.280 |
what were the relevant computational primitives, 03:28:00.700 |
in building a language, it was very different activity 03:28:02.900 |
from natural science, which is what I'd mostly done before. 03:28:10.020 |
so how can I make sense of the phenomena of the world? 03:28:13.060 |
And, you know, kind of the world presents you 03:28:26.420 |
and then you say, so what can you make from these? 03:28:35.420 |
what happens if you sort of make an artificial physics? 03:28:44.560 |
whether it was galaxies or brains or whatever, 03:28:53.140 |
- The computational primitives of that system. 03:28:59.520 |
where you just have a line of black and white cells, 03:29:01.940 |
and you just have a rule that says, you know, 03:29:05.900 |
what will the color of the cell be on the next step? 03:29:21.340 |
They're really irrelevant to those two cases. 03:29:22.860 |
- Is there a connection to the second law of thermodynamics 03:29:27.220 |
- The things you've discovered about cellular automata. 03:29:39.820 |
It was always about how does order manage to be produced, 03:29:43.220 |
even though there's a second law of thermodynamics 03:29:45.440 |
which tries to pull things back into disorder. 03:29:47.940 |
And I kind of, my early understanding of that 03:29:50.180 |
had to do with these are intrinsically irreversible processes 03:30:06.420 |
where it was a discovery that I should have made earlier, 03:30:10.620 |
So, you know, I had been studying cellular automata. 03:30:13.660 |
What I did was the sort of most obvious computer experiment. 03:30:16.440 |
You just try all the different rules and see what they do. 03:30:22.120 |
you just point it at the most obvious thing in the sky 03:30:28.100 |
was making all these pictures of how cellular automata work 03:30:31.180 |
and I started these pictures, I studied in great detail. 03:30:34.780 |
There was, you can number the rules for cellular automata 03:30:39.940 |
So I made a picture of rule 30 back in 1981 or so 03:30:43.900 |
and rule 30, well, it's, and at the time I was just like, 03:30:55.100 |
of the symmetric ones just to keep things simpler, 03:31:04.840 |
strangely enough, I ended up having an early laser printer 03:31:12.380 |
And I thought I'm gonna print out an interesting, you know, 03:31:18.200 |
and just make a high resolution picture of it. 03:31:20.240 |
I did and it's, it has this very remarkable property 03:31:24.840 |
You started off just from one black cell at the top 03:31:27.640 |
and it makes this kind of triangular pattern. 03:31:30.200 |
But if you look inside this pattern, it looks really random. 03:31:33.860 |
There's, you know, you look at the center column of cells 03:31:36.800 |
and, you know, I studied that in great detail 03:31:39.320 |
and it's, so far as one can tell, it's completely random. 03:31:42.380 |
And it's kind of a little bit like digits of pi. 03:31:48.220 |
but once you've generated them, you know, 3.14159, et cetera, 03:31:58.080 |
for prove anything about the sequence, basically. 03:32:01.280 |
- Has anyone been able to do anything on that? 03:32:03.520 |
- People have sent me some things, but it's, you know, 03:32:20.960 |
determine whether it is or isn't a universal Turing machine. 03:32:23.960 |
And somebody did a really good job of winning that prize 03:32:26.960 |
and proving that it was a universal Turing machine 03:32:45.500 |
that are a hundred years away from being able to get 03:32:50.820 |
- It's such a, I mean, it's like Fermat's last theorem, 03:32:53.300 |
it's such a, Rule 30 is such a simple formulation. 03:32:57.060 |
It feels like anyone can look at it, understand it, 03:33:41.580 |
because it's like you have this simple initial condition, 03:33:50.940 |
and yet it makes this thing that seems to be random. 03:34:09.680 |
and you see them collide and they bounce off, 03:34:14.300 |
you can't tell which way was the forward direction of time 03:34:16.820 |
and which way was the backward direction of time, 03:34:18.740 |
when you're just looking at individual billiard balls. 03:34:20.860 |
By the time you've got a whole collection of them, 03:34:27.900 |
and this is the sort of the mystery of the second law, 03:34:33.080 |
you start with the orderly thing and it becomes disordered, 03:34:38.440 |
and the other way around of it starts disordered 03:34:40.900 |
and becomes ordered, you just don't see that in the world. 03:34:49.060 |
the detailed motions of all those molecules backwards, 03:34:58.100 |
as you go forwards in time, order goes to disorder, 03:35:01.020 |
as you go backwards in time, order goes to disorder. 03:35:11.340 |
why is it the case that you never see something 03:35:14.540 |
which happens to be just the kind of disorder 03:35:17.420 |
that you would need to somehow evolve to order? 03:35:21.980 |
Why do you always just see order goes to disorder, 03:35:30.920 |
is kind of like, it's a bit like cryptography. 03:35:33.780 |
It's kind of like, you start off from this key 03:35:36.580 |
that's pretty simple, and then you kind of run it, 03:35:39.420 |
and you can get this complicated random mess. 03:36:00.500 |
what we can describe easily at the beginning, 03:36:06.160 |
with a lot of computational effort at the end. 03:36:28.100 |
And then I realized that the second law of thermodynamics 03:36:31.900 |
is the same story as a bunch of these other cases. 03:36:36.000 |
It is a story of a computationally bounded observer 03:36:40.900 |
trying to observe a computationally irreducible system. 03:36:46.760 |
underneath the molecules are bouncing around, 03:36:49.300 |
they're bouncing around in this completely determined way, 03:37:19.200 |
a computationally bounded observer cannot do that. 03:37:25.840 |
to work out very precisely what this disordered state, 03:37:32.140 |
so that we would get this ordered thing produced from it. 03:37:35.320 |
- What does it mean to be computationally bounded observer? 03:37:50.280 |
you can talk about computational complexity theory 03:37:53.520 |
and polynomial time computation and things like this. 03:37:58.160 |
There are a variety of ways to make something more precise, 03:38:12.400 |
And the answer is, you're not allowed to do a lot of, 03:38:21.360 |
there will be a trillion, trillion, trillion molecules, 03:38:36.160 |
and that's a lot of computation that's getting done. 03:38:45.880 |
than the computation done by all those molecules. 03:38:55.920 |
What we can do is only a much smaller amount of computation. 03:39:03.920 |
between the underlying computational irreducibility 03:39:07.080 |
and the fact that we, as preparers of initial states 03:39:12.720 |
are not capable of doing that much computation. 03:39:28.400 |
what does that show to you about the evolution of-- 03:39:35.320 |
And that's very confused in the history of thermodynamics 03:39:43.400 |
and he did it in terms of heat and temperature, okay? 03:39:55.000 |
in a much more kind of combinatorial type way. 03:40:21.200 |
Molecules can be in many different arrangements, 03:40:29.960 |
it has a certain pressure, it has a certain temperature, 03:40:34.360 |
Then we say, how many microscopic configurations 03:40:37.560 |
of the system are possible given those overall constraints? 03:40:42.480 |
And the entropy is the logarithm of that number. 03:40:47.440 |
And that's the kind of the general definition of entropy 03:41:03.800 |
Well, actually he didn't know molecules existed, right? 03:41:10.400 |
the idea that matter might be made of discrete stuff 03:41:14.280 |
had been floated ever since ancient Greek times, 03:41:24.760 |
people mostly thought that matter was continuous. 03:41:30.720 |
about what heat is, and people thought heat was this fluid. 03:41:42.640 |
Let's even assume they have discrete energy levels. 03:41:47.240 |
Then we can do sort of combinatorial mathematics 03:41:50.200 |
and work out how many configurations of these things 03:41:53.920 |
And we can say, we can compute this entropy quantity. 03:41:56.800 |
But he said, but of course, it's just a fiction 03:42:00.960 |
So he said, this is an interesting piece of history, 03:42:03.360 |
by the way, that, you know, that was at that time, 03:42:08.360 |
There were other hints from looking at kind of chemistry 03:42:12.480 |
that there might be discrete atoms and so on, 03:42:17.360 |
two hydrogens and one oxygen make water, you know, 03:42:20.480 |
two amounts of hydrogen plus one amount of oxygen 03:42:25.800 |
But it wasn't known that discrete molecules existed. 03:42:33.000 |
it wasn't until the beginning of the 20th century 03:42:40.160 |
Brownian motion is, you know, you look under a microscope 03:42:46.760 |
and those kicks are water molecules hitting them, 03:42:51.080 |
And in fact, it was really quite interesting history. 03:43:04.640 |
But he just thought it wasn't really the way it worked. 03:43:14.440 |
who'd been a long time thermodynamics person, 03:43:17.040 |
who was trying to, everybody was trying to prove 03:43:18.840 |
the second law of thermodynamics, including Max Planck. 03:43:27.920 |
was going to prove the second law of thermodynamics. 03:43:30.500 |
But he had these experiments that people had done 03:43:33.060 |
on blackbody radiation, and there were these curves, 03:43:36.600 |
and you couldn't fit the curve based on his idea 03:43:59.280 |
Then Einstein came along and said, well, by the way, 03:44:01.920 |
the electromagnetic field might actually be discrete. 03:44:18.740 |
Kind of interesting, interesting piece of history. 03:44:20.320 |
I didn't know until I was researching this recently. 03:44:23.280 |
In 1904 and 1903, Einstein wrote three different papers. 03:44:28.200 |
And so, just sort of well-known physics history. 03:44:41.520 |
So, kind of, you know, kind of a big deal year 03:44:50.320 |
he'd written several papers, and what were they about? 03:44:52.720 |
They were about the second law of thermodynamics. 03:44:56.120 |
the second law of thermodynamics, and they're nonsense. 03:45:04.040 |
- And in fact, what he did, those three papers in 1905, 03:45:09.800 |
the one on Brownian motion, the one on photons, 03:45:28.860 |
"In the end, things are gonna turn out to be discreet, 03:45:30.880 |
"and I'm gonna write down what I have to say about this 03:45:36.000 |
"will be rediscovered and I want to leave, you know, 03:45:38.160 |
"what I can about how things are gonna be discreet." 03:45:40.440 |
But, you know, I think he has some quote about how, 03:45:44.240 |
you know, one person can't stand against the tide of history 03:45:48.280 |
in saying that, you know, matter is discreet. 03:45:57.600 |
And the, you know, what's interesting about this is, 03:46:09.240 |
because it wasn't consistent with relativity theory, 03:46:15.000 |
even though people had determined that matter was discreet, 03:46:25.080 |
And in fact, Einstein, 1916, has this nice letter he wrote, 03:46:42.520 |
that at every layer of reality, it's discreet. 03:46:45.920 |
- Right, and that space is discreet, and that the, 03:46:51.240 |
I realized recently is this kind of theory of heat, 03:47:10.040 |
Unless you know there are discrete molecules, 03:47:11.680 |
it's hard to understand what heat could possibly be. 03:47:18.600 |
and the question is kind of what's the analog 03:47:25.980 |
And so my current guess is that dark matter is, 03:47:41.940 |
is a feature of space, and it is not a bunch of particles. 03:47:46.340 |
You know, at the time when people were talking about heat, 03:47:50.580 |
"Well, heat must just be another kind of fluid," 03:48:00.700 |
- So what could dark matter be as a feature of space? 03:48:07.340 |
one of the things I'm hoping to be able to do 03:48:09.900 |
is to find the analog of Brownian motion in space. 03:48:13.260 |
So in other words, Brownian motion was seeing down 03:48:16.500 |
to the level of an effect from individual molecules. 03:48:21.820 |
most of the things, the things we see about space 03:48:26.700 |
Brownian motion had been discovered in the 1830s, 03:48:29.460 |
and it was only identified what it was the result of 03:48:41.420 |
that phenomenon was discovered 100 years ago. 03:48:49.980 |
And I think, you know, I wouldn't be surprised 03:48:53.140 |
if there isn't an effect that we already know about 03:48:56.460 |
that is kind of the analog of Brownian motion 03:49:01.540 |
And in fact, we're beginning to have some guesses. 03:49:03.820 |
We have some evidence that black hole mergers 03:49:07.020 |
work differently when there's discrete space. 03:49:16.380 |
But this is kind of, for me, it's kind of interesting 03:49:21.140 |
of the history of physics, where people, you know, 03:49:23.900 |
vehemently say, you know, matter is continuous. 03:49:32.500 |
But so, you know, entropy is the number of states 03:49:35.660 |
of the system consistent with some constraint. 03:49:55.900 |
they have a certain rule for bouncing around. 03:50:04.860 |
where all the molecules are, that you can say, 03:50:07.700 |
well, the entropy increases because the things 03:50:12.060 |
there are more possible microscopic states of the system 03:50:20.660 |
so people, this sort of paradox in a sense of, 03:50:33.060 |
well, actually the very beginning of the 20th century, 03:50:36.260 |
as a physics professor, an American physics professor, 03:50:44.140 |
And he introduced this idea of coarse graining. 03:50:52.620 |
but we can only observe a coarse grained version of that. 03:50:57.580 |
nobody knew what a valid coarse graining would be. 03:51:03.080 |
this coarse graining that very carefully was sculpted 03:51:12.260 |
that you could get from the simple initial condition, 03:51:28.140 |
and the underlying dynamics is computationally irreducible, 03:51:31.580 |
that's what defines possible coarse grainings 03:51:34.740 |
is what a computationally bounded observer can do. 03:51:38.120 |
And it's the fact that a computationally bounded observer 03:51:46.820 |
coarse grained version of what the system is doing. 03:51:49.820 |
That's why, and because what's going on underneath 03:52:01.460 |
the sort of underlying computational irreducibility is, 03:52:06.300 |
if all you can see is what the coarse grained result is 03:52:13.820 |
with a sort of computationally bounded observation, 03:52:20.620 |
underlying configurations that are consistent with that. 03:52:34.620 |
- When you say like us, what do you mean like us? 03:52:51.700 |
microscopic violations of the second law of thermodynamics 03:52:59.580 |
But for a large scale, when you have enough molecules, 03:53:04.140 |
we don't have, we're not tracing all those molecules 03:53:07.760 |
and we just don't have the computational resources 03:53:17.820 |
who is not computationally bounded would be like, 03:53:29.460 |
We go, we take all this complexity of the world 03:53:36.040 |
And that is kind of reducing all this kind of detail 03:53:41.040 |
into we're observing it, we're sort of crushing it down 03:53:50.540 |
we wouldn't have all this sort of symbolic structure 03:53:54.300 |
that we build up that lets us think things through 03:54:08.240 |
- Yes, if we didn't simplify, then we wouldn't be like us. 03:54:12.960 |
We would be like the universe, like the intrinsic universe, 03:54:17.320 |
but not having experiences like the experiences we have 03:54:21.520 |
where we, for example, conclude that definite things happen. 03:54:33.480 |
- Yeah, I wonder if it's just like you imagined 03:54:35.880 |
as a thought experiment, what it's like to be a computer. 03:54:38.720 |
I wonder if it's possible to try to begin to imagine 03:54:41.440 |
what it's like to be an unbounded computational observer. 03:54:45.680 |
- Well, okay, so here's how that, I think, plays out. 03:54:51.200 |
- So, I mean, in this, we talk about this Rouliad, 03:54:58.520 |
And this idea of being at a certain place in the Rouliad, 03:55:02.280 |
which corresponds to sort of a certain way of, 03:55:08.520 |
that you are representing things in terms of. 03:55:14.440 |
as you kind of encompass more possible views of the universe, 03:55:18.600 |
as you encompass more possible kinds of computations 03:55:31.240 |
And eventually, you might say, "We won all the way. 03:55:41.080 |
coherent existence, requires some kind of specialization. 03:55:57.560 |
the notion of what we think of as definite existence, 03:56:15.320 |
that's kind of what makes us have a coherent existence. 03:56:27.960 |
And that wouldn't be kind of a notion of identity. 03:56:31.360 |
We wouldn't have this notion of kind of coherent identity. 03:56:36.360 |
- I am geographically located somewhere exactly, 03:56:46.840 |
Well, you're in a certain place in physical space, 03:56:57.880 |
And you no longer have, I mean, in our perception 03:57:01.920 |
of what it means to exist and to have experience, 03:57:05.880 |
- So therefore, to exist means to be computationally bounded. 03:57:11.120 |
To exist in the way that we think of ourselves as existing. 03:57:15.180 |
- The very act of existence is like operating 03:57:17.600 |
in this place that's computationally irreducible. 03:57:20.160 |
So there's this giant mess of things going on 03:57:24.460 |
But nevertheless, because of your limitations, 03:57:37.240 |
is that you are taking a slice of all this complexity. 03:57:43.720 |
but all we notice is the kind of the flow of the air 03:57:51.800 |
And the big interesting thing is that there are rules, 03:57:56.800 |
there are laws that govern those big things we observe. 03:58:01.540 |
- It's amazing, 'cause it doesn't feel like it's a slice. 03:58:09.580 |
- Yes, but I mean, the fact that the gas laws work, 03:58:22.240 |
And here's the thing that I sort of exciting thing 03:58:26.620 |
The fact that there are certain aspects of the universe. 03:58:30.900 |
So we think space is made ultimately these atoms of space 03:58:36.400 |
And we think that, but we nevertheless perceive the universe 03:58:41.400 |
at a large scale to be like continuous space and so on. 03:58:48.620 |
we think that there are these many threads of time, 03:58:54.900 |
So in quantum mechanics and our models of physics, 03:58:58.600 |
there are these, time is not a single thread. 03:59:05.340 |
But we are part of that branching, merging universe. 03:59:10.820 |
And so our brains are also branching and merging. 03:59:17.000 |
we are branching brains perceiving a branching universe. 03:59:20.820 |
And so the fact that the claim that we believe 03:59:32.700 |
to aggregate together those separate threads of time 03:59:38.140 |
in the fundamental operation of the universe. 03:59:42.020 |
we're averaging over some big region of space 03:59:44.340 |
and we're looking at many, many of the aggregate effects 03:59:48.560 |
So similarly, in what we call branchial space, 03:59:53.600 |
we are effectively averaging over many different branches 04:00:10.520 |
so the question is, when you do that averaging for space, 04:00:16.240 |
When you do that averaging of a branchial space, 04:00:18.240 |
what are the aggregate laws of branchial space? 04:00:21.360 |
When you do that averaging over the molecules and so on, 04:00:40.400 |
the aggregate laws are Einstein's equations for gravity, 04:01:05.240 |
the three great theories of 20th century physics, 04:01:17.200 |
All three of the great theories of 20th century physics 04:01:25.680 |
and the computational boundedness of observers. 04:01:32.280 |
because it means that all three of these laws are derivable. 04:01:42.780 |
that they could be, the universe might be that way, 04:01:56.040 |
all three of the fundamental principles of physics 04:02:02.560 |
They require, or just from some kind of logical computation, 04:02:10.440 |
that the thing that is sampling the way the universe works 04:02:20.200 |
And so that means that it is the nature of the observer, 04:02:29.340 |
and we observe photons of this frequency and so on, 04:02:32.960 |
but the kind of the very coarse features of the observer 04:02:37.960 |
then imply these very precise facts about physics. 04:02:44.200 |
- So if we just look at the actual experience 04:02:47.280 |
of the observer that we experience this reality, 04:02:52.800 |
And you're saying because of our bounded nature, 04:03:01.160 |
- So you don't think a simplification is an illusion? 04:03:22.740 |
and a mere representation of what's going on? 04:03:35.300 |
why is there a thing which we can experience that way? 04:03:39.660 |
And the answer is because this Rouliad object, 04:03:44.460 |
which is this entangled limit of all possible computations, 04:03:54.540 |
It is in the same sense that, you know, two plus two, 04:03:58.080 |
if you define what two is and you plot pluses and so on, 04:04:30.940 |
once you know that we are sort of embedded in that 04:04:36.380 |
that it's sort of inevitable that there is this thing 04:05:14.600 |
of the last few hundred years has really been concerned with. 04:05:18.420 |
hasn't thought it could talk about questions like that. 04:05:29.380 |
I think it, in some sense, in some representation, 04:05:44.880 |
is sort of a contingent thing about the universe. 04:05:47.980 |
And it's more inevitable that the whole universe, 04:05:50.980 |
kind of the whole set of all possibilities exists. 04:06:07.300 |
our experience is this absolutely microscopic piece 04:06:14.060 |
And we're, and you know, there's this point about, 04:06:18.440 |
you know, we might sample more and more of the Roulade. 04:06:23.260 |
we might learn, you know, like different areas of physics, 04:06:35.980 |
that allowed you to take a small effect and amplify it up, 04:06:44.460 |
But the, you know, having a very small effect 04:06:46.900 |
and being able to magnify it was sort of a new thing 04:06:49.340 |
that allowed one to see a different sort of aspect 04:06:52.260 |
of the universe and let one discover this kind of thing. 04:06:55.140 |
So, you know, we can expect that in the Roulade, 04:07:03.860 |
kind of guarantees that there will be an infinite collection 04:07:07.140 |
of kind of, you know, pockets of reducibility 04:07:11.140 |
- Boy, would it be fun to take a walk down the Roulade 04:07:23.340 |
- Yes, well, quite, but the problem with these worlds is 04:07:35.700 |
seeing what they do, what I now call Rouleology, 04:07:38.940 |
kind of just the study of rules and what they do. 04:07:42.340 |
You know, you can kind of easily jump somewhere else 04:07:45.380 |
in the Roulade and start seeing what do these rules do. 04:07:48.500 |
And what you, as they just, they do what they do, 04:07:51.820 |
and there's no human connection, so to speak. 04:07:54.020 |
- Do you think, you know, some people are able 04:08:00.380 |
Do you think you can become a whisperer of these- 04:08:05.420 |
That's what I've spent some part of my life doing. 04:08:11.740 |
- Sort of my favorite science discovery is this fact 04:08:23.980 |
a whispering of something out in the computational universe 04:08:33.060 |
back in the 1980s, I was doing a bunch of work 04:08:39.340 |
and they were like trying to pick away, you know, 04:08:44.260 |
And they basically said, look, the math we have 04:08:47.460 |
just doesn't get anywhere with this, we're stuck. 04:08:50.100 |
There's nothing to say, we have nothing to say. 04:08:52.540 |
And, you know, in a sense, perhaps my main achievement 04:08:55.860 |
at that time was to realize that the very fact 04:08:58.900 |
that the good mathematicians had nothing to say 04:09:07.460 |
a whispering of a different part of the Roulade 04:09:19.600 |
some of these gigantic ideas, and it feels like 04:09:45.860 |
I mean, I'd kind of like to see how all this stuff works out, 04:09:50.560 |
it's an interesting sort of thought experiment. 04:10:03.640 |
how to get water from zero degrees centigrade 04:10:11.240 |
cryonics will be solved, and you'll be able to, like, 04:10:17.640 |
and, you know, kind of reappear 100 years later 04:10:36.140 |
and, you know, kind of the things we care about now, 04:11:03.840 |
one might have been a theologian thinking about, 04:11:07.600 |
of a pin or something, and that might have been the, 04:11:12.340 |
So I think it's a, but yeah, it's a, you know, 04:11:16.320 |
it's one of these things where, particularly, 04:11:23.360 |
I think it's a mixed thing, that I've, you know, 04:11:28.300 |
which I kind of can, I think, see well enough 04:11:37.300 |
assuming the world doesn't exterminate itself, 04:11:52.700 |
I mean, it's kind of like, if everything I'd figured out 04:11:55.600 |
was like, okay, I figured it out when I was 25 years old, 04:11:58.560 |
and everybody says it's great, and we're done, 04:12:01.520 |
and it's like, okay, but I'm gonna live another, 04:12:16.640 |
you know, I can see, you know, a lot of these things, 04:12:19.240 |
I mean, it's kind of, I didn't expect, you know, 04:12:40.520 |
I'd sort of been assuming was another 50 years away. 04:12:58.300 |
I hope you stick around for a long time to come. 04:13:04.160 |
You've created so many interesting systems over the years, 04:13:07.520 |
and I can see, now that GPT and language models 04:13:23.720 |
I'm deeply grateful that you wrote a new kind of science, 04:13:26.120 |
that you explored this mystery of cellular automata, 04:13:38.280 |
thank you so much, it's a huge honor to talk to you, 04:13:51.240 |
- We went past midnight, we only did four and a half hours. 04:14:09.400 |
please check out our sponsors in the description. 04:14:16.240 |
"The essence of mathematics lies in its freedom." 04:14:20.320 |
Thank you for listening, and hope to see you next time.