back to indexMarcus Hutter: Universal Artificial Intelligence, AIXI, and AGI | Lex Fridman Podcast #75
Chapters
0:0 Introduction
3:32 Universe as a computer
5:48 Occam's razor
9:26 Solomonoff induction
15:5 Kolmogorov complexity
20:6 Cellular automata
26:3 What is intelligence?
35:26 AIXI - Universal Artificial Intelligence
65:24 Where do rewards come from?
72:14 Reward function for human existence
73:32 Bounded rationality
76:7 Approximation in AIXI
78:1 Godel machines
81:51 Consciousness
87:15 AGI community
92:36 Book recommendations
96:7 Two moments to relive (past and future)
00:00:00.000 |
The following is a conversation with Marcus Hutter, 00:00:03.480 |
senior research scientist at Google DeepMind. 00:00:08.280 |
including with Juergen Schmidhuber and Shane Legg, 00:00:13.920 |
in and around the field of artificial general intelligence, 00:00:17.480 |
including the development of AIXI, spelled A-I-X-I, model, 00:00:25.360 |
that incorporates ideas of Kolmogorov complexity, 00:00:28.840 |
Solomonov induction, and reinforcement learning. 00:00:32.000 |
In 2006, Marcus launched the 50,000 Euro Hutter Prize 00:00:41.160 |
The idea behind this prize is that the ability 00:00:43.700 |
to compress well is closely related to intelligence. 00:00:50.320 |
Specifically, if you can compress the first 100 megabytes 00:00:54.880 |
or one gigabyte of Wikipedia better than your predecessors, 00:00:58.320 |
your compressor likely has to also be smarter. 00:01:08.360 |
In conjunction with his podcast release just a few days ago, 00:01:34.060 |
I'm a big fan of benchmarks for developing AI systems, 00:01:40.600 |
that will spark some good ideas for approaches 00:01:43.120 |
that will make progress on the path of developing AGI systems. 00:01:56.000 |
or simply connect with me on Twitter @LexFriedman, 00:02:02.600 |
As usual, I'll do one or two minutes of ads now, 00:02:27.920 |
Brokerage services are provided by Cash App Investing, 00:02:39.160 |
and security in all digital transactions is very important, 00:02:42.760 |
let me mention the PCI Data Security Standard 00:02:48.120 |
I'm a big fan of standards for safety and security. 00:02:57.180 |
and agreed that there needs to be a global standard 00:03:02.520 |
Now, we just need to do the same for autonomous vehicles 00:03:21.280 |
that is helping to advance robotics and STEM education 00:03:26.760 |
And now, here's my conversation with Markus Hutter. 00:03:41.560 |
I think it's a very interesting hypothesis or idea, 00:03:47.960 |
so I know a little bit about physical theories, 00:03:57.920 |
and they're all, in a sense, computable theories. 00:04:04.360 |
which describe virtually everything in the universe. 00:04:24.780 |
are so nice and beautiful and simple and compressible? 00:04:39.520 |
Our human minds just enjoy something about that simplicity, 00:04:46.720 |
- No, I strongly believe, and I'm pretty convinced 00:04:49.400 |
that the universe is inherently beautiful, elegant, 00:04:52.560 |
and simple, and described by these equations, 00:05:06.720 |
but we understand that it's an emergent phenomena, 00:05:14.640 |
All of chemistry follows that, and we know that. 00:05:18.140 |
because we have limited computational resources. 00:05:25.640 |
Maybe there's some corners very far out in the universe, 00:05:28.320 |
or super, super tiny below the nucleus of atoms, 00:05:33.000 |
or, well, parallel universes which are not nice and simple, 00:05:42.240 |
and choose the simplest tree consistent with it, 00:05:57.480 |
which sort of if you translate it to proper English means, 00:06:03.440 |
that if you have two theories or hypotheses or models 00:06:20.040 |
Perhaps we'll kind of discuss it and think about it, 00:06:23.480 |
but what's the intuition of why the simpler answer 00:06:28.080 |
is the one that is likelier to be more correct descriptor 00:06:36.560 |
is probably the most important principle in science. 00:06:44.520 |
but science is about understanding the world, 00:06:53.680 |
which explain everything but predict nothing, 00:06:55.980 |
but the simple model seem to have predictive power 00:07:09.200 |
and we use this principle and it seems to be successful. 00:07:12.820 |
We don't know why, but it just happens to be. 00:07:15.900 |
Or you can try, you know, find another principle 00:07:49.000 |
- So I apologize for the romanticized question, 00:07:51.620 |
but why do you think, outside of its effectiveness, 00:07:59.920 |
Why does it just, why does E equals MC squared 00:08:10.600 |
many things can be explained by an evolutionary argument. 00:08:14.960 |
And, you know, there's some artifacts in humans 00:08:17.320 |
which are just artifacts and not evolutionary necessary. 00:09:01.120 |
but, I mean, it's best, of course, if they are, 00:09:11.480 |
but indeed, in terms of just for survival purposes, 00:09:18.680 |
for why we find the work of Einstein so beautiful. 00:09:26.840 |
Could you describe what Solomonov induction is? 00:09:35.040 |
and where Solomonov sort of claimed a long time ago 00:09:37.680 |
that this solves the big philosophical problem of induction. 00:09:52.040 |
induction can be interpreted narrowly and widely. 00:09:59.000 |
And widely means also then using these models 00:10:06.560 |
So I'm a little sloppy sort of with the terminology, 00:10:09.440 |
and maybe that comes from Ray Solomonov being sloppy. 00:10:18.160 |
So let me explain a little bit this theory in simple terms. 00:10:31.000 |
The natural answer, I'm gonna speed up a little bit, 00:10:42.800 |
And why should it suddenly after 100 1s be different? 00:10:45.520 |
So what we're looking for is simple explanations or models 00:10:51.720 |
a model has to be presented in a certain language. 00:11:03.880 |
So abstractly on a Turing machine, for instance, 00:11:12.880 |
and then 100 0s, and 100 1s, that's a model, right? 00:11:30.920 |
It will not stop, it will continue, naturally. 00:11:36.120 |
And on the sequence of 1s, it's very plausible, right? 00:11:46.080 |
The short program is again, you know, counter. 00:11:53.160 |
The extra twist is that it can also deal with noisy data. 00:12:08.720 |
oh, the next coin flip will be head with probability 60%. 00:12:19.600 |
Well, in Solomon of induction, precisely what you do is, 00:12:22.240 |
so you combine, so looking for the shortest program 00:12:54.640 |
you have a prior, which is precisely two to the minus 00:13:02.680 |
And you weigh all this hypothesis and take this mixture 00:13:13.840 |
That seems to me, maybe a very human-centric concept, 00:13:25.080 |
You've used the term compression quite a bit. 00:13:33.080 |
and maybe science or just all of our intellectual pursuits 00:13:37.760 |
is basically the attempt to compress the complexity 00:13:43.600 |
So what does this word mean to you, compression? 00:13:54.440 |
finding short programs for the data or the phenomenon at hand 00:13:59.440 |
you could interpret it more widely as finding simple theories 00:14:09.480 |
Compression means finding short descriptions, 00:14:34.840 |
- Well, at least all of science I see as a endeavor 00:14:40.160 |
And well, there are also some other aspects of science 00:14:49.240 |
And that is then part of the decision-making process. 00:14:52.800 |
But once we have the data to understand the data 00:14:58.640 |
So I don't see any difference between compression, 00:15:05.060 |
- So we're jumping around topics a little bit, 00:15:10.560 |
a fascinating concept of Kolmogorov complexity. 00:15:31.200 |
And it takes the compression view to the extreme. 00:15:36.000 |
So I explained before that if you have some data sequence, 00:15:41.760 |
and best sort of, you know, just a string of bits. 00:15:49.440 |
like we compress big files into say zip files 00:15:53.720 |
And you can also produce self-extracting archives. 00:16:02.840 |
It's just the decompressor plus the archive together in one. 00:16:06.240 |
And now there are better and worse compressors. 00:16:08.800 |
And you can ask, what is the ultimate compressor? 00:16:11.120 |
So what is the shortest possible self-extracting archive 00:16:14.840 |
you could produce for a certain data set, yeah? 00:16:19.560 |
And the length of this is called the Kolmogorov complexity. 00:16:23.280 |
And arguably, that is the information content 00:16:27.960 |
I mean, if the data set is very redundant or very boring, 00:16:34.760 |
And you know, it is low according to this definition. 00:16:42.040 |
- And what's your sense of our sort of universe 00:16:46.280 |
when we think about the different objects in our universe, 00:16:51.280 |
that we try concepts or whatever at every level, 00:16:55.440 |
do they have high or low Kolmogorov complexity? 00:17:01.400 |
in being able to summarize much of our world? 00:17:08.520 |
So as I said before, I believe that the whole universe, 00:17:13.520 |
based on the evidence we have, is very simple. 00:17:19.240 |
- Sorry, to linger on that, the whole universe, 00:17:24.040 |
Do you mean at the very basic fundamental level 00:17:42.120 |
- Is noise a problem or is it a bug or a feature? 00:17:45.280 |
- I would say it makes our life as a scientist 00:17:55.980 |
- But then maybe we wouldn't feel like there's a free will. 00:18:12.800 |
So we can't get away with statistics even then. 00:18:21.260 |
But I mean, it's still so hard to compute the trajectory 00:18:39.960 |
then arguably you could describe the whole universe 00:18:43.240 |
as well as a standard model plus generativity. 00:18:47.460 |
I mean, we don't have a theory of everything yet, 00:18:49.680 |
but sort of assuming we are close to it or have it, yeah. 00:18:59.120 |
But that's spoiled by noise or by chaotic systems 00:19:03.560 |
or by initial conditions, which may be complex. 00:19:23.080 |
but when you just take a small window, then-- 00:19:26.120 |
- It may become complex and that may be counterintuitive, 00:19:34.280 |
So imagine you have a normal library with interesting books 00:19:43.360 |
which contains all possible books, say, of 500 pages. 00:19:46.840 |
So the first book just has AAAA over all the pages. 00:19:49.720 |
The next book, AAAA and ends with B and so on. 00:20:02.920 |
and suddenly you have a lot of information in there. 00:20:10.480 |
seems to be understudied or under-talked about 00:20:14.960 |
What lessons do you draw from sort of the game of life 00:20:18.600 |
for cellular automata where you start with the simple rules 00:20:20.840 |
just like you're describing with the universe 00:20:34.120 |
where, like you said, some chaotic behavior could happen, 00:20:39.400 |
it could die out in some very rigid structures? 00:20:54.240 |
is really great because these rules are so simple. 00:20:57.720 |
and even by hand you can simulate a little bit, 00:21:04.000 |
and people have proven that it's even Turing-complete. 00:21:06.800 |
You cannot just use a computer to simulate game of life, 00:21:09.840 |
but you can also use game of life to simulate any computer. 00:21:16.520 |
and it's the prime example probably to demonstrate 00:21:21.240 |
that very simple rules can lead to very rich phenomena. 00:21:44.600 |
You asked also about whether I understand this phenomenon, 00:21:58.360 |
And I think I'm pretty used to cellular automata, 00:22:09.280 |
I didn't play too much with this converse game of life, 00:22:20.000 |
And well, when the computers were really slow 00:22:25.360 |
and programmed my own programs in assembler too. 00:22:37.400 |
So I returned to this, you know, every couple of years, 00:22:40.320 |
and then I tried to understand what is going on, 00:22:48.720 |
you know, there are these circles and the apple shape, 00:23:08.080 |
And by sort of mathematically approaching this problem, 00:23:12.520 |
you slowly get a feeling of why things are like they are. 00:23:18.080 |
And that sort of is a first step to understanding 00:23:24.920 |
- Do you think it's possible, what's your intuition? 00:23:27.200 |
Do you think it's possible to reverse engineer 00:23:28.920 |
and find the short program that generated these fractals 00:23:37.800 |
So, I mean, in principle, what you can do is, 00:23:42.000 |
you take any data set, you take these fractals, 00:23:44.560 |
or you take whatever your data set, whatever you have, 00:23:53.200 |
You take a program of size one, two, three, four, 00:23:55.280 |
and all these programs, run them all in parallel 00:23:59.080 |
give them computational resources, first one 50%, 00:24:02.360 |
second one half resources, and so on, and let them run. 00:24:09.160 |
and if some of these programs produce the correct data, 00:24:12.360 |
then you stop, and then you have already some program. 00:24:14.480 |
It may be a long program because it's faster. 00:24:19.680 |
until you eventually find the shortest program. 00:24:25.520 |
because there could be an even shorter program, 00:24:32.160 |
But asymptotically, and actually after finite time, 00:24:36.440 |
So, this is a theoretical but completely impractical way 00:24:50.600 |
In practice, of course, we have to approach the problem 00:24:55.120 |
if you take resource limitations into account, 00:25:00.680 |
there's, for instance, the field of pseudo-random numbers, 00:25:17.840 |
I mean, random numbers, maybe not that interesting, 00:25:34.480 |
that's a big challenge for our search for simple programs 00:25:39.760 |
in the space of artificial intelligence, perhaps. 00:25:42.280 |
- Yes, it definitely is for artificial intelligence, 00:25:48.680 |
physicists worked really hard to find these theories, 00:25:52.040 |
but apparently, it was possible for human minds 00:26:02.920 |
So, let me ask another absurdly big question. 00:26:35.680 |
is that intelligence measures an agent's ability 00:26:38.880 |
to perform well in a wide range of environments. 00:26:46.000 |
and these words have been very carefully chosen, 00:26:49.760 |
and there is a mathematical theory behind that, 00:26:55.160 |
And if you look at this definition by itself, 00:27:12.840 |
which we usually associate with intelligence, 00:27:27.600 |
So, you don't have to explicitly mention that 00:27:31.000 |
So, yeah, so the consciousness, abstract reasoning, 00:27:34.200 |
all these kinds of things are just emergent phenomena 00:27:53.520 |
to achieve goals in a wide range of environments. 00:28:09.440 |
but it's much closer to something that could be formalized. 00:28:24.240 |
like how good are they at fulfilling that definition, 00:28:32.880 |
I mean, the humans are performing best among all-- 00:28:40.800 |
- Depends, you could say that trees and plants 00:28:46.840 |
- Yeah, but they are in a much more narrow environment. 00:28:49.520 |
I mean, you just have a little bit of air pollutions 00:29:04.760 |
from wide intelligence, also in the AI research. 00:29:20.680 |
but I wanna kind of hear your thoughts on it. 00:29:24.480 |
Can machines be made to fulfill this definition 00:30:01.880 |
which is able just by self-play to play chess, 00:30:26.960 |
And nothing in there, nothing was put in there. 00:30:31.080 |
the self-play mechanism, starting from scratch, 00:30:33.560 |
being able to learn, actually new strategies is-- 00:30:38.560 |
- Yeah, it rediscovered all these famous openings 00:30:47.520 |
I'm a terrible chess player, but I like Queen Gambi, 00:30:50.200 |
and AlphaZero figured out that this is the best opening. 00:31:01.680 |
yes, I believe that general intelligence is possible. 00:31:05.040 |
And it also, I mean, it depends how you define it. 00:31:14.520 |
if you achieve human level or a subhuman level, 00:31:17.440 |
but quite broad, is it also general intelligence? 00:31:25.120 |
- Is there a test in your mind, like the Turing test, 00:31:39.880 |
- Well, the Turing test, well, it has been criticized a lot, 00:31:43.000 |
but I think it's not as bad as some people think. 00:31:47.720 |
So it tests not just for a system to be intelligent, 00:31:59.000 |
And on the other hand, they say it's too weak, 00:32:01.200 |
because it just maybe fakes emotions or intelligent behavior. 00:32:09.440 |
But I don't think that's the problem, or a big problem. 00:32:14.520 |
so a conversation over terminal with a bot for an hour, 00:32:20.640 |
or maybe a day or so, and you can fool a human 00:32:23.360 |
into not knowing whether this is a human or not, 00:32:26.160 |
so that's the Turing test, I would be truly impressed. 00:32:30.280 |
And we have this annual competitions, the Leupner Prize. 00:32:38.240 |
And what is it called, the Japanese Mitsuko or so, 00:32:41.800 |
that's the winner of the last couple of years. 00:32:50.280 |
Just recently, that's an open domain conversational bot, 00:33:01.720 |
I mean, maybe it's obvious to you, it wasn't to me, 00:33:04.040 |
of setting sort of a length of a conversation. 00:33:07.760 |
Like you want the bot to be sufficiently interesting 00:33:10.960 |
that you'd want to keep talking to it for like 20 minutes. 00:33:13.680 |
And that's a surprisingly effective in aggregate metric, 00:33:19.560 |
'cause you really, like nobody has the patience 00:33:24.560 |
to be able to talk to a bot that's not interesting 00:33:29.000 |
and is able to go on to different tangents, jump domains, 00:33:36.720 |
- And maybe many humans will also fail this test. 00:33:42.840 |
just like with autonomous vehicles, with chatbots, 00:33:45.400 |
we also set a bar that's way too high to reach. 00:33:51.200 |
but what is really not useful about the Turing test, 00:33:55.960 |
it gives us no guidance how to develop these systems 00:34:00.640 |
Of course, we can develop them by trial and error 00:34:06.880 |
But a mathematical definition of intelligence 00:34:11.880 |
gives us an objective, which we can then analyze 00:34:25.120 |
And we will come back to that later with the IXE model. 00:34:40.240 |
and then you see how well it performs on the task. 00:34:55.920 |
and then they measure the compression length, 00:35:04.640 |
and then the system's performing well at the task at hand. 00:35:07.560 |
It's not perfect, but it's good enough for them 00:35:25.800 |
- So you mentioned you're one of the only people 00:35:38.680 |
to have a mathematical framework for intelligence, 00:35:53.400 |
- Okay, so let me first say what it stands for because-- 00:36:01.640 |
The first question is usually how it's pronounced, 00:36:04.400 |
but finally I put it on the website how it's pronounced, 00:36:10.560 |
- The name comes from AI, artificial intelligence, 00:36:22.040 |
which I'm not willing to repeat here in front of camera. 00:36:27.080 |
So it just happened to be more or less arbitrary, 00:36:34.720 |
So there are actions and perceptions in this model, 00:36:40.520 |
and over time, so this is A-index-I, X-index-I, 00:36:55.720 |
- So at some point, maybe five years ago or 10 years ago, 00:37:08.520 |
and the word Aix appeared there a couple of times. 00:37:17.000 |
and I looked it up, so it is Catalan language, 00:37:21.720 |
so that's it, that's the right thing to do, yeah, Eureka. 00:37:32.080 |
- And similarly, there's a Chinese word, Aixi, 00:37:37.480 |
and the final one is that it's AI crossed with induction, 00:37:41.160 |
because that is, and it's going more to the content now, 00:37:44.680 |
so good old-fashioned AI is more about planning 00:38:06.760 |
I think we can, there might be some interesting ideas there, 00:38:19.520 |
- Yeah, so it consists essentially of two parts, 00:38:22.600 |
one is the learning and induction and prediction part, 00:38:32.880 |
which essentially I explained already before, 00:38:43.520 |
I mean, if you have no idea what your actions do, 00:38:46.080 |
how can you decide which actions are good or not, 00:38:48.960 |
so you need to have some model of what your actions effect, 00:38:56.200 |
you build models like scientists of your experience, 00:38:59.440 |
then you hope these models are roughly correct, 00:39:01.440 |
and then you use these models for prediction. 00:39:05.240 |
and the model is based on your perception of the world, 00:39:16.040 |
but at this stage we can just think about predicting, 00:39:22.240 |
one, two, three, four, five, what comes next, yeah? 00:39:24.520 |
So of course our actions affect what we're doing, 00:39:32.160 |
so just to draw a line between prediction and planning, 00:39:43.640 |
without your long-term action in that environment, 00:39:48.240 |
- Okay, if you want to put the actions in now, 00:40:01.320 |
that you just have data which you passively observe, 00:40:08.960 |
weather forecasting, stock market, IQ sequences, 00:40:16.240 |
And Solomonov's theory of induction based on compression, 00:40:24.440 |
it reproduces your data sequence by definition, 00:40:30.880 |
and you can rigorously prove that for any prediction task, 00:40:35.880 |
this is essentially the best possible predictor. 00:40:46.720 |
yeah, I cannot predict the next fair coin flip, 00:40:54.040 |
Solomonov will also not magically predict it. 00:40:56.600 |
But if there is some pattern and predictability, 00:40:59.640 |
then Solomonov induction will figure that out eventually, 00:41:24.120 |
I mean, Ray Solomonov already tried it at the time. 00:41:28.120 |
- But so the basic task is you're in the environment, 00:41:35.400 |
and the model is in the space of all these programs, 00:41:38.720 |
and your goal is to get a bunch of programs that are simple. 00:41:49.680 |
but I usually sort of just jump to the decision part. 00:42:03.200 |
but you have an observation, then you act somehow, 00:42:06.720 |
and then you want to predict the next observation 00:42:09.080 |
based on the past observation and your action. 00:42:14.640 |
you don't care about predicting it because you're doing it, 00:42:21.840 |
based on your past action and observation sequence. 00:42:30.480 |
that you also try to predict your own actions. 00:42:50.160 |
- But that is a really interesting variation. 00:42:53.080 |
I don't know if you want to insert that in here, 00:42:55.440 |
but you're looking at that, in terms of observations, 00:42:59.200 |
you're looking at the entire, the big history, 00:43:07.680 |
And we can come back to that also why this is important. 00:43:13.560 |
micro-decision processes, which are much more limiting. 00:43:15.840 |
Okay, so now we can predict conditioned on actions. 00:43:21.560 |
but prediction is not all we want to do, right? 00:43:26.920 |
And the question is how to choose the actions. 00:43:29.080 |
And we don't want to greedily choose the actions, 00:43:39.920 |
So we measure performance by giving the agent reward. 00:43:43.320 |
That's the so-called reinforcement learning framework. 00:43:45.600 |
So every time step, you can give it a positive reward 00:43:51.880 |
Like if you play chess, just at the end of the game, 00:43:54.120 |
you give plus one for winning or minus one for losing. 00:43:56.880 |
So in the IXE framework, that's completely sufficient. 00:44:04.000 |
but not greedily sort of, you know, the next one, next one, 00:44:06.360 |
because that's very bad in the long run if you're greedy. 00:44:12.400 |
so let's assume the agent lives for M time steps, 00:44:14.560 |
let's just say it dies in sort of 100 years, sharp. 00:44:16.960 |
That's just, you know, the simplest model to explain. 00:44:44.400 |
under the assumption that the opponent plays the move 00:44:50.640 |
under the assumption that I play, again, the best move, 00:45:02.120 |
which von Neumann already figured out a long time ago, 00:45:14.960 |
so it can be, if there are other humans, even cooperative, 00:45:18.400 |
or nature is usually, I mean, the dead nature is stochastic. 00:45:22.680 |
Things just happen randomly, or don't care about you. 00:45:26.800 |
So what you have to take into account is the noise 00:45:30.720 |
So you replace the minimum on the opponent's side 00:45:36.000 |
which is general enough to include also adversarial cases. 00:45:49.440 |
on which probability distribution do you base that? 00:45:56.920 |
There's dice and there's certain randomness involved. 00:46:04.120 |
come up with the optimal decision if I have enough compute. 00:46:09.720 |
What is the probability the driver in front of me breaks? 00:46:19.600 |
So this is this unknown thing about prediction 00:46:24.200 |
So what you do is in sequential decision tree, 00:46:28.640 |
which we don't know, by this universal distribution. 00:46:36.760 |
and plug it into the sequential decision tree mechanism. 00:46:44.600 |
but it doesn't need to know anything about the world 00:46:56.040 |
and how Solomonov induction plays a role here? 00:47:00.760 |
- So what it does it, so in the simplest case, 00:47:07.160 |
have a prediction which would be deterministic. 00:47:09.880 |
- Okay, but you should not just take the shortest program, 00:47:33.320 |
you know, I could explain length of this program. 00:47:46.680 |
is how likely is this model given the data at hand. 00:47:56.760 |
So even if the model is simple, it gets penalized by that. 00:48:00.240 |
And what you do is then you take just the sum, 00:48:04.440 |
And this gives you a probability distribution, 00:48:10.520 |
- So it's weighed by the simplicity of the program 00:48:19.640 |
And then you said there's, you're planning N or M, 00:48:32.080 |
- So you have a planning problem up to horizon M, 00:48:34.920 |
and that's exponential time in the horizon M, 00:48:38.080 |
which is, I mean, it's computable, but intractable. 00:48:41.760 |
I mean, even for chess, it's already intractable 00:48:45.440 |
- But it could be also discounted kind of framework, or? 00:48:48.720 |
- Yeah, so having a hard horizon, you know, at 100 years, 00:48:53.000 |
it's just for simplicity of discussing the model, 00:49:09.320 |
let's let the parameter M tend to infinity, right? 00:49:12.920 |
You want an agent which lives forever, right? 00:49:22.760 |
and getting reward 0.1 every time step is infinity, 00:49:25.600 |
and giving reward one every time step is infinity, 00:49:31.120 |
Other problem is that if you have an infinite life, 00:49:35.800 |
you can be lazy for as long as you want for 10 years, 00:49:38.600 |
and then catch up with the same expected reward. 00:49:47.280 |
If they knew they lived forever, why work hard now? 00:49:51.480 |
Just enjoy your life, and then catch up later. 00:49:54.320 |
So that's another problem with infinite horizon. 00:49:56.640 |
And you mentioned, yes, we can go to discounting. 00:50:08.360 |
So if you do this so-called geometric discounting, 00:50:15.400 |
to look ahead a certain amount of time effectively. 00:50:23.880 |
there is a problem to solve which requires larger horizon. 00:50:28.120 |
So if I look ahead, you know, five time steps, 00:50:34.600 |
If I play Go, I probably have to look ahead even longer. 00:50:40.280 |
there is a problem which this horizon cannot solve. 00:50:43.800 |
But I introduced the so-called near harmonic horizon, 00:50:48.360 |
rather than exponentially T, which produces an agent 00:50:55.200 |
So if it's five years old, it plans for five years. 00:50:57.360 |
If it's 100 years old, it then plans for 100 years. 00:51:00.440 |
And it's a little bit similar to humans too, right? 00:51:04.320 |
but then we get adult, we play ahead more longer. 00:51:24.800 |
empirically, it would probably be a good idea 00:51:27.960 |
extend the horizon as you experience more of the world. 00:51:32.960 |
But is there some mathematical conclusions here 00:51:37.280 |
- With the solomon-hawking sort of prediction part, 00:51:57.720 |
- Asymptotic means you can prove, for instance, 00:51:59.960 |
that in the long run, if the agent acts long enough, 00:52:03.440 |
then it performs optimal or some nice thing happens. 00:52:19.560 |
So that is what asymptotic means, sort of eventually, 00:52:28.000 |
then I cannot prove asymptotic results, right? 00:52:35.040 |
then 100 years is over, I cannot say eventually. 00:52:42.800 |
- So just to clarify, so, okay, I've built up a model. 00:52:50.960 |
I have this way of looking several steps ahead. 00:53:12.120 |
The action which maximizes the future expected reward 00:53:26.800 |
- And then maybe you can even predict your own action. 00:53:34.120 |
I mean, it's kind of a beautiful mathematical framework 00:53:38.920 |
to think about artificial general intelligence. 00:54:10.880 |
So I switched back and forth between computer science 00:54:16.000 |
- The theory of everything, yeah, just like-- 00:54:17.400 |
- Those are basically the two biggest problems 00:54:21.440 |
- Yeah, I can explain if you wanted some later time, 00:54:32.120 |
if it was one to be solved, which one would you, 00:54:49.160 |
- Definitely AGI, because once the AGI problem is solved, 00:54:51.760 |
I can ask the AGI to solve the other problem for me. 00:55:01.160 |
- Okay, so, and the reason why I didn't settle, 00:55:04.920 |
I mean, this thought about, once you have solved AGI, 00:55:11.240 |
but all kinds of more useful problems to humanity 00:55:18.280 |
but I was quite disappointed with the state of the art 00:55:25.480 |
There was some theory about logical reasoning, 00:55:28.160 |
but I was never convinced that this will fly, 00:55:30.640 |
and then there was this more heuristic approaches 00:55:33.360 |
with neural networks, and I didn't like these heuristics, 00:55:37.520 |
so, and also I didn't have any good idea myself. 00:55:42.200 |
So, that's the reason why I toggled back and forth 00:55:44.280 |
quite some while, and even worked four and a half years 00:55:49.720 |
but then I had this idea about the AXI model, 00:55:52.840 |
and so what it gives you, it gives you a gold standard. 00:55:57.800 |
So, I have proven that this is the most intelligent agents 00:56:02.400 |
which anybody could "build" in quotation mark, 00:56:11.200 |
But this is the limit, and this is completely specified. 00:56:21.240 |
which are just skeletons, and then pieces are missing, 00:56:27.400 |
and so this is completely and uniquely defined, 00:56:33.520 |
and we have also developed some approximations. 00:56:44.280 |
that's the theoretical optimal play of games, 00:56:48.880 |
put heuristics in, prune the tree, blah, blah, blah, 00:56:51.080 |
and so on, so we can do that also with the AXI model, 00:57:00.880 |
They have their systems, they try to make it more general, 00:57:18.080 |
if my standard regularization doesn't work well, yeah? 00:57:22.560 |
We have a system which does everything that's AXI. 00:57:25.080 |
It's just, you know, completely in the ivory tower, 00:57:27.840 |
completely useless from a practical point of view, 00:57:31.960 |
ah, yeah, maybe, you know, I can take some aspects, 00:57:34.960 |
and, you know, instead of Kolmogorov complexity, 00:58:03.200 |
and at some point it was completely formalized. 00:58:12.960 |
and then they have been rigorously formalized. 00:58:15.160 |
Speed and acceleration has been confused, right, 00:58:23.240 |
who don't have any background, you know, still confuse it. 00:58:26.200 |
So, and this AXI model or the intelligence definitions, 00:58:38.920 |
- So in a sense, it serves as kind of the light 00:58:43.880 |
- So for, so, I mean, there's a million questions 00:58:54.760 |
but in general, been a lot of breakthrough ideas, 00:58:56.980 |
just like we've been saying around reinforcement learning. 00:59:10.600 |
maybe the Markov assumption is made quite often 00:59:42.040 |
on the previous observation and not the whole history, 00:59:45.280 |
which makes, of course, the mathematics much easier 01:00:04.040 |
So for instance, usually you need an ergodicity assumption 01:00:11.680 |
Ergodicity essentially means that you can recover 01:00:14.760 |
from your mistakes and that there are no traps 01:00:20.000 |
then essentially you can go back to a previous state, 01:00:24.160 |
and then learn what statistics and what the state is like. 01:00:29.960 |
And then in the long run, perform well in this state. 01:00:35.880 |
But in real life, we know there can be one single action. 01:00:39.120 |
One second of being inattentive while driving a car fast 01:00:50.360 |
So the real world is not ergodic, I always say. 01:00:56.440 |
And very little theory has been developed for this case. 01:01:01.440 |
- What about, what do you see in the context of Aixia 01:01:14.680 |
we can get into trouble when we make the wrong decisions 01:01:19.240 |
But exploration seems to be fundamentally important 01:01:22.200 |
for learning about this world, for gaining new knowledge. 01:01:29.120 |
Another way to ask it, what are the parameters 01:01:44.120 |
I mean, you can modify Aixia so that you have some knobs 01:02:08.240 |
for simple problems like so-called bandit problems, 01:02:18.000 |
say you have two medical treatments, A and B, 01:02:23.360 |
but you don't want to harm too many patients. 01:02:48.400 |
then the Bayes optimal decision with long-term planning 01:02:50.600 |
that is important, automatically implies exploration 01:03:07.720 |
although they're only asymptotic, not finite time bounds. 01:03:21.360 |
the Markov assumption and looking at the full history? 01:03:25.320 |
Sort of intuitively, of course, it's important, 01:03:41.880 |
The question is whether there's a way to deal with it 01:03:47.240 |
in a more heuristic and still sufficiently well way. 01:03:52.240 |
So I have to come up with an example on the fly, 01:04:02.080 |
you realize that's a really dangerous street or whatever, 01:04:04.680 |
right, yeah, and you want to remember that forever, right, 01:04:11.560 |
So you remember all the important events in the past, 01:04:17.480 |
- That's very hard, yeah, and I'm not concerned 01:04:27.640 |
how much data comes in through the vision system 01:04:31.800 |
and the auditory system, you compress it a little bit, 01:04:40.520 |
but you still need to do selection for the planning part 01:04:44.920 |
and the compression for the understanding part. 01:04:47.280 |
The raw storage, I'm really not concerned about, 01:04:50.000 |
and I think we should just store, if you develop an agent, 01:04:53.640 |
preferably just store all the interaction history, 01:04:59.400 |
and then you build, of course, models on top of it, 01:05:04.960 |
but occasionally, you go back to the old data 01:05:08.120 |
and reanalyze it based on your new experience you have. 01:05:13.840 |
you learn all these things you think is totally useless, 01:05:16.800 |
and much later, you realize, oh, they were not 01:05:24.080 |
Right, so maybe let me ask about objective functions, 01:05:27.720 |
because that reward, it seems to be an important part. 01:05:44.040 |
of the objective function is a key part of intelligence, 01:05:48.080 |
like the agent itself figuring out what is important. 01:06:19.200 |
So, and then, you know, I give you now a couple of answers. 01:06:22.480 |
So if we want to build agents, now let's start simple. 01:06:29.920 |
based on the IXE model, which performs a particular task. 01:06:35.800 |
like, I mean, super simple, like playing chess, 01:06:39.880 |
Then you just, you know, the reward is, you know, 01:06:47.360 |
If you have enough compute, you let itself play, 01:06:59.520 |
then you may believe that you have the right reward, 01:07:18.320 |
to how long people wait in front of the elevator. 01:07:24.240 |
and what happens is the elevator eagerly picks up 01:07:33.040 |
in the elevator also counts, so you minimize the sum, yeah? 01:07:36.800 |
And the elevator does that, but never picks up the people 01:08:10.440 |
So in this case, the human should give the reward on the fly. 01:08:15.040 |
I mean, maybe it's pre-trained in the factory 01:08:16.800 |
and that there's some sort of internal reward 01:08:23.600 |
You know, you punish the robot, it does it good. 01:08:25.280 |
You reward the robot, and then train it to a new task, 01:08:31.720 |
if you want a system which is useful to the human. 01:08:35.080 |
And as long as this agent stays sub-human level, 01:08:43.560 |
It becomes critical if they become, you know, 01:08:51.160 |
The reward technique doesn't work so well anymore. 01:08:59.160 |
which are just, you could say, slaves to the humans, yeah? 01:09:04.560 |
we want to build a new species of intelligent beings, 01:09:09.960 |
and we want them to develop this planet or whatever. 01:09:20.280 |
come up with some reward functions like, you know, 01:09:23.960 |
It should maybe multiply, build more robots, right? 01:09:28.720 |
And, you know, maybe, well, all kinds of things 01:09:32.320 |
which you find useful, but that's pretty hard, right? 01:09:38.160 |
Should it be exact copy, an approximate copy? 01:09:57.920 |
to how much the agent had learned about the world. 01:10:00.760 |
And you can rigorously, formally, uniquely define that 01:10:13.160 |
than for the general agent, which is also nice. 01:10:18.120 |
it will be, in a sense, the optimal scientist. 01:10:27.200 |
In order to learn, it needs to at least survive, right? 01:10:35.840 |
acquiring more information, it will do that, yeah? 01:10:39.160 |
If exploration, space exploration or whatever is necessary, 01:10:43.720 |
right, to gathering information and develop it. 01:10:51.040 |
And this agent is completely autonomous of us. 01:10:59.760 |
and get stuck in that library that you mentioned beforehand 01:11:10.760 |
It would get stuck in front of an old TV screen, 01:11:16.560 |
- But the second version can deal with at least stochasticity. 01:11:32.000 |
is that similar to idea of kind of injecting exploration 01:11:41.960 |
Do you find this at all appealing, interesting? 01:12:03.120 |
but for actually learning about the environment 01:12:08.520 |
So I think most curiosity is tied, in the end, 01:12:22.480 |
currently passing the Turing test quite effectively. 01:12:39.800 |
the biological reward function is to survive and to spread, 01:13:06.240 |
But the survival and spreading is, I would say, 01:13:09.720 |
the goal or the reward function of humans, the core one. 01:13:14.400 |
- I like how you avoided answering the second question, 01:13:20.920 |
- Your own meaning of life and the reward function. 01:13:36.960 |
kind of, infinity keeps creeping up everywhere. 01:13:50.040 |
in intelligent systems is that we're operating 01:13:52.000 |
always under constraints, under limited time, 01:14:06.760 |
- Yeah, that is one of the criticisms about IXE, 01:14:13.800 |
is inherently tied to what's bounded resources. 01:14:21.120 |
Do you think it's, do you think the bounded resources 01:14:31.160 |
which ignores computational limits is extremely useful. 01:14:35.520 |
A good intelligence notion which includes these resources 01:14:39.160 |
would be even more useful, but we don't have that yet. 01:14:42.140 |
And so look at other fields outside of computer science. 01:14:48.480 |
Computational aspects never play a fundamental role. 01:15:00.320 |
Well, in the end, of course, we need to do something 01:15:02.400 |
with this model, but there's more nuisance than a feature. 01:15:05.520 |
And I'm sometimes wondering if artificial intelligence 01:15:10.040 |
would not sit in a computer science department, 01:15:22.080 |
There's virtually no paper who cares about, you know, 01:15:28.320 |
Of course, once we have figured out the first problem, 01:15:31.660 |
so intelligence without computational resources, 01:15:39.360 |
could we improve it by including computational resources? 01:15:51.580 |
the right department to belong to is philosophy. 01:15:59.580 |
or even to at least to think about big picture 01:16:01.940 |
philosophical questions, big picture questions, 01:16:16.340 |
that within the IEC framework that are useful? 01:16:19.820 |
- Yeah, we have developed a couple of approximations. 01:16:23.140 |
And what we do there is that the Solomoff induction part, 01:16:28.140 |
which was, you know, find the shortest program 01:16:33.680 |
which has replaced it by standard data compressors, right? 01:16:52.260 |
So we use that for the approximation of the induction 01:17:19.100 |
So we approximate this planning part by sampling. 01:17:23.220 |
And it's successful on some small toy problems. 01:17:34.940 |
If you want to be general, you have to give up something. 01:17:41.140 |
you know, small games like coon poker and tic-tac-toe 01:17:52.060 |
The agent doesn't know the rules of the game, 01:18:22.380 |
Yeah, so the Gate-On Machine, to explain it briefly, 01:18:39.300 |
Optimally, you write the algorithm to minimax 01:18:42.140 |
so you write down what the Girdle machine should do. 01:18:50.820 |
and other part of its sources to improve this program. 01:19:02.420 |
It needs to prove by itself that this change of program 01:19:13.220 |
and by definition does the same job, but just faster, okay? 01:19:17.140 |
And then, you know, it proves over it and over it. 01:19:22.380 |
all parts of this Girdle machine can self-improve, 01:19:31.860 |
So from this perspective, it has nothing to do with IXE, 01:19:36.140 |
but if you would now put IXE as the starting axioms in, 01:19:40.580 |
it would run IXE, but, you know, that takes forever. 01:19:44.860 |
But then if it finds a provable speedup of IXE, 01:19:48.540 |
it would replace it by this, and this, and this, 01:19:51.020 |
and maybe eventually it comes up with a model 01:19:54.540 |
It cannot be, I mean, just for the knowledgeable reader, 01:20:02.900 |
therefore, there cannot be a computable exact 01:20:10.380 |
and this is not dealt with the Girdle machine, 01:20:13.260 |
But there's the IXETL model, which is finitely computable, 01:20:22.260 |
- But there is ways of getting computable approximations 01:20:30.060 |
It is still way beyond any resources anybody will ever have, 01:20:33.740 |
but then the Girdle machine could sort of improve it 01:20:41.220 |
that the Girdle machine process could improve? 01:20:51.900 |
- It is optimal in terms of the revert collected 01:20:59.420 |
but it takes infinite time to produce one action. 01:21:03.540 |
And the world continues whether you want it or not. 01:21:12.980 |
or the reaction time you need gives the answer, 01:21:19.740 |
also from learning efficiency and data efficiency, 01:21:27.580 |
but probably not provably could make it go faster. 01:21:45.540 |
in the sense you're always getting the correct answer 01:22:04.980 |
So is there a place for consciousness within IXE? 01:22:20.900 |
- You say humans are an instantiation of an IXE agent? 01:22:25.260 |
but I think that's not really for the smartest 01:22:29.000 |
I think maybe we are very crude approximations. 01:22:33.760 |
I mean, I tend to believe, again, I'm Russian, 01:22:35.740 |
so I tend to believe our flaws are part of the optimal. 01:22:40.740 |
So we tend to laugh off and criticize our flaws 01:22:45.660 |
and I tend to think that that's actually close 01:22:50.700 |
But some flaws, if you think more carefully about it, 01:23:00.860 |
As a student of history, I think all the suffering 01:23:06.820 |
it's possible that that's the optimal amount of suffering 01:23:10.240 |
we need to endure to minimize long-term suffering. 01:23:18.140 |
whether we humans are or not instantiations of an AIC agent, 01:23:21.900 |
do you think consciousness is something that could emerge 01:23:25.700 |
in a computational form of framework like AIC? 01:23:47.180 |
- If an agent were to solve the imitation game 01:23:49.780 |
posed by Turing, I think they would be dressed 01:23:51.740 |
similarly to you, because there's a kind of flamboyant, 01:24:01.060 |
that sells that you're human and you're conscious. 01:24:26.780 |
if it behaves in a way sufficiently close to humans, 01:24:39.940 |
typically consciousness, we would attribute consciousness 01:24:50.860 |
And that's the big, hard problem of consciousness. 01:24:55.700 |
I mean, not the movie zombie, but the philosophical zombie. 01:25:02.620 |
close enough to consciousness from a perspective of AGI 01:25:06.740 |
that the distinction of the hard problem of consciousness 01:25:13.980 |
especially the hard problem for developing AGI. 01:25:39.900 |
or is it a zombie which just, you know, fakes everything? 01:25:49.900 |
but it may also be sort of practically interesting. 01:25:54.300 |
if it's just faking consciousness and feelings, 01:25:56.260 |
you know, then we don't need to be concerned about rights. 01:26:22.660 |
I mean, remember, maybe it was before you were born, 01:27:00.380 |
- Probably would have said yes, I would guess. 01:27:01.660 |
- I think that's kind of a beautiful thing, actually, 01:27:04.780 |
'cause that consciousness, ascribing consciousness, 01:27:12.660 |
but we have to be careful on the ethics side of that. 01:27:15.940 |
Well, let me ask about the AGI community broadly. 01:27:18.500 |
You kind of represent some of the most serious work on AGI, 01:27:24.260 |
and DeepMind represents serious work on AGI these days. 01:27:29.260 |
But why, in your sense, is the AGI community so small, 01:27:34.100 |
or has been so small, until maybe DeepMind came along? 01:27:38.100 |
Like, why aren't more people seriously working 01:27:41.660 |
on human-level and superhuman-level intelligence 01:27:58.500 |
and then there were big promises which were not fulfilled. 01:28:05.740 |
and that narrow AI, solving particular problems 01:28:19.500 |
And if you build something which is, you know, 01:28:28.540 |
which drives people to develop specific systems, 01:28:38.780 |
you should be able to do ivory tower research. 01:28:42.860 |
And that was probably better a long time ago, 01:28:46.020 |
but even nowadays, there's quite some pressure 01:28:48.340 |
of doing applied research or translational research, 01:28:52.260 |
and, you know, it's harder to get grants as a theorist. 01:29:03.180 |
So I think enough people, I mean, maybe a small number, 01:29:05.940 |
were still interested in formalizing intelligence 01:29:27.620 |
what do you think it takes to build an AGI system? 01:29:30.380 |
Is it, and I don't know if that's a stupid question 01:29:35.100 |
from everything we've been talking about at IAXE, 01:29:46.340 |
- That's the whole point of this conversation, 01:29:49.780 |
Now, is there, I mean, what's your intuition? 01:29:58.940 |
like the efforts with AlphaZero and AlphaStar 01:30:02.860 |
how you can solve it through in the simulation, 01:30:08.220 |
all the transformer work in natural language processing, 01:30:13.220 |
sort of maybe attacking the open domain dialogue? 01:30:15.780 |
Like what, where do you see the promising pathways? 01:30:29.540 |
I don't believe that we need a physical robot 01:30:40.260 |
interacting with the real world in order to achieve AGI. 01:30:51.380 |
It's sort of confusing the body with the mind. 01:30:54.580 |
For industrial applications or near-term applications, 01:30:58.940 |
of course we need robots for all kinds of things, 01:31:04.140 |
at least at this stage, I think it's not necessary. 01:31:15.300 |
and that can be a virtual agent in a computer 01:31:28.900 |
Even if you don't intend to later put it sort of, 01:32:01.620 |
on top of each other in space and flying in cars and so on, 01:32:19.580 |
and becomes a better and better mathematician, 01:32:53.900 |
'cause maybe people could also read those books 01:33:00.180 |
- Yeah, luckily I asked books and not singular book. 01:33:03.580 |
It's very hard and I tried to pin down one book. 01:33:10.620 |
So the most, the books which were most transformative 01:33:25.500 |
I would always start with Russell and Norbeck, 01:33:59.740 |
the Reinforcement Learning Book by Sutton and Bartow. 01:34:07.940 |
it makes RL feel and look much easier than it actually is. 01:34:17.620 |
You can very quickly get some RL systems to run, 01:34:20.380 |
you know, on very toy problems, but it's a lot of fun. 01:34:23.300 |
And in a couple of days, you feel you know what RL is about, 01:34:36.260 |
And maybe, I mean, there's so many books out there. 01:34:41.260 |
If you like the information theoretic approach, 01:34:43.460 |
then there's "Colmogorov Complexity" by Leon Bitani, 01:34:46.780 |
but probably, you know, some short article is enough. 01:34:54.420 |
And if you have to mention one all-time favorite book, 01:34:59.420 |
so different flavor, that's a book which is used 01:35:05.220 |
for high school students in several countries. 01:35:09.020 |
That's from Nicholas Alchen, "Theory of Knowledge." 01:35:12.940 |
Second edition or first, not the third, please. 01:35:25.620 |
or to me, interesting philosophical questions 01:35:27.620 |
about how we acquire knowledge from all perspectives, 01:35:36.220 |
And the book is called "Theory of Knowledge." 01:35:38.020 |
- From which, it's almost like a philosophical exploration 01:35:43.140 |
- Yes, yeah, I mean, can religion tell us, you know, 01:35:46.180 |
Can science tell us something about the world? 01:35:48.060 |
Can mathematics, or is it just playing with symbols? 01:35:56.220 |
so they have the resources from "Hitchhiker's Guide 01:36:01.780 |
And it's fun to read, but it's also quite deep. 01:36:07.580 |
- If you could live one day of your life over again, 01:36:33.340 |
- You're really gonna say from the future, I love it. 01:36:53.100 |
but I discovered sort of this compression idea myself, 01:36:56.220 |
but immediately I knew I can't be the first one, 01:36:59.220 |
and then I knew about sequential decision tree, 01:37:02.100 |
and I knew if I put it together, this is the right thing. 01:37:06.260 |
And yeah, still when I think back about this moment, 01:37:12.300 |
- Was there any more details and context that moment? 01:37:17.900 |
So like if you look at Ian Goodfellow talking about GANs, 01:37:25.820 |
Is there some more context of what sparked your thought, 01:37:37.500 |
And I worked on an image interpolation problem, 01:37:43.620 |
and I developed a quite neat new interpolation techniques, 01:37:54.260 |
yeah, that's pretty good, but it's not the best, 01:37:56.100 |
so what is the best possible way of doing interpolation? 01:37:59.660 |
And then I thought, yeah, you want the simplest picture, 01:38:06.420 |
and then I thought about the simplicity concept 01:38:34.060 |
- So what about if you could really one day in the future, 01:38:41.540 |
- I don't think-- - In practice, in practice, 01:38:44.580 |
so in theory I have solved it with the Ix-A model, 01:38:54.740 |
- I don't think there's a better way to end it. 01:39:01.420 |
- Yeah, thank you too, it was a pleasure of mine, too. 01:39:07.420 |
and thank you to our presenting sponsor, Cash App. 01:39:15.500 |
an organization that inspires and educates young minds 01:39:18.560 |
to become science and technology innovators of tomorrow. 01:39:22.340 |
If you enjoy this podcast, subscribe on YouTube, 01:39:27.020 |
support on Patreon, or simply connect with me on Twitter 01:39:32.040 |
And now, let me leave you with some words of wisdom 01:39:37.900 |
"The measure of intelligence is the ability to change." 01:39:42.120 |
Thank you for listening, and hope to see you next time.