back to indexGarry Kasparov: IBM Deep Blue, AlphaZero, and the Limits of AI in Open Systems | AI Podcast Clips
Chapters
0:0 Intro
1:0 Losing was painful
2:0 Analyzing the match
3:0 Closed systems
4:0 Chess standards
5:0 Deep Blue
6:0 Humans vs Machines
7:0 Chess vs AlphaZero
00:00:05.640 |
In my eyes, that is one of the most seminal moments in the history. 00:00:09.040 |
Again, I apologize for being romanticizing the notion, but in the history of our civilization, 00:00:15.480 |
because humans as a civilization for centuries saw chess as the peak of what man can accomplish, 00:00:26.840 |
And that moment when a machine could beat a human being was inspiring to just an entire 00:00:34.840 |
anyone who cares about science, innovation, an entire generation of AI researchers. 00:00:41.080 |
And yet, to you that loss, at least if reading your face, seemed like a tragedy, extremely 00:00:51.400 |
When you look back at your psychology of that loss, why was it so painful? 00:00:56.040 |
Were you not able to see the seminal nature of that moment? 00:01:05.960 |
As I already said, losing was painful, physically painful. 00:01:12.040 |
And the match I lost in 1997 was not the first match I lost to a machine. 00:01:33.200 |
Now I lost, and the reason I was so angry that I just, I had suspicions that my loss 00:01:44.480 |
So though I played quite poorly, just when you started looking at the games today, I 00:01:49.760 |
But I had all reasons to believe that there were other factors that had nothing to do 00:02:02.000 |
We can analyze this match and this is with everything you said. 00:02:05.320 |
I agree with probably one exception, is that considering chess as the sort of, as a pinnacle 00:02:16.320 |
Because we just thought, "Oh, it's a game of the highest intellect and it's just you 00:02:21.080 |
have to be so intelligent and you could see things that the ordinary mortals could not 00:02:32.280 |
And all machines had to do in this game is just to make fewer mistakes, not to solve 00:02:40.120 |
I mean, according to Koval Shanin, the number of legal moves is 10 to the 46th power. 00:02:44.800 |
Too many zeros, just for any computer to finish the job in next few billion years. 00:02:59.680 |
And what's happened afterwards with other games, with Go, with Shogi, with video games, 00:03:07.480 |
it's a demonstration that the machines will always beat humans in what I call closed systems. 00:03:14.200 |
The moment you build a closed system, no matter how the system is called, chess, Go, Shogi, 00:03:22.520 |
Dota, machines will prevail simply because they will bring down number of mistakes. 00:03:34.040 |
The way they outplay us, it's not by just being more intelligent. 00:03:38.240 |
It's just by doing something else, but eventually it's capitalizing on our mistakes. 00:03:44.600 |
When you look at the chess machines ratings today, and compare this to Magnus Carlsen, 00:03:50.000 |
is the same as comparing Ferrari to Usain Bolt. 00:03:54.960 |
The gap is, I mean, by chess standards is insane. 00:04:03.520 |
It's like difference between Magnus and an ordinary player from an open international 00:04:10.800 |
It's not because machine understanding is better than Magnus Carlsen, but simply because 00:04:18.760 |
And I think that is what we have to learn from 1997 experience and from further encounters 00:04:27.220 |
with computers and sort of the current state of affairs was AlphaZero, you were beating 00:04:34.860 |
The idea that we can compete with computers in so-called intellectual fields, it was wrong 00:04:44.960 |
By the way, the 1997 match was not the first victory of machines over... 00:04:53.600 |
And I played against first decent chess computers from late '80s. 00:04:58.960 |
So I played with the prototype of Deep Blue called Deep Thought in 1989, two rapid chess 00:05:07.940 |
We played against new chess engines like Fritz and other programs. 00:05:13.720 |
And then it was Israeli program Junior that appeared in 1995. 00:05:22.920 |
I lost one match against the computer chess engine in 1994, rapid chess. 00:05:27.920 |
So I lost one game to Deep Blue in 1996 match, the match I won. 00:05:33.040 |
Some people tend to forget about it that I won the first match. 00:05:37.120 |
But we made a very important psychological mistake thinking that the reason we lost Blitz 00:05:45.000 |
matches, five minutes games, the reason we lost some of the rapid chess matches, 25 minutes 00:05:53.000 |
If you play a longer match, we will not make the same mistakes. 00:05:57.760 |
So yeah, we had more time, but we still make mistakes. 00:06:02.420 |
And machine will always be steady and consistent compared to humans' instabilities and inconsistencies. 00:06:13.040 |
And today we are at the point where nobody talks about humans playing against machines. 00:06:19.640 |
Humans can offer handicap to top players, still will be favored. 00:06:25.840 |
I think we're just learning that it's no longer human versus machines. 00:06:32.720 |
That's what I recognized in 1998, just after leaking my wounds and spending one year and 00:06:38.720 |
just ruminating so what's happened in this match. 00:06:43.180 |
And I knew that we still could play against the machines. 00:06:46.440 |
I had two more matches in 2003 playing both deep free and deep junior. 00:06:54.400 |
Though these machines were not weaker, at least probably stronger than deep blue. 00:06:59.760 |
And by the way, today, chess app on your mobile phone is probably stronger than deep blue. 00:07:05.160 |
I'm not speaking about chess engines that are so much superior. 00:07:08.720 |
And by the way, when you analyze games we played against deep blue in 1997 on your chess 00:07:15.400 |
And it also shows us how chess changed because chess commentators, they'll look at some of 00:07:20.880 |
our games like game four, game five, brilliant idea. 00:07:24.120 |
Now you ask Stockfish, you ask Houdini, you ask Commodore, all the leading chess engines. 00:07:31.880 |
Within 30 seconds, they will show you how many mistakes both Gary and deep blue made 00:07:37.160 |
in the game that was trumpeted as a great chess match in 1997. 00:07:45.400 |
- Well, okay, so you've made an interesting, if you can untangle that comment. 00:07:51.200 |
So now in retrospect, it was a mistake to see chess as the peak of human intellect. 00:08:01.400 |
So in Europe, because you move to the Far East, they will go, they're showing you- 00:08:16.760 |
- So if I push back a little bit, so now you say that, okay, but it was a mistake to see 00:08:25.320 |
And then now there's other things maybe like language, like conversation, like some of 00:08:30.000 |
the things that in your view is still way out of reach of computers, but inside humans. 00:08:35.560 |
Do you think, can you talk about what those things might be? 00:08:39.160 |
And do you think just like chess that might fall soon with the same set of approaches, 00:08:45.720 |
if you look at alpha zero, the same kind of learning approaches as the machines grow in 00:08:54.120 |
It's about, again, it's about understanding the difference between closed system and open 00:08:59.680 |
- So you think that key difference, so the board games are closed in terms of the rules 00:09:05.440 |
that they actions, the state space, everything is just constrained. 00:09:10.760 |
You think once you open it, the machines are lost? 00:09:14.760 |
- Not lost, but again, the effectiveness is very different because machine does not understand 00:09:19.800 |
the moment it's reaching territory of diminishing returns. 00:09:23.400 |
It's the, to put it in a different way, machine doesn't know how to ask right questions. 00:09:30.640 |
It can ask questions, but it will never tell you which questions are relevant. 00:09:37.120 |
So I think it's in human machine relations, we have to consider so our role and many people 00:09:43.400 |
feel uncomfortable that the territory that belongs to us is shrinking. 00:09:50.440 |
I'm saying so what, this is eventually will belong to the last few decimal points, but 00:09:55.800 |
it's like having so very powerful gun and all you can do there is slightly alter direction 00:10:05.640 |
of the bullet, maybe 0.1 degree of this angle, but that means a mile away, 10 meters of target. 00:10:16.440 |
So that's, we have to recognize that is a certain unique human qualities that machines 00:10:22.680 |
in a foreseeable future will not be able to reproduce. 00:10:28.400 |
And the effectiveness of this cooperation, collaboration depends on our understanding 00:10:35.160 |
So the greatest danger is when we try to interfere with machine superior knowledge. 00:10:40.320 |
So that's why I always say that sometimes you'd rather have, by reading this picture 00:10:44.680 |
is in radiology, you may probably prefer an experienced nurse than rather than having 00:10:51.280 |
top professor, because she will not try to interfere with machines understanding. 00:10:56.880 |
So it's very important to know that if machines knows how to do better things in 95%, 96% 00:11:02.720 |
of territory, we should not touch it because it's happened. 00:11:06.160 |
It's like in chess, recognize, they do it better. 00:11:12.120 |
You mentioned AlphaZero, I mean, AlphaZero, it's actually a first step into what you may 00:11:18.160 |
call AI, because everything that's being called AI today, it's one or another variation of 00:11:25.840 |
what Claude Shannon characterized as a brute force, is a type A machine. 00:11:30.440 |
Whether it's Deep Blue, whether it's Watson, and all these things, the modern technologies 00:11:36.040 |
that are being trumpeted as AI, it's still brute force. 00:11:40.120 |
It's the, all they do, it's they do optimization. 00:11:43.560 |
It's this, they are, they keep improving the way to process human generated data. 00:11:51.640 |
Now AlphaZero is the first step towards machine produced knowledge, which is by the way, it's 00:12:00.360 |
quite ironic that the first company that championed that was IBM. 00:12:08.640 |
Yes, you should look at IBM, it's a new gammon, it's the scientist, he's still working at 00:12:20.720 |
It's the program that played in all the AlphaZero types, so just trying to come up with own 00:12:27.000 |
But because of success of Deep Blue, this project had been not abandoned, but just it 00:12:34.760 |
And now it's, everybody talks about the machines generated knowledge, so as revolutionary, 00:12:43.240 |
and it is, but there's still many open-ended questions. 00:12:53.040 |
Many ideas that AlphaZero generated in chess were quite intriguing. 00:12:56.920 |
So I looked at these games with, not just with interest, but it was quite exciting to 00:13:04.720 |
learn how machine could actually juggle all the pieces and just play positions with a 00:13:10.840 |
broken material balance, sacrificing material, always being ahead of other programs, one 00:13:15.640 |
or two moves ahead by foreseeing the consequences, not over calculating, because other machines 00:13:22.480 |
were at least as powerful in calculating, but it's having this unique knowledge based 00:13:37.040 |
Now, the simple question, if AlphaZero faces superior point, let's say another powerful 00:13:44.360 |
computer accompanied by a human who could help just to discover certain problems, because 00:13:51.000 |
I already, I looked at many AlphaZero games, I visited their lab, spoke to Demis Kasabis 00:13:55.920 |
and his team, and I know there's certain weaknesses there. 00:13:58.960 |
Now if these weaknesses are exposed, then the question is, how many games will it take 00:14:06.400 |
Even if it keeps losing, it's just because the whole system is based. 00:14:11.160 |
So it's now, imagine, so you can have a human by just making a few tweaks. 00:14:16.440 |
So humans are still more flexible, and as long as we recognize what is our role, where 00:14:23.200 |
we can play sort of the most valuable part in this collaboration, so it will help us 00:14:30.600 |
to understand what are the next steps in human-machine collaboration.