back to indexEscaping the Local Optimum of Low Expectation
Chapters
0:0 Overview - The Voice poem
6:46 Artificial intelligence
13:44 Open problems in AI
14:10 Problem 1: Learning to understand
17:15 Problem 2: Learning to act
19:28 Problem 3: Reasoning
20:44 Problem 4: Connection between humans & AI systems
23:57 Advice about life as an optimization problem
24:10 Advice 1: Listen to your inner voice - ignore the gradient
25:12 Advice 2: carve your own path
26:28 Advice 2: Measure passion not progress
28:10 Advice 4: work hard
29:5 Advice 5: forever oscillate between gratitude and dissatisfaction
31:10 Q&A: Meaning of life
33:11 Q&A: Simulation hypothesis
36:15 Q&A: How do you define greatness?
00:00:00.000 |
It's wonderful to be here, wonderful to see so many faces 00:00:08.880 |
My advisor, my family's here, my mom, brother. 00:00:13.880 |
You know, I did ask security to make sure my dad doesn't, 00:00:20.200 |
is not allowed in, but he somehow found his way in, 00:00:46.840 |
different Jew, similar haircut for those of you. 00:00:53.280 |
You know, there's a saying, there's an old saying 00:00:58.040 |
that goes, "Give a man a fish and you feed him for a day, 00:01:02.120 |
"teach a man to fish and you feed him for a lifetime." 00:01:05.040 |
A little known fact, it actually goes on to say, 00:01:09.800 |
"So that he may never discover how much he loves steak." 00:01:18.680 |
And the key there, the key idea is society tries to, 00:01:24.280 |
impose lessons to teach, to drive the human being, 00:01:29.280 |
each of us, but it's you discovering your own passion 00:01:39.000 |
And the central poem by Shel Silverstein called "The Voice" 00:01:43.560 |
is one I think that will resonate throughout the talk. 00:01:50.040 |
it's a voice that's in the air, it's a voice that's in the air. 00:01:55.040 |
It's a voice that's in the air, it's a voice that's in the air. 00:02:00.040 |
It's a voice that's in the air, it's a voice that's in the air. 00:02:05.200 |
It's a voice that's in the air, it's a voice that's in the air. 00:02:10.280 |
It's a voice that's in the air, it's a voice that's in the air. 00:02:15.760 |
And that's the poem we have together over two small topics, 00:02:32.880 |
when you show a plot, you have to describe the X axis 00:02:41.040 |
The X axis is competence, the Y axis is confidence. 00:02:45.520 |
And there's something called the Dunning-Kruger effect, 00:02:53.320 |
And that is at the beginning of your journey of competence, 00:03:01.000 |
of learning something, as some of you here are 00:03:03.440 |
in the engineering fields, you're overly confident. 00:03:14.280 |
And it's funny that I'm speaking here before you today 00:03:19.280 |
in a place of a complete sort of self-doubt and despair 00:03:30.680 |
And I feel like I have zero expertise to impart on you. 00:03:38.120 |
to be speaking with, especially some of the lessons, 00:03:44.720 |
And some of you sitting in the audience today 00:03:48.600 |
may be at the very peak, especially if you're 00:03:50.720 |
at the beginning of the college journey, university journey. 00:03:57.840 |
the biggest impact of college and university education 00:04:01.580 |
is the dismantling of the ego that's involved 00:04:10.680 |
to the valley of despair that I'm currently in. 00:04:13.220 |
Oh, and I should mention that this is also the time for me 00:04:17.880 |
and perhaps for you where folks like Dostoevsky 00:04:31.400 |
I think as he says, have great sadness on earth. 00:04:52.120 |
and the valley of despair, there's a gradient 00:04:55.320 |
provided to you by your advisors, by your parents, 00:04:58.720 |
by your friends, your loved ones, society in general. 00:05:12.360 |
What everybody else tells you you're supposed to do. 00:05:15.020 |
What everybody else at the small scale, on a daily scale, 00:05:32.000 |
of how to listen just enough to hear the lessons 00:05:50.420 |
So I was introduced as a research scientist at MIT. 00:05:53.180 |
And very recently, I decided to step down from MIT 00:06:00.880 |
but sort of give up the salary, give up everything, 00:06:06.840 |
the definition under academic colleagues of what success is, 00:06:26.680 |
but perhaps it's interesting to speak from this position, 00:06:30.240 |
'cause I would argue it's the most beautiful position 00:06:34.640 |
The opportunity, the freedom in the struggles 00:06:43.100 |
that comes at the end of this journey of college. 00:06:49.400 |
And what is the dream that I mentioned there at the end? 00:06:58.080 |
and engineering artificial intelligence systems. 00:07:01.520 |
Visualized on the left here is just 3% of the neurons 00:07:10.020 |
It's easy to forget how little we know about this mystery 00:07:26.960 |
this mysterious, beautiful thing that brings to life 00:07:31.820 |
And the dream of creating intelligence systems, 00:07:36.800 |
companions, ones that you can have a deep connection with. 00:07:49.720 |
has been on robotics and autonomous vehicles. 00:07:52.260 |
But now the dream is to create a system that you can love 00:08:02.540 |
to give you a sense, to give you a quick review 00:08:17.180 |
So it started on the theoretical end with Alan Turing 00:08:21.020 |
and many of the ideas from philosophy to mathematics 00:08:24.140 |
that he presented and from whom the field was born. 00:08:27.680 |
And on the engineering side, Frank Crosonblatt 00:08:31.620 |
in building the Perceptron, the first machine. 00:08:34.220 |
So engineering machines that can do some aspect of learning, 00:08:41.020 |
And then there's been accomplishments throughout, 00:08:49.920 |
There's been two branches of artificial intelligence 00:08:54.740 |
The early days have been, you can think of a search, 00:08:59.740 |
It's not quite as captivating to our imagination. 00:09:05.780 |
because it's brute force searching through possible answers 00:09:11.460 |
It's converting every single problem to a search problem 00:09:24.040 |
is when IBM D-Blue defeated Garry Kasparov in 1997. 00:09:29.040 |
This is a seminal moment in artificial intelligence 00:09:31.940 |
where the game that was associated with thought, 00:09:34.760 |
with intelligence, with reason, was overcome, 00:09:47.160 |
of artificial intelligence, which is learning systems, 00:09:55.640 |
was able to defeat the greatest player in the world. 00:10:01.560 |
Little side note, the first moment did have human assistance 00:10:06.200 |
in the AlphaGo system from DeepMind and Google DeepMind. 00:10:11.900 |
the system called AlphaZero was able to learn from scratch 00:10:27.140 |
like I said, I worked a lot in autonomous vehicles. 00:10:29.200 |
This is one of the most exciting applications 00:10:31.240 |
with autonomous and semi-autonomous vehicles. 00:10:33.600 |
There's been deployments, lessons, explorations, 00:10:39.840 |
This is the most exciting space of artificial intelligence. 00:10:44.720 |
autonomous vehicles is the space you will do so 00:10:54.920 |
in artificial intelligence that were key breakthroughs. 00:10:58.040 |
So neural networks and Perceptron, like I said, 00:11:03.160 |
With the algorithms that dominate today's world 00:11:07.120 |
have been invented in many, many decades ago, 00:11:13.080 |
with convolutional for the computer vision aspect 00:11:23.400 |
they work with language, work with sequence of data, 00:11:26.440 |
were developed in the '90s and proven out in the aughts. 00:11:31.000 |
And then the deep learning quote unquote revolution, 00:11:33.640 |
the term and the ideas of large-scale machine learning 00:11:43.160 |
and then proven out in the seminal ImageNet moment 00:11:54.040 |
image recognition, and the ImageNet data set, 00:11:58.560 |
neural networks were able to far outperform the competition 00:12:03.560 |
and do so easily from just learning from data. 00:12:14.200 |
that were born in the '14, '15, '16, just a few years ago, 00:12:18.040 |
and a lot of exciting ideas in the past few years. 00:12:31.480 |
Transformers in particular, with natural languages, 00:12:34.880 |
some of the most beautiful and exciting ideas 00:12:39.000 |
you can learn to model language sufficiently well 00:12:42.920 |
to outperform anything we've done previously, 00:12:52.840 |
And especially exciting is that bigger is better, 00:12:56.760 |
meaning that as long as we can scale compute, 00:13:11.240 |
there's a concept of Big Bang for the start of the universe, 00:13:16.320 |
a silly name for one of the most incredible mysteries 00:13:22.320 |
Same way, self-play is one of the silliest names 00:13:32.960 |
to improve continuously without any human supervision. 00:13:41.480 |
and I recommend if you love learning that you explore. 00:13:44.880 |
So the open problems in artificial intelligence 00:13:50.920 |
And one of the things, and I'll focus on number four, 00:14:04.040 |
Learning to understand, learning to act, reason, 00:14:08.360 |
and a deep connection between humans and AI systems. 00:14:13.040 |
there's a lot of exciting possibilities here. 00:14:15.480 |
This is a lot of the breakthroughs in machine learning 00:14:18.320 |
have been in something called supervised learning, 00:14:21.360 |
where you have a set of data and you have a neural network 00:14:24.640 |
or a model that's able to learn from that data 00:14:41.920 |
you could recognize other vehicles, pedestrians, 00:14:49.360 |
Now that's all good, but to solve real world problems, 00:15:01.760 |
that our ability to do reasoning and common sense reasoning 00:15:06.520 |
So to be able to learn over those edge cases, 00:15:10.920 |
And for that, you have to be much more selective 00:15:13.600 |
and clever about which data you annotate with human beings. 00:15:18.680 |
Same way with, as children, we explore the world, 00:15:22.600 |
we interact with the world to pick up the lessons from it. 00:15:28.560 |
to select only small parts of it to learn from. 00:15:34.000 |
that's using autonomous driving and its system autopilot 00:15:49.120 |
They're creating a pipeline for each individual task. 00:15:51.880 |
They take the task of driving and break it apart 00:15:59.200 |
Each subtask gets its own pipeline, its own dataset. 00:16:12.000 |
And when the vehicle fails in a particular case, 00:16:15.440 |
that's an edge case that's marked for the system 00:16:19.400 |
and is brought back to the pipeline to annotate. 00:16:22.800 |
So there's ongoing pipeline that continuously goes on. 00:16:27.320 |
The system is not very good in the beginning, 00:16:29.280 |
but the whole purpose of it is to discover edge cases. 00:16:32.920 |
In the same way that us humans learn something, 00:16:37.160 |
and you can think of our actually existence in the world 00:16:55.200 |
And we do that thousands of times a day still, 00:17:11.400 |
in terms of scale impact area in the next few years. 00:17:17.880 |
the second set of open problems in artificial intelligence. 00:17:20.920 |
This is where the idea of self-play comes in, 00:17:27.560 |
whether through a reinforcement learning mechanism 00:17:29.720 |
or otherwise, that are actually acting in the world. 00:17:36.000 |
the idea is that you have a really dumb system 00:17:46.280 |
you have other systems that also know nothing, 00:17:52.360 |
So you formulate the problem as a competitive setting. 00:18:02.440 |
The one that's slightly less dumb starts winning. 00:18:26.280 |
It can all be done in computation in a distributed sense. 00:18:34.160 |
create a system that beats the world champion at go. 00:18:39.840 |
that have they've defeated the world champion in chess, 00:18:44.440 |
is the best chess playing program, Stockfish, 00:18:54.680 |
This is both the exciting and the scary thing 00:19:02.840 |
What we hit is the limits of our computational power, 00:19:08.840 |
especially the kind of mechanisms that are happening now, 00:19:16.440 |
So computation, if you just wait a few years, 00:19:19.480 |
So we were yet to see the ceiling of the capabilities 00:19:40.800 |
dog intelligence system solving a particular problem. 00:19:48.120 |
So we know nothing how to do reasoning systems 00:19:54.840 |
This is the actually not very often talked about area 00:20:01.760 |
There's been subsets called program synthesis, 00:20:06.120 |
communities that kind of try to formulate a subset 00:20:10.120 |
of the reasoning problem and try to solve it, 00:20:18.480 |
to be able to reason about the physics of the world, 00:20:21.160 |
about the basic, especially with human beings, 00:20:24.120 |
human to human, human to physical world dynamics. 00:20:38.480 |
This process is a really exciting area of research 00:20:47.360 |
don't really have anything to do with humans necessarily. 00:21:18.760 |
the recommendation engines behind social networks, 00:21:24.280 |
and about the content of your friends that you see, 00:21:33.280 |
The personalization of IOT, of smart systems, 00:21:47.000 |
Whenever you have AI systems between you and a machine. 00:21:55.000 |
There's you human that are tasked with sitting there 00:21:59.960 |
And there is an AI system in the middle that manages that. 00:22:03.680 |
It manages the tension, the dance, the uncertainty, 00:22:09.360 |
all the mess of human beings, it manages that. 00:22:18.520 |
What I show there is where my sense is, where we stand. 00:22:28.720 |
Some of you may even be old enough to have used them. 00:22:31.720 |
AltaVista, Excite, AskG is like us and so on. 00:22:35.360 |
Then Google came along, the Google search engine 00:22:46.960 |
and the fundamental ideas behind their approach was flawed. 00:23:03.200 |
Many people have in their home an Alexa device, 00:23:08.280 |
But most people don't use it for almost anything 00:23:18.720 |
but artificial intelligence plays a minimal role 00:23:24.360 |
in recommending how you interact with the platform 00:23:41.860 |
except whether your hands are on the steering wheel or not. 00:23:50.000 |
how much opportunity there is to learn about human beings 00:24:08.760 |
So in this context, in this optimization context, 00:24:14.800 |
my first piece of advice is to listen to your inner voice. 00:24:20.880 |
including a lot of very smart professors, advisors, 00:24:29.200 |
have in them a kind of mutually agreed upon gradient 00:24:36.060 |
It's so difficult for me to articulate this in a clear way. 00:24:46.960 |
a silly sounding, crazy voice that told me to do things. 00:24:52.340 |
One of which was to try to put a robot in every home. 00:24:57.680 |
There's dreams that are difficult for me to articulate. 00:25:00.840 |
But if you allow your mind to be quiet enough, 00:25:03.840 |
you'll hear such voices, you'll hear such dreams. 00:25:07.480 |
And it's important to really listen and to pursue them. 00:25:17.840 |
And if that means taking a few detours, take the detours. 00:25:23.880 |
Again, this is coming from the valley of despair. 00:25:39.400 |
I gave a lot of myself to the practice of martial arts. 00:25:48.440 |
And both music and martial arts have given me, 00:25:55.140 |
but it have given me something quite profound. 00:25:59.060 |
It gave flavor and color to the pursuit of that dream 00:26:07.280 |
listened to my heart in pursuing these detours. 00:26:09.600 |
From poetry to excessive reading, like I mentioned, 00:26:21.020 |
that seemingly have nothing to do with the main pursuit. 00:26:24.200 |
And starting the silliest of pursuits, starting a podcast. 00:26:29.200 |
Advice number three is to measure passion, not progress. 00:26:35.780 |
So most of us get an average of about 27,000 days of life. 00:26:42.280 |
I think a good metric by which you should live 00:26:49.800 |
that are filled with a passionate pursuit of something. 00:26:58.020 |
Because goals are grounded in your comparison 00:27:01.640 |
to something that's already been done before. 00:27:08.320 |
is the way you achieve something totally new. 00:27:51.720 |
Advisors, colleagues will tell you to be pragmatic 00:27:57.120 |
from the main effort that you should be focusing on. 00:28:38.320 |
who has not truly worked hard for thousands of hours 00:28:47.640 |
you first have to put in those few tens of thousands 00:29:02.000 |
And to do that is you have to love what you do. 00:29:13.840 |
and appreciate every single moment you're alive. 00:29:28.200 |
between deep, profound doubt and self-dissatisfaction 00:29:38.960 |
for all the people around you that give you their love, 00:29:49.680 |
In the desert, I saw a creature, a naked bestial, 00:30:11.040 |
So I would say the bitter is the self-dissatisfaction, 00:30:15.360 |
and that's the restless energy that drives us forward. 00:30:47.960 |
I'd like to continue on the gratitude and say thank you. 00:31:03.400 |
and all the friends that I've had along the way. 00:31:12.720 |
I've never been introduced with this much energy. 00:31:18.260 |
- You're hanging out at the wrong places, man. 00:31:22.260 |
- First of all, great to see you in person, Dr. Kristine. 00:32:03.740 |
there's something in you that makes life beautiful. 00:32:29.940 |
And see that the work you're doing is part of that. 00:32:40.260 |
there's something in me that's deeply fulfilling 00:32:50.180 |
that we human beings can create intelligent systems. 00:32:57.900 |
And to me, that makes it somehow deeply exciting, 00:33:10.620 |
There's something within that that's so exciting. 00:33:16.860 |
and he basically said that we're in a civilization, 00:33:30.460 |
and scientists are uncomfortable with this question. 00:33:34.900 |
I love to ask it just 'cause it makes them uncomfortable. 00:33:41.780 |
It's a good, I don't know, maybe in French cuisine, 00:33:52.660 |
We're not now talking about the latest paper. 00:33:55.980 |
We're now talking about the bigger questions of life. 00:33:58.620 |
The simulation question is a nice one to do that. 00:34:05.180 |
I think there's two interesting things to say. 00:34:19.140 |
even primitive as they are now that are virtual. 00:34:21.580 |
I can already imagine that more and more people 00:34:38.720 |
in order for you to enjoy it better than this one? 00:34:48.100 |
'cause you can create virtual reality systems 00:34:53.860 |
by having people wanna stay in the virtual worlds. 00:34:56.860 |
And then the other question is the physics question 00:35:01.380 |
like what is the fundamental fabric of reality? 00:35:15.000 |
What are the mechanisms, the underlying mechanisms? 00:35:21.420 |
And there's actually people that written papers 00:35:28.860 |
And now quantum computers are coming forward, 00:35:46.180 |
way out of reach of our engineering capabilities. 00:35:48.780 |
But it's just, it's a nice party over the beer, 00:35:51.340 |
over beers thing to bring up with scientists. 00:35:55.140 |
There's two things that make scientists uncomfortable 00:36:03.820 |
And the other is, what do you think about the idea 00:36:29.780 |
to be one of the greatest runners of all time, 00:36:32.100 |
even though he's quite outpaced by the runners today. 00:36:38.900 |
because of his advancements in classical mechanics 00:36:49.620 |
Does it involve looking for the most advancements 00:36:54.940 |
Does it come from the journey and the work associated 00:37:01.420 |
or is it something you understand across humanity? 00:37:13.020 |
from the perspective of the individual for me. 00:37:21.060 |
It's just like, to me, I'm the greatest human 00:37:47.100 |
they tend to then tell stories about these pursuits. 00:37:50.980 |
And they like to, like greatness is something 00:37:56.260 |
They give Nobel prize, they give prizes to accomplishment. 00:37:59.540 |
They kind of tell stories about human beings, 00:38:06.860 |
And some are completely ignored through history. 00:38:09.460 |
Some are glorified through history, like over glorified. 00:38:13.620 |
I recently found out that the Pythagorean theorem 00:38:28.540 |
But that's an example of somebody I at least thought 00:38:32.140 |
was kind of an actual entity, an actual human being 00:38:34.740 |
that was great and associated with this idea. 00:38:37.660 |
So to me, I think greatness is doing the things you love. 00:38:45.740 |
whether they tell a good story about you or not. 00:38:48.500 |
- Give it up for our speaker, Dr. Leslie Kuhn.