Back to Index

Escaping the Local Optimum of Low Expectation


Chapters

0:0 Overview - The Voice poem
6:46 Artificial intelligence
13:44 Open problems in AI
14:10 Problem 1: Learning to understand
17:15 Problem 2: Learning to act
19:28 Problem 3: Reasoning
20:44 Problem 4: Connection between humans & AI systems
23:57 Advice about life as an optimization problem
24:10 Advice 1: Listen to your inner voice - ignore the gradient
25:12 Advice 2: carve your own path
26:28 Advice 2: Measure passion not progress
28:10 Advice 4: work hard
29:5 Advice 5: forever oscillate between gratitude and dissatisfaction
31:10 Q&A: Meaning of life
33:11 Q&A: Simulation hypothesis
36:15 Q&A: How do you define greatness?

Transcript

It's wonderful to be here, wonderful to see so many faces that I've come to love over the years. My advisor, my family's here, my mom, brother. You know, I did ask security to make sure my dad doesn't, is not allowed in, but he somehow found his way in, so good job.

(audience laughing) The topic of today's talk reminds me of something my dad once told me. I wrote it down. Fewer those who see with their own eyes and feel with their own hearts. No, wait, that actually was Albert Einstein, different Jew, similar haircut for those of you. (audience laughing) Similar.

You know, there's a saying, there's an old saying that goes, "Give a man a fish and you feed him for a day, "teach a man to fish and you feed him for a lifetime." A little known fact, it actually goes on to say, "So that he may never discover how much he loves steak." Or vegetarian lasagna for those of you who are vegetarian in the audience.

And the key there, the key idea is society tries to, impose lessons to teach, to drive the human being, each of us, but it's you discovering your own passion is the key, and that's what the talk I'd like to talk about today. And there'll be a lot of poems throughout.

And the central poem by Shel Silverstein called "The Voice" is one I think that will resonate throughout the talk. There's a voice that's in the air, it's a voice that's in the air, it's a voice that's in the air. It's a voice that's in the air, it's a voice that's in the air.

It's a voice that's in the air, it's a voice that's in the air. It's a voice that's in the air, it's a voice that's in the air. It's a voice that's in the air, it's a voice that's in the air. And that's the poem we have together over two small topics, life and artificial intelligence.

Now, from an optimization perspective, and one of my co-advisors has always told me when you show a plot, you have to describe the X axis and the Y axis as a good engineer. There you go, that's lesson number one. The X axis is competence, the Y axis is confidence.

And there's something called the Dunning-Kruger effect, which is captured by this plot. And that is at the beginning of your journey of competence, when you're not very good at something, when you're first taking the first steps of learning something, as some of you here are in the engineering fields, you're overly confident.

It's the peak of confidence, and you're at the lowest stage of actually of your abilities, of your expertise. And it's funny that I'm speaking here before you today in a place of a complete sort of self-doubt and despair and not knowing what I'm doing at all. And I feel like I have zero expertise to impart on you.

And so in that sense, it's a funny position to be speaking with, especially some of the lessons, some of the advice I'll try to give. So take that with a grain of salt. And some of you sitting in the audience today may be at the very peak, especially if you're at the beginning of the college journey, university journey.

And I'd say to me, the biggest positive, the biggest impact of college and university education is the dismantling of the ego that's involved in going from that peak overconfidence to the valley of despair that I'm currently in. Oh, and I should mention that this is also the time for me and perhaps for you where folks like Dostoevsky start making a lot of sense, talking about suffering and pain and how the really great men and women must, I think as he says, have great sadness on earth.

This resonates with everybody in their undergraduate years in engineering. Now, the real thing I'd like to talk about is the broader optimization problem formed by the Dunning-Kruger effect, which is after the peak of confidence and the valley of despair, there's a gradient provided to you by your advisors, by your parents, by your friends, your loved ones, society in general.

The gradient over which you're optimized to achieve some definition of success. This is what I call the local optimum. What everybody else tells you you're supposed to do. What everybody else at the small scale, on a daily scale, and on the weekly scale, monthly, yearly, and for the rest of your life tells you what the definition of success is.

That's the local optimum. What I'd like to argue is some ideas of how to break out of that convention, of how to listen just enough to hear the lessons in society, advisors, friends, and parents, but for the rest of it, ignore their voices and only listen to your own voice.

And I'll tell you through my own story here. So I was introduced as a research scientist at MIT. And very recently, I decided to step down from MIT to do my own startup. I'm still affiliated there, but sort of give up the salary, give up everything, give up what I'm supposed to be, the definition under academic colleagues of what success is, of what the pursuit of the academic life is, because I'm listening to the voice inside.

And so I'm speaking to you at the very beginning of this journey, again, full of self-doubt. And so take with a grain of salt, but perhaps it's interesting to speak from this position, 'cause I would argue it's the most beautiful position to be in in life. The opportunity, the freedom in the struggles that I'm undergoing now is really a gift that comes at the end of this journey of college.

Now, who am I? And what is the dream that I mentioned there at the end? The global optimum. For me, that's understanding the human mind and engineering artificial intelligence systems. Visualized on the left here is just 3% of the neurons in the human brain. It's a mysterious, beautiful thing.

It's easy to forget how little we know about this mystery that's just between our two ears. And engineering machines that can reason, that can think, that can perceive the world is one of the ways we can understand this mysterious, beautiful thing that brings to life everything around us. And the dream of creating intelligence systems, companions, ones that you can have a deep connection with.

That's what drives me. That's my startup work. That's what my entrepreneurship work, that's my research work is focused on. Most of the work at MIT and before that has been on robotics and autonomous vehicles. But now the dream is to create a system that you can love and it can love you back.

A brief history of artificial intelligence to give you a sense, to give you a quick review if this is a totally new field. Again, if you're undergraduate, perhaps this is a field that you want to, that you want to take on as your journey. So it started on the theoretical end with Alan Turing and many of the ideas from philosophy to mathematics that he presented and from whom the field was born.

And on the engineering side, Frank Crosonblatt in building the Perceptron, the first machine. So engineering machines that can do some aspect of learning, some aspect of search that we associate with artificial intelligence. And then there's been accomplishments throughout, none greater, at least to me, than in this, at least for now in a span of games.

There's been two branches of artificial intelligence that have dominated the field. The early days have been, you can think of a search, as brute force search. It's not quite as captivating to our imagination. It doesn't quite feel like intelligence because it's brute force searching through possible answers until you find one that's optimal.

It's converting every single problem to a search problem and then bringing computational power to it to try to solve it. But nevertheless, the peak of that, especially for those who play chess, especially for those who might be a Russian, is when IBM D-Blue defeated Garry Kasparov in 1997. This is a seminal moment in artificial intelligence where the game that was associated with thought, with intelligence, with reason, was overcome, was the greatest champion and human champion was defeated by a machine.

And the seminal moment on the second branch of artificial intelligence, which is learning systems, systems that learn from scratch, knowing nothing, with zero human assistance, was able to defeat the greatest player in the world. Little side note, the first moment did have human assistance in the AlphaGo system from DeepMind and Google DeepMind.

And then the follow on a few months later, the system called AlphaZero was able to learn from scratch by playing itself. This is, to me, the greatest accomplishment of artificial intelligence. And I'll mention when I discuss it about open problems in the field. And then in a real world application, like I said, I worked a lot in autonomous vehicles.

This is one of the most exciting applications with autonomous and semi-autonomous vehicles. There's been deployments, lessons, explorations, a lot of different debates. This is the most exciting space of artificial intelligence. If you wanna have an impact as an engineer, autonomous vehicles is the space you will do so in the next, in the 2020s.

And a quick whirlwind overview of key ideas in artificial intelligence that were key breakthroughs. So neural networks and Perceptron, like I said, was born in the '40s, '50s, and '60s. With the algorithms that dominate today's world of deep learning and machine learning have been invented in many, many decades ago, in the '70s and '80s, with convolutional for the computer vision aspect of things in the '80s and '90s, with LSTM, recurrent neural networks, they work with language, work with sequence of data, were developed in the '90s and proven out in the aughts.

And then the deep learning quote unquote revolution, the term and the ideas of large-scale machine learning using neural networks was reborn in 2006 in the early aughts, and then proven out in the seminal ImageNet moment when computer vision systems were able to, in the challenge of object recognition, image recognition, and the ImageNet data set, and the ImageNet challenge, neural networks were able to far outperform the competition and do so easily from just learning from data.

And a few other developments. There's a lot of unsupervised learning, self-supervised learning ideas that were born in the '14, '15, '16, just a few years ago, and a lot of exciting ideas in the past few years. The past few years have been dominated by ideas in natural language processing with ideas of transformers.

Anyway, this might be outside the scope of what you're familiar with. I encourage you to look into it. Transformers in particular, with natural languages, some of the most beautiful and exciting ideas that without any human supervision, you can learn to model language sufficiently well to outperform anything we've done previously, to do things like machine translation to a level that's unprecedented.

It's really exciting. And especially exciting is that bigger is better, meaning that as long as we can scale compute, we can perform better and better and better. And it's a totally open question how, what the ceiling of that is. And finally, the most exciting thing in artificial intelligence is the idea, there's a concept of Big Bang for the start of the universe, a silly name for one of the most incredible mysteries of our human existence.

Same way, self-play is one of the silliest names for one of the most powerful ideas in artificial intelligence. It's the mechanism behind alpha zero. It's a system playing against itself to improve continuously without any human supervision. That is the most exciting aspect, the most exciting area that I'm excited and I recommend if you love learning that you explore.

So the open problems in artificial intelligence and possible solutions. And one of the things, and I'll focus on number four, which is something that is my dream, that is sort of my life aspiration, but I'll give a whirlwind introduction. Learning to understand, learning to act, reason, and a deep connection between humans and AI systems.

So learning to understand, there's a lot of exciting possibilities here. This is a lot of the breakthroughs in machine learning have been in something called supervised learning, where you have a set of data and you have a neural network or a model that's able to learn from that data in order to generalize sufficiently to infer on cases it hasn't seen before.

You could recognize cat versus dog. In the case of domain, in the domains of like autonomous driving, you can recognize lane markings, you could recognize other vehicles, pedestrians, all the different subtasks involved in solving a particular problem. Now that's all good, but to solve real world problems, you have to actually, you have to deal with endless edge cases that we human beings effortlessly take care of, that our ability to do reasoning and common sense reasoning effortlessly takes care of.

So to be able to learn over those edge cases, you have to do much larger scale learning. And for that, you have to be much more selective and clever about which data you annotate with human beings. And that's the idea of active learning. Same way with, as children, we explore the world, we interact with the world to pick up the lessons from it.

The same way you can interact with a dataset to select only small parts of it to learn from. And I'll take Tesla, which is a car company that's using autonomous driving and its system autopilot that uses deep learning to learn how to solve all these different problems. I'll use them as a case study.

What they're doing is quite interesting in the space of active learning. They're creating a pipeline for each individual task. They take the task of driving and break it apart into now over a hundred different subtasks. Each subtask gets its own pipeline, its own dataset. And there's a machine learning system that learns from that dataset and is then deployed back into the vehicles.

And when the vehicle fails in a particular case, that's an edge case that's marked for the system and is brought back to the pipeline to annotate. So there's ongoing pipeline that continuously goes on. The system is not very good in the beginning, but the whole purpose of it is to discover edge cases.

In the same way that us humans learn something, and you can think of our actually existence in the world as an edge case discovery mechanism. So you learn something, you construct a mental model of the world, and you move about the world until you run up against a case, a situation that you totally didn't expect.

And we do that thousands of times a day still, and we learn from those. And that pipeline of active learning is a really exciting area that very few people are working on, especially in the space of research. To me, that's the most exciting in terms of scale impact area in the next few years.

Learning to act, the second set of open problems in artificial intelligence. This is where the idea of self-play comes in, is learning to build systems, whether through a reinforcement learning mechanism or otherwise, that are actually acting in the world. In the case of self-play, the idea is that you have a really dumb system in the beginning that knows nothing.

Again, no human supervision. And through randomization, you have other systems that also know nothing, but know a different set of nothing. And they compete against each other. So you formulate the problem as a competitive setting. And when you have two dumb systems that compete against each other, a magical thing happens.

The one that's slightly less dumb starts winning. And this little incremental step can be repeated arbitrarily and without any constraints on human supervision, annotation costs, without any constraints on having to sort of bring the human in the loop or bring the physical world in the loop. It can all be done in computation in a distributed sense.

So you can, in a matter of hours on a distributed compute setting, create a system that beats the world champion at go. And in fact, with DeepMind and all the games that have they've defeated the world champion in chess, not just the world champion, is the best chess playing program, Stockfish, in a matter of hours of training.

And the ceiling hasn't yet been reached. This is both the exciting and the scary thing about self-play is very few times is the ceiling ever reached. What we hit is the limits of our computational power, which is computation power, especially the kind of mechanisms that are happening now, developments happening now.

The Moore's law is continuing in many ways. So computation, if you just wait a few years, computation is increasing. So we were yet to see the ceiling of the capabilities that these approaches are able to achieve. This should be both exciting and terrifying. Okay, the total biggest open problem that nobody even knows how to do.

This is an example of a state-of-the-art dog intelligence system solving a particular problem. So we know nothing how to do reasoning systems in artificial intelligence. This is the actually not very often talked about area because nobody knows what to do about it. There's been subsets called program synthesis, communities that kind of try to formulate a subset of the reasoning problem and try to solve it, but we don't know much to do, particularly common sense reasoning, how to formulate enough about the world to be able to reason about the physics of the world, about the basic, especially with human beings, human to human, human to physical world dynamics.

Just there's millions of facts seemingly that are intricately connected that we learn and we accumulate in a knowledge base. This process is a really exciting area of research that nobody knows what to do with. The things I've described previously don't really have anything to do with humans necessarily. The by-passion in my interest is that space between machine and human.

The community broadly could be called human-robot interaction, but there's a lot of different areas in which there's a deep connection between the human and machine that you all experience every day. So recommender systems from Netflix to much more importantly, social networks, the recommendation engines behind social networks, recommending what you see next in terms of both advertisement and about the content of your friends that you see, which friends you get to see more from.

The personalization of IOT, of smart systems, semi-autonomous systems like Tesla Autopilot and different semi-autonomous vehicles like the Cadillac Super Cruise systems. Whenever you have AI systems between you and a machine. So there's a machine that does, that automates some particular task. There's you human that are tasked with sitting there and supervising the machine.

And there is an AI system in the middle that manages that. It manages the tension, the dance, the uncertainty, the human, all the T word, the trust, all the mess of human beings, it manages that. That's a really exciting space that is in the very early days. What I show there is where my sense is, where we stand.

In 1998, there was a lot of search engines. Some of you may even be old enough to have used them. AltaVista, Excite, AskG is like us and so on. Then Google came along, the Google search engine and blew them all out of the water. They were all working on a very interesting, very important problem, but the approach and the fundamental ideas behind their approach was flawed.

I believe that personal assistance and a personal deep, meaningful connection between an AI system and a human being that's exactly where we're at. Many people have in their home an Alexa device, a Google home device. But most people don't use it for almost anything except to play music or check the weather.

Many of you use Twitter and social networks, but artificial intelligence plays a minimal role and understands almost nothing about you in recommending how you interact with the platform or the advertisements you see. And autonomous vehicles, robotics platforms know almost nothing about you. So shown there is the Tesla vehicle.

It knows almost nothing about you except whether your hands are on the steering wheel or not. I believe it'll be obvious in retrospect how much opportunity there is to learn about human beings from the devices and from that to form a deep, meaningful connection. So now to return to my valley of despair to give some words of advice.

And again, take them with a grain of salt. So in this context, in this optimization context, my first piece of advice is to listen to your inner voice. I think a lot of people, including a lot of very smart professors, advisors, parents, friends, significant others, have in them a kind of mutually agreed upon gradient along which they push you.

It's so difficult for me to articulate this in a clear way. But early on, I heard within myself a silly sounding, crazy voice that told me to do things. One of which was to try to put a robot in every home. There's dreams that are difficult for me to articulate.

But if you allow your mind to be quiet enough, you'll hear such voices, you'll hear such dreams. And it's important to really listen and to pursue them. Advice number two is carve your own path. And if that means taking a few detours, take the detours. Again, this is coming from the valley of despair.

(audience laughing) So I hope this pans out in the end. But I had many detours. In music, I was in a band, I had long hair. I gave a lot of myself to the practice of martial arts. And both music and martial arts have given me, again, very difficult to put into words, but it have given me something quite profound.

It gave flavor and color to the pursuit of that dream that's hard to articulate. It's because I listened to my instinct, listened to my heart in pursuing these detours. From poetry to excessive reading, like I mentioned, I took a James Joyce course here. So pursuing these avenues of knowledge through philosophy and history that seemingly have nothing to do with the main pursuit.

And starting the silliest of pursuits, starting a podcast. Advice number three is to measure passion, not progress. So most of us get an average of about 27,000 days of life. I think a good metric by which you should live is to maximize the number of those days that are filled with a passionate pursuit of something.

Not by how much you've progressed towards a particular goal. Because goals are grounded in your comparison to other human beings, to something that's already been done before. Passionate pursuit of something is the way you achieve something totally new. And a quick warning about passion. Again, I'm a little bit of Russian, so maybe I romanticize this whole suffering and passion thing.

(audience laughing) But the people who love you, the people who care for you, like I mentioned, your friends, your family, should not be trusted. Accept their love, but not their advice. Parents and significant others will tell you to find a secure job because passion looks dangerous. It looks insecure.

Advisors, colleagues will tell you to be pragmatic because passion looks like a distraction from the main effort that you should be focusing on. And society will tell you to find balance, work-life balance in your life because passion looks unhealthy. Advice number four, continuing on the unhealthy part, is work hard.

Make a habit of working hard every day, putting in the hours. There's a lot of books and a lot of advice that have been written on working smart and not working hard. I'm yet to meet anyone who has not truly worked hard for thousands of hours in order to accomplish something great.

In order to work smart, you first have to put in those few tens of thousands of hours of really dumb, brute force, hard work of all-nighters. The key there is to minimize stress, not to minimize the amount of hours of work. And to do that is you have to love what you do.

And the final piece of advice, I love that picture, okay, is to look up to the stars and appreciate every single moment you're alive. At the mystery of this world, at the beauty of this world. Again, this is my perspective, take it with a grain of salt, but I advise to forever oscillate between deep, profound doubt and self-dissatisfaction and a deep gratitude for the moment, for just being alive, for all the people around you that give you their love, with whom you get to share those moments and share the love.

A poem by Stephen Crane that I especially like in the desert. In the desert, I saw a creature, a naked bestial, who squatting up on the ground, held his heart in his hands and ate of it. I said, "Is it good, friend?" "It is bitter." "Bitter," he answered. "But I like it, because it is bitter, "and because it is my heart." So I would say the bitter is the self-dissatisfaction, and that's the restless energy that drives us forward.

And then enjoying that bitterness and enjoying the moment and enjoying the sweetness that comes from eating your own heart in this poem is a thing that makes life worthwhile. And that is, to me, happiness. So with those silly few pieces of advice, I'd like to continue on the gratitude and say thank you.

Thank you to my advisor. Thank you to this university for giving me a helping hand. There you go. And thank you to my family and all the friends that I've had along the way. Thank you for their love. I appreciate it. (audience applauding) I've never been introduced with this much energy.

I really appreciate it. (audience laughing) - You're hanging out at the wrong places, man. (laughing) - Yes. - First of all, great to see you in person, Dr. Kristine. Big fan of your lectures, big fan of your show, "Tell Me" podcast. Just listening to your conversation on these phones this morning, my way to my phone.

My question for you was, is your perspective in any way influenced by the ultimate being-less-ness of it all? (laughing) - By the way, thank you for that question. How is your daily life affected by the meaninglessness of it all? (audience laughing) So the answer is yes. And it's hard to use reason to justify that life is meaningful.

I think you have to listen to, there's something in you that makes life beautiful. So if you look at somebody like Elon Musk, he believes that interplanetary, so colonizing Mars, that's one of the most exciting things we human beings can do. And so if you allow yourself to think, what is the most exciting thing that we human beings can do?

And see that the work you're doing is part of that. For me, if I were to psychoanalyze myself, there's something in me that's deeply fulfilling about creating intelligent systems. That's so exciting to me, that we human beings can create intelligent systems. I see artificial intelligence as the next evolution of human civilization.

And to me, that makes it somehow deeply exciting, even though eventually the whole universe will collapse on itself or the other cold death of the universe. There's something within that that's so exciting. - There was an interview with Elon Musk and he basically said that we're in a civilization, so this might not be actual reality.

What's your take on that? - So my first take is, I love it how much fellow colleagues and scientists are uncomfortable with this question. So I love it. I love to ask it just 'cause it makes them uncomfortable. (audience laughing) Yeah, I appreciate it. It's a good, I don't know, maybe in French cuisine, you have to cleanse the palate.

It's a good question to ask. We're not now talking about the latest paper. We're now talking about the bigger questions of life. The simulation question is a nice one to do that. In terms of actually practically, I think there's two interesting things to say. So one, it's interesting to me, I'm a big fan of virtual reality.

I love entering worlds, even primitive as they are now that are virtual. I can already imagine that more and more people would wanna live in those worlds. It's an interesting question to me, how real do those worlds need to become in order for you to wanna stay there and not return to the real world?

So the question of the simulation is, how real do we need to simulate the world in order for you to enjoy it better than this one? That's a computer science question. That's really interesting. That's a, it's a, it's like practical engineering question 'cause you can create virtual reality systems that'll make a lot of money, perhaps have a detrimental effect on society by having people wanna stay in the virtual worlds.

And then the other question is the physics question of quantum mechanics of, like what is the fundamental fabric of reality? And is it, what does it take to simulate that reality? And that's like a physics question. How, is it finite, is it infinite? What are the mechanisms, the underlying mechanisms?

Does it go as low as string theory? Does it go below string theory? And there's actually people that written papers on how big a computer needs to be in order to simulate that kind of system. And now quantum computers are coming forward, which is one of the exciting applications of quantum computing is to be able to simulate quantum mechanical systems.

And this is the question, how big does a quantum computer have to be to simulate the universe? It's a fun, but a real physics question, way out of reach of our engineering capabilities. But it's just, it's a nice party over the beer, over beers thing to bring up with scientists.

There's two things that make scientists uncomfortable that I love bringing up. One is the simulation question. And the other is, what do you think about the idea that's become popular recently that the earth might be flat? (audience laughing) They get really, they get angry actually. So. - I wanna say, I appreciate your work and I love the podcast and stuff like that.

So people talk about athletes and academics being the greatest of their field. People consider Jesse Ellens to be one of the greatest runners of all time, even though he's quite outpaced by the runners today. People consider scientists like Isaac Newton one of the greatest science evers because of his advancements in classical mechanics and calculus, which is considered pretty basic physics nowadays.

What do you define greatness as when it comes to the pursuit of an endeavor? Does it involve looking for the most advancements in the field given your starting point? Does it come from the journey and the work associated or the destination? Is it a personal concept or is it something you understand across humanity?

- So thank you for that question. Very well written out and thought out. There's a personal greatness from the perspective of the individual for me. Like for me, greatness is doing what I love. That ignores the rest of society. It's just like, to me, I'm the greatest human to have ever lived in my own little world for having to do the things I love.

And that's from my perspective. And I love the craftsmanship of it. Anything, it could be anything. It's just doing the skill. So that's not about accomplishment. That's not about anything. That's about just doing the things you love. From the perspective of society, they tend to then tell stories about these pursuits.

And they like to, like greatness is something that people invent. They give Nobel prize, they give prizes to accomplishment. They kind of tell stories about human beings, about Steve Jobs, about different icons. And some are completely ignored through history. Some are glorified through history, like over glorified. I recently found out that the Pythagorean theorem was not developed by Pythagoras.

But I read it on Wikipedia. I don't know if it's true. But that's an example of somebody I at least thought was kind of an actual entity, an actual human being that was great and associated with this idea. So to me, I think greatness is doing the things you love.

And the rest is just luck, whether they tell a good story about you or not. - Give it up for our speaker, Dr. Leslie Kuhn. (audience applauding) (audience cheering) (audience cheering) (audience cheering) (audience cheering) (audience cheering) (audience cheering)