back to indexMichael I. Jordan: Machine Learning, Recommender Systems, and Future of AI | Lex Fridman Podcast #74
Chapters
0:0 Introduction
3:2 How far are we in development of AI?
8:25 Neuralink and brain-computer interfaces
14:49 The term "artificial intelligence"
19:0 Does science progress by ideas or personalities?
19:55 Disagreement with Yann LeCun
23:53 Recommender systems and distributed decision-making at scale
43:34 Facebook, privacy, and trust
61:11 Are human beings fundamentally good?
62:32 Can a human life and society be modeled as an optimization problem?
64:27 Is the world deterministic?
64:59 Role of optimization in multi-agent systems
69:52 Optimization of neural networks
76:8 Beautiful idea in optimization: Nesterov acceleration
79:2 What is statistics?
89:21 What is intelligence?
97:1 Advice for students
99:57 Which language is more beautiful: English or French?
00:00:00.000 |
The following is a conversation with Michael I. Jordan, 00:00:14.640 |
and has mentored many of the world-class researchers 00:00:27.480 |
All of this to me is as impressive as the over 32,000 points 00:01:06.040 |
In many ways, the underlying spirit of this podcast 00:01:22.640 |
from the individual to our civilization as a whole. 00:01:40.720 |
As usual, I'll do one or two minutes of ads now, 00:02:07.640 |
Since Cash App does fractional share trading, 00:02:11.040 |
let me mention that the order execution algorithm 00:02:15.240 |
to create the abstraction of the fractional orders 00:02:27.760 |
that takes a step up to the next layer of abstraction 00:02:32.200 |
making trading more accessible for new investors 00:02:50.240 |
that is helping to advance robotics and STEM education 00:02:55.800 |
And now, here's my conversation with Michael I. Jordan. 00:03:01.380 |
Given that you're one of the greats in the field of AI, 00:03:05.200 |
machine learning, computer science, and so on, 00:03:15.600 |
so technically MJ is the Michael I. Jordan of basketball, 00:03:22.160 |
calling you the Miles Davis of machine learning, 00:03:25.280 |
because as he says, you reinvent yourself periodically 00:03:28.100 |
and sometimes leave fans scratching their heads 00:03:32.400 |
So can you put, at first, your historian hat on 00:03:37.120 |
and give a history of computer science and AI 00:03:42.280 |
including the four generations of AI successes 00:03:49.480 |
Yeah, first of all, I much prefer Yan's metaphor. 00:03:59.860 |
So I think I have one, but it's not just the one you lived, 00:04:08.800 |
I think what's happening right now is not AI. 00:04:22.380 |
or electrical engineering from electromagnetism. 00:04:34.180 |
But people pretty clearly viewed interesting goals 00:04:39.360 |
to try to build factories that make chemicals products 00:04:42.220 |
and do it viably, safely, make good ones, do it at scale. 00:04:47.960 |
So people started to try to do that, of course, 00:05:06.520 |
It's the real thing, real concepts are needed. 00:05:11.600 |
There was Maxwell's equations, which in some sense 00:05:13.960 |
were everything you need to know about electromagnetism, 00:05:16.380 |
but you needed to figure out how to build circuits, 00:05:17.980 |
how to build modules, how to put them together, 00:05:19.680 |
how to bring electricity from one point to another 00:05:23.220 |
So a whole field developed called electrical engineering. 00:05:27.940 |
is that we have a proto-field, which is statistics, 00:05:41.260 |
and use human data and mix in human decisions. 00:05:48.420 |
In fact, if you want to call machine learning a field, 00:05:51.340 |
That's a proto-form of engineering based on statistical 00:05:54.260 |
and computational ideas of previous generations. 00:05:56.500 |
- But do you think there's something deeper about AI 00:06:01.740 |
to chemical engineering and electrical engineering? 00:06:08.020 |
I think that that's like the Greeks sitting there 00:06:09.620 |
and saying it would be neat to get to the moon someday. 00:06:12.940 |
- I think we have no clue how the brain does computation. 00:06:19.180 |
on most anything interesting scientifically of our era. 00:06:28.500 |
but a little bit unique in the clarity of that. 00:06:34.820 |
like where we stand in our understanding of the human brain? 00:06:39.660 |
we're not very far in understanding the human brain, 00:06:48.940 |
that really study real synapses or real neurons, 00:06:55.700 |
and they're building it up slowly and surely. 00:07:02.600 |
We think it's electrical, maybe it's chemical. 00:07:09.420 |
And that's even around like a single synapse. 00:07:11.060 |
If you look at a electron micrograph of a single synapse, 00:07:15.820 |
And that's one little thing on a dendritic tree, 00:07:24.060 |
are even flying around and then proteins are taking that 00:07:26.100 |
and taking it down into the DNA and who knows what. 00:07:29.380 |
So it is the problem of the next few centuries. 00:07:50.500 |
All right, that's like the Greeks speculating 00:07:55.180 |
And I think that I like to say this fairly strongly 00:08:00.780 |
Because a lot of people who don't talk about it clearly 00:08:38.900 |
that's working on putting electrodes into the brain 00:08:46.260 |
Just as you said, even the basic mechanism of communication 00:08:57.980 |
the fundamental principles of how the brain works, 00:09:07.300 |
So I hope in the sense like anybody else hopes 00:09:09.860 |
for some interesting things to happen from research. 00:09:12.540 |
I would expect more something like Alzheimer's 00:09:14.760 |
will get figured out from modern neuroscience. 00:09:17.100 |
That, you know, there's a lot of human suffering 00:09:20.620 |
And we throw things like lithium at the brain, 00:09:25.700 |
That's not quite true, but you know, mostly we don't know. 00:09:28.020 |
And that's even just about the biochemistry of the brain 00:09:34.620 |
We just, we were really, really completely dim. 00:09:47.580 |
It's just, so it's like kind of sitting in a satellite 00:09:53.100 |
and trying to infer things about the microeconomy, 00:09:55.140 |
even though you don't have microeconomic concepts. 00:10:02.380 |
Can you control a cursor or mouse with your brain? 00:10:07.740 |
You know, and I can imagine business models based on that. 00:10:11.180 |
And even, you know, medical applications of that. 00:10:20.580 |
you know, I just, no, I don't agree with Elon Musk. 00:10:22.580 |
I don't think that's even, that's not for our generation. 00:10:30.780 |
you've mentioned Kolmogorov and Turing might pop up. 00:10:34.460 |
Do you think that there might be breakthroughs 00:10:38.380 |
that will get you to sit back in five, 10 years 00:10:45.220 |
but I don't think that there'll be demos that impress me. 00:10:49.220 |
I don't think that having a computer call a restaurant 00:10:56.340 |
And people, you know, some people present it as such. 00:11:07.020 |
And so fine, the world runs on those things too. 00:11:11.060 |
And I don't want to diminish all the hard work 00:11:13.780 |
and engineering that goes behind things like that 00:11:19.620 |
And I know the people that work on these things, 00:11:23.580 |
You know, in the meantime, they've got to kind of, 00:11:26.260 |
and they got mouths to feed and they got things to do. 00:11:35.980 |
that originally emerged, that had real principles 00:11:39.900 |
and you have a little scientific understanding, 00:11:48.620 |
And so we don't want to muddle too much these waters 00:11:52.460 |
of what we're able to do versus what we really can do 00:11:59.420 |
but I think that someone comes along in 20 years, 00:12:02.200 |
a younger person who's absorbed all the technology 00:12:07.840 |
I think they have to be more deeply impressed. 00:12:13.860 |
- The demos, but do you think the breakthroughs 00:12:25.460 |
to have fundamental breakthroughs in engineering? 00:12:31.020 |
that Elon Musk is working with SpaceX and then others 00:12:34.620 |
sort of trying to revolutionize the fundamentals 00:12:38.940 |
of saying here's a problem we know how to do a demo of 00:12:44.580 |
- Yeah, so there's gonna be all kinds of breakthroughs. 00:12:48.320 |
I'm a scientist and I work on things day in and day out 00:12:55.900 |
Also, I don't like to prize theoretical breakthroughs 00:13:02.620 |
and I think there's lots to do in that arena right now. 00:13:11.680 |
But Musk, God bless him, also will say things about AI 00:13:18.020 |
and he leads people astray when he talks about things 00:13:23.540 |
Trying to program a computer to understand natural language, 00:13:26.180 |
to be involved in a dialogue like we're having right now, 00:13:32.500 |
sort of take old sentences that humans use and retread them, 00:13:38.500 |
And so from that, I hope you can perceive that deeper, 00:13:51.220 |
Bring value to human beings at scale in brand new ways 00:13:58.660 |
because there's not a kind of a engineering field 00:14:10.980 |
so we don't think these things through very well. 00:14:19.220 |
just like electrical engineering was a breakthrough 00:14:24.500 |
- So the scale, the markets that you talk about 00:14:26.500 |
and we'll get to, will be seen as sort of breakthrough. 00:14:37.300 |
Can you give, we kind of threw off the historian hat. 00:14:42.140 |
I mean, you briefly said that the history of AI 00:14:46.420 |
kind of mimics the history of chemical engineering, 00:14:51.260 |
just to let you know, I don't, you know, I resist that. 00:14:55.340 |
AI really was John McCarthy as almost a philosopher 00:15:03.380 |
"If we could mimic the human capability to think 00:15:06.180 |
"or put intelligence in in some sense into a computer?" 00:15:11.860 |
and he wanted to make it more than philosophy. 00:15:13.540 |
He wanted to actually write down logical formula 00:15:17.220 |
And that is a perfectly valid, reasonable thing to do. 00:15:25.380 |
and I'd love to hear what you think about it, 00:15:33.140 |
- Maybe your version of it is that mine doesn't. 00:15:37.700 |
It does optimization, it does sampling, it does-- 00:15:40.060 |
- So systems that learn is what machine learning is. 00:15:45.460 |
So it's not just pattern recognition and finding patterns. 00:15:48.620 |
It's all about making decisions in real worlds 00:15:52.420 |
- So something like symbolic AI, expert systems, 00:15:55.300 |
reasoning systems, knowledge-based representation, 00:16:04.620 |
- So I don't even like the word machine learning. 00:16:06.140 |
I think that with the field you're talking about 00:16:07.940 |
is all about making large collections of decisions 00:16:10.140 |
under uncertainty by large collections of entities. 00:16:13.740 |
And there are principles for that at that scale. 00:16:15.980 |
You don't have to say the principles are for a single entity 00:16:18.100 |
that's making decisions, a single agent or a single human. 00:16:20.480 |
It really immediately goes to the network of decisions. 00:16:27.140 |
So we can continue the conversation to use AI for all that. 00:16:32.860 |
that this is not about, we don't know what intelligence is 00:16:38.060 |
We don't know much about abstraction and reasoning 00:16:41.780 |
We're not trying to build that because we don't have a clue. 00:16:47.380 |
but eventually we'll start to get glimmers of that. 00:16:52.820 |
we're trying to make good decisions based on that. 00:17:03.700 |
They will look, computers were so dumb before. 00:17:11.200 |
But, so machine learning, you can scope it narrowly 00:17:14.860 |
as just learning from data and pattern recognition. 00:17:17.820 |
But whatever, when I talk about these topics, 00:17:23.580 |
It really is important that the decisions are, 00:17:47.420 |
But it's, let's not say that what the goal of that is 00:17:52.920 |
The goal of that is really good working systems 00:17:54.800 |
at planetary scale that we've never seen before. 00:17:56.540 |
- So reclaim the word AI from the Dartmouth conference 00:17:59.260 |
from many decades ago of the dream of human-- 00:18:07.080 |
the history was basically that McCarthy needed a new name 00:18:15.280 |
Norbert Wiener was kind of an island to himself. 00:18:17.320 |
And he felt that he had encompassed all this. 00:18:48.600 |
But thinking forward about creating a sober academic 00:18:50.840 |
and real world discipline, it was a terrible choice 00:18:52.760 |
because it led to promises that are not true, 00:19:00.400 |
because you're one of the great personalities 00:19:02.320 |
of machine learning, whatever the heck you call the field, 00:19:05.160 |
do you think science progresses by personalities 00:19:09.520 |
or by the fundamental principles and theories 00:19:11.880 |
and research that's outside of personalities? 00:19:15.880 |
And I wouldn't say there should be one kind of personality. 00:19:19.440 |
and I have a kind of network around me that feeds me 00:19:23.520 |
and some of them agree with me and some of them disagree, 00:19:34.720 |
And I do think that there's some good to that. 00:19:38.080 |
It certainly attracts lots of young people to our field, 00:19:40.600 |
but a lot of those people come in with strong misconceptions 00:19:48.240 |
And so I think there's just gotta be some multiple voices 00:19:51.480 |
and I wasn't hearing enough of the more sober voice. 00:20:00.460 |
what would you say is the most interesting disagreement 00:20:09.880 |
that I don't think we disagree about very much really. 00:20:13.360 |
He and I both kind of have a let's build it kind of mentality 00:20:17.320 |
and does it work kind of mentality and kind of concrete. 00:20:21.280 |
We both speak French and we speak French more together 00:20:25.640 |
And so if one wanted to highlight a disagreement, 00:20:31.840 |
I think it's just kind of where we're emphasizing. 00:20:40.920 |
And it's interesting to try to take that as far as you can. 00:20:46.600 |
what would that give you kind of as a thought experiment? 00:20:57.840 |
that allow me to figure out what you're about ready to do, 00:21:02.900 |
Moreover, most of us find ourselves during the day 00:21:06.140 |
in all kinds of situations we had no anticipation of 00:21:09.220 |
that are kind of various, novel in various ways. 00:21:13.240 |
And in that moment, we want to think through what we want. 00:21:16.200 |
And also there's gonna be market forces acting on us. 00:21:18.940 |
I'd like to go down that street, but now it's full 00:21:23.780 |
I gotta think about what I might really want here. 00:21:26.120 |
And I gotta sort of think about how much it costs me 00:21:34.580 |
and prediction systems don't do any risk evaluations. 00:21:38.980 |
I gotta think about other people's decisions around me. 00:21:41.060 |
I gotta think about a collection of my decisions. 00:21:43.500 |
Even just thinking about like a medical treatment. 00:21:45.740 |
I'm not gonna take the prediction of a neural net 00:21:48.740 |
about my health, about something consequential. 00:21:55.820 |
that's ever been collected about heart attacks, 00:22:00.920 |
I'm not gonna trust the output of that neural net 00:22:03.900 |
I'm gonna wanna ask what if questions around that. 00:22:06.460 |
I'm gonna wanna look at some other possible data 00:22:10.420 |
I'm gonna wanna have a dialogue with a doctor 00:22:17.580 |
And I think that if you say prediction is everything, 00:22:23.160 |
And so prediction plus decision-making is everything, 00:22:30.020 |
Aeon, rightly so, has seen how powerful that is. 00:22:35.700 |
that decision-making is where the rubber really 00:22:37.300 |
hits the road, where human lives are at stake, 00:22:45.900 |
you gotta think about the economy around your decisions, 00:22:56.020 |
But if you go out in the real world, in industry, 00:23:02.140 |
and the other half are doing the pattern recognition. 00:23:05.620 |
- And the words of pattern recognition and prediction, 00:23:07.660 |
I think the distinction there, not to linger on words, 00:23:11.020 |
but the distinction there is more a constraint 00:23:13.780 |
sort of in the lab data set versus decision-making 00:23:17.340 |
is talking about consequential decisions in the real world 00:23:20.320 |
under the messiness and the uncertainty of the real world. 00:23:25.300 |
the whole mess of it that actually touches human beings 00:23:30.700 |
- It helps add that perspective, that broader perspective. 00:23:35.620 |
On the other hand, if you're a real prediction person, 00:23:37.380 |
of course you want it to be in the real world, 00:23:39.700 |
I'm just saying that's not possible with just data sets, 00:23:53.480 |
- So one of the things that you're working on, 00:24:02.060 |
especially in the clarity that you talk about it, 00:24:12.820 |
that help make decisions that scale in a distributed way, 00:24:28.500 |
you're absolutely getting into some territory 00:24:33.580 |
that are gonna be very not obvious to think about. 00:24:35.580 |
Just like, again, I like to think about history a little bit 00:24:38.860 |
but think about, put yourself back in the '60s, 00:24:54.560 |
that are actually all valid and so on and so forth. 00:25:00.140 |
when you start to get serious about things like this. 00:25:13.340 |
there is no music market in the world right now, 00:25:30.820 |
But there's a long tail of huge numbers of people 00:25:37.100 |
They are not in a market, they cannot have a career. 00:25:45.360 |
- The creators, the so-called influencers or whatever, 00:25:49.280 |
So there are people who make extremely good music, 00:25:52.140 |
especially in the hip hop or Latin world these days. 00:26:00.940 |
and they put it up on SoundCloud or other sites. 00:26:15.140 |
especially young kids listening to music all the time. 00:26:18.880 |
very little of the music they're listening to 00:26:21.840 |
and none of it's old music, it's all the latest stuff. 00:26:30.000 |
Of course, there'll be a few counter examples. 00:26:31.440 |
The record companies incentivize to pick out a few 00:26:35.160 |
Long story short, there's a missing market there. 00:26:37.600 |
There is not a consumer-producer relationship 00:26:47.960 |
they make money off of subscriptions or advertising 00:26:52.340 |
And then they will offer bits and pieces of it 00:27:02.180 |
that you actually are somebody who's good enough 00:27:12.220 |
here's all the places your songs were listened to. 00:27:17.100 |
vettable so that if someone down in Providence 00:27:19.580 |
sees that you're being listened to 10,000 times 00:27:21.900 |
in Providence, that they know that's real data, 00:27:25.260 |
they will have you come give a show down there. 00:27:28.460 |
who've been listening to you that you're coming. 00:27:34.380 |
You do that three times a year, you start to have a career. 00:27:40.660 |
it's creating new jobs because it creates a new market. 00:27:44.300 |
you've now connected up producers and consumers. 00:27:48.180 |
can say to someone who comes to their shows a lot, 00:27:49.820 |
"Hey, I'll play at your daughter's wedding for $10,000." 00:27:55.100 |
Then again, you can now get an income up to $100,000. 00:28:01.940 |
And now even think about really the value of music 00:28:06.660 |
even so much so that a young kid wants to wear a T-shirt 00:28:11.380 |
with their favorite musician's signature on it, right? 00:28:14.740 |
So if they listen to the music on the internet, 00:28:24.300 |
Well, because the kid who bought the shirt will be happy, 00:28:26.340 |
but more the person who made the music will get the money. 00:28:32.260 |
So you can create markets between producers and consumers, 00:28:35.160 |
take 5% cut, your company will be perfectly sound, 00:28:47.780 |
kind of create some connections and all that. 00:28:50.860 |
you think about the challenges in the real world 00:28:53.900 |
And there are actually new principles gonna be needed. 00:28:56.100 |
You're trying to create a new kind of two-way market 00:28:57.980 |
at a different scale that's ever been done before. 00:28:59.880 |
There's gonna be unwanted aspects of the market, 00:29:09.500 |
it'll fail in some ways, it won't deliver value. 00:29:18.780 |
And so that maybe doesn't get at all the huge issues 00:29:21.300 |
that can arise when you start to create markets, 00:29:22.620 |
but it starts, at least for me, solidify my thoughts 00:29:25.740 |
and allow me to move forward in my own thinking. 00:29:28.540 |
- Yeah, so I talked to, had a research at Spotify, 00:29:31.140 |
actually, I think their long-term goal, they've said, 00:29:36.660 |
make a comfortable living putting on Spotify. 00:29:41.020 |
So, and I think you articulate a really nice vision 00:29:53.400 |
What do you think companies like Spotify or YouTube 00:30:08.440 |
Is it some other kind of, is it an economics problem? 00:30:13.080 |
Who should they hire to solve these problems? 00:30:28.680 |
but really, I think missing that kind of culture. 00:30:31.200 |
All right, so it's literally that 16-year-old 00:30:34.720 |
You don't create that as a Silicon Valley entity. 00:30:39.260 |
You have to create an ecosystem in which they are wanted 00:30:44.340 |
And so you have to have some cultural credibility 00:30:53.900 |
It's such a terrible word, but it's culture, right? 00:31:04.400 |
but it's kind of like rich white people's thing to do. 00:31:07.200 |
You know, and American culture has not been so much 00:31:13.100 |
all the Africans who came and brought that culture 00:31:23.660 |
And so companies can't artificially create that. 00:31:33.640 |
not to denigrate, these companies are all trying 00:31:35.260 |
and they should, and I'm sure they're asking these questions 00:31:44.380 |
You got to blend your technology with cultural meaning. 00:31:49.380 |
- How much of a role do you think the algorithm, 00:31:52.220 |
so machine learning has in connecting the consumer 00:31:55.180 |
to the creator, sort of the recommender system 00:32:08.860 |
And recommender systems was a billion dollar industry 00:32:13.800 |
And it continues to be extremely important going forward. 00:32:24.460 |
'cause they put the book people are out of business, 00:32:34.180 |
and poor people reading them than ever before. 00:32:39.340 |
So, you know, that's how economics sometimes work. 00:32:42.780 |
But anyway, when I finally started going there 00:32:47.460 |
I was really pleased to see another few books 00:32:49.900 |
being recommended to me that I never would have thought of. 00:32:56.120 |
And I still to this day kind of browse using that service. 00:33:00.860 |
And I think lots of people get a lot, you know, 00:33:03.780 |
that is a good aspect of a recommendation system. 00:33:05.780 |
I'm learning from my peers in an indirect way. 00:33:08.960 |
And their algorithms are not meant to have them impose 00:33:13.860 |
It really is trying to find out what's in the data. 00:33:16.600 |
It doesn't work so well for other kinds of entities, 00:33:18.620 |
but that's just the complexity of human life like shirts. 00:33:20.880 |
You know, I'm not gonna get recommendations on shirts. 00:33:37.440 |
with other economic ideas, matchings and so on 00:33:42.140 |
is really, really still very open research wise. 00:33:45.200 |
And there's new companies that are gonna emerge 00:33:47.360 |
- What do you think is going to the messy, difficult land 00:34:02.980 |
So they're having, recommend the kind of news 00:34:06.580 |
that are most likely for you to be interesting. 00:34:13.580 |
Do you think it's a solvable problem for machines 00:34:15.860 |
or is this a deeply human problem that's unsolvable? 00:34:20.260 |
I think that what's broken with some of these companies, 00:34:25.300 |
They're not, at least Facebook, I wanna critique them, 00:34:28.020 |
that they didn't really try to connect a producer 00:34:34.620 |
And so they all, starting with Google, then Facebook, 00:34:37.260 |
they went back to the playbook of the television companies 00:34:43.180 |
They will pay for the TV box, but not for the signal, 00:34:50.380 |
and it somehow didn't take over our lives quite. 00:35:05.620 |
It didn't seem like that was gonna happen, at least to me. 00:35:08.340 |
These little things on the right-hand side of the screen 00:35:10.060 |
just did not seem all that economically interesting, 00:35:12.220 |
but that companies had maybe no other choice. 00:35:14.340 |
The TV market was going away and billboards and so on. 00:35:22.940 |
it was doing so well with that and making such money, 00:35:24.980 |
it didn't think much more about how, wait a minute, 00:35:32.860 |
Is there an actual market between the producer and consumer? 00:35:40.620 |
the person who could adjust it as a function of demand, 00:35:59.820 |
And I think they also didn't think very much about that. 00:36:02.980 |
So fast forward and now they are making huge amounts 00:36:13.540 |
So you want more people to click on certain things 00:36:24.060 |
you try to adjust it with your smart AI algorithms, right? 00:36:33.740 |
It does get into all the complexity of human life 00:36:38.460 |
but you could also fix the whole business model. 00:37:06.740 |
And I could go on the web right now and search. 00:37:12.140 |
I'll have lots of advertisers in my face, right? 00:37:14.740 |
What I really wanna do is broadcast to the world 00:37:18.500 |
and have someone on the other side of a market look at me 00:37:23.940 |
So they're not looking at all possible people 00:37:26.020 |
They're looking at the people who are relevant to them. 00:37:33.340 |
but I'm happy because what I'm gonna get back 00:37:35.020 |
is this person can make a little video for me 00:37:37.140 |
or they're gonna write a little two page paper 00:37:38.860 |
on here's the cool things that you want to do 00:37:45.140 |
I'm gonna pay, you know, $100 or whatever for that. 00:37:52.260 |
it's that I'm gonna pay that person in that moment. 00:38:06.140 |
they would make more of those things independently. 00:38:08.900 |
You don't have to incentivize them any other way. 00:38:13.580 |
I don't think I've thought long and heard about that. 00:38:19.020 |
who's just not thought about these things at all. 00:38:20.460 |
I think thinking that AI will fix everything. 00:38:25.260 |
because they were already out in the real world. 00:38:26.500 |
They were delivering packages to people's doors. 00:38:33.500 |
but, you know, they're in that business model. 00:38:36.300 |
And then I'd say Google sort of hovers somewhere in between. 00:38:38.420 |
I don't think for a long, long time they got it. 00:38:47.140 |
and that they're probably heading that direction. 00:38:49.740 |
But, you know, Silicon Valley has been dominated 00:39:06.780 |
- So advertisement, if we're gonna linger on that, 00:39:11.220 |
I don't know if everyone really deeply thinks about that. 00:39:28.740 |
you know, find the better angels of our nature 00:39:31.740 |
and do good by society and by the individual. 00:39:38.660 |
there's a difference between should and could. 00:39:51.700 |
can we, because the advertising model is so successful now 00:39:55.300 |
in terms of just making a huge amount of money 00:39:58.460 |
and therefore being able to build a big company 00:40:00.540 |
that provides, has really smart people working 00:40:05.460 |
And just to clarify, you think we should move away? 00:40:16.940 |
'cause I kind of want to learn more about how they do things. 00:40:18.820 |
So, you know, I'm not speaking for Amazon in any way, 00:40:22.820 |
because I actually believe they get a little bit of this 00:40:35.260 |
And, you know, without revealing too many deep secrets 00:40:43.480 |
You do not want a world where there's zero advertising. 00:40:51.720 |
is trying to bring products to customers, right? 00:40:55.700 |
and then you get more, you want to buy a vacuum cleaner, 00:40:57.300 |
say, you want to know what's available for me. 00:40:59.420 |
And, you know, it's not gonna be that obvious. 00:41:02.140 |
The recommendation system will sort of help, right? 00:41:16.540 |
It's not gonna be in the recommendation system. 00:41:20.740 |
and AI will not fix that, okay, at all, right? 00:41:31.900 |
it's a signal that you believe in your product enough 00:41:35.480 |
that you're willing to pay some real money for it. 00:41:37.340 |
And to me as a consumer, I look at that signal. 00:41:44.000 |
I know that, you know, these are super cheap, you know, 00:41:49.140 |
I know the company is only doing a few of these 00:41:51.020 |
and they're making, you know, real money is kind of flowing 00:41:53.220 |
and I see an ad, I may pay more attention to it. 00:41:55.140 |
And I actually might want that because I see, 00:41:57.300 |
hey, that guy spent money on his vacuum cleaner. 00:42:03.180 |
And so that's part of the overall information flow 00:42:22.440 |
stick with things that annoy a lot of people. 00:42:28.360 |
So I think that a Google probably is smart enough 00:42:30.720 |
to figure out that this is a dead, this is a bad model, 00:42:37.920 |
And I'm sure the CEO there will figure it out, 00:42:46.100 |
but you reduce it at the same time you bring up 00:42:48.780 |
producer, consumer, actual real value being delivered, 00:42:58.020 |
from the kind of the poor kind of advertising. 00:43:00.060 |
And I think that a good company will do that, 00:43:08.140 |
They bring, you know, grandmothers, you know, 00:43:11.800 |
they bring children's pictures into grandmothers' lives, 00:43:15.680 |
But they need to think of a new business model. 00:43:22.240 |
Until they start to connect producer, consumer, 00:43:24.280 |
I think they will just continue to make money 00:43:34.800 |
- So I apologize that we kind of returned to words. 00:43:43.580 |
don't you think the kind of direct connection 00:44:00.840 |
So that is best advertisement is literally now, 00:44:08.620 |
and will be able to actually start automatically 00:44:14.460 |
So like, I apologize if it's just a matter of terms, 00:44:21.300 |
just better and better and better algorithmically 00:44:25.540 |
almost a direct connection? - That's a good question. 00:44:32.500 |
So I was defending it as a way to get signals into a market 00:44:34.980 |
that don't come any other way, especially algorithmically. 00:44:44.520 |
then if I trust other people, I might be willing to listen. 00:44:56.560 |
And I find it creepy that they know I'm going to India 00:45:01.920 |
Can we, so what, can you just put your PR hat on? 00:45:10.500 |
and not trust them as do majority of the population? 00:45:14.100 |
So they're, out of the Silicon Valley companies, 00:45:18.640 |
but there's ranking of how much people trust companies 00:45:23.020 |
- In the gutter, including people inside of Facebook. 00:45:29.980 |
that right now we're talking that I might walk out 00:45:31.840 |
on the street right now that some unknown person 00:45:33.720 |
who I don't know kind of comes up to me and says, 00:45:38.800 |
That's just a, I want transparency in human society. 00:45:42.600 |
I want to have, if you know something about me, 00:45:44.580 |
there's actually some reason you know something about me. 00:45:51.460 |
You know something about me 'cause you care in some way. 00:45:58.000 |
Not just that you're someone who could exploit it 00:46:08.640 |
And that Facebook knows things about a lot of people 00:46:11.560 |
and could exploit it and does exploit it at times. 00:46:20.040 |
'cause they care about them, right, in any real sense. 00:46:24.160 |
They should not be a big brother caring about us. 00:46:29.800 |
Not the big brother part, but the caring, the trusting. 00:46:34.500 |
just to linger on it because a lot of companies 00:46:38.360 |
I would argue that there's companies like Microsoft 00:46:41.120 |
that has more information about us than Facebook does. 00:46:47.400 |
Microsoft, you know, under Satya Nadella has decided 00:46:56.200 |
that they really would approve of, that we don't decide. 00:47:00.120 |
And I'm just kind of adding that the health of a market 00:47:04.660 |
is that when I connect to someone who, producer, consumer, 00:47:18.060 |
in moments that I choose, of my choosing, then fine. 00:47:22.700 |
So, and also think about the difference between, 00:47:29.420 |
I just wanna buy, you know, a gadget or something. 00:47:33.080 |
I need some ammonia for my house or something, 00:47:43.160 |
I want to just go and have it be extremely easy 00:47:56.800 |
I don't want the company's algorithms to decide for me. 00:48:02.460 |
if Facebook thinks they should take the control from us 00:48:09.060 |
how much it relates to what they know about us 00:48:10.960 |
that we didn't really want them to know about us. 00:48:13.200 |
They're not, I don't want them to be helping me in that way. 00:48:23.040 |
So Facebook, by the way, I have this optimistic thing 00:48:26.560 |
where I think Facebook has the kind of personal information 00:48:29.520 |
about us that could create a beautiful thing. 00:48:32.400 |
So I'm really optimistic of what Facebook could do. 00:48:36.600 |
It's not what it's doing, but what it could do. 00:48:45.380 |
- Create a beautiful thing is really, let's be clear. 00:48:48.340 |
It's about something that people would value. 00:48:51.000 |
And I don't think they have that business model. 00:48:53.720 |
And I don't think they will suddenly discover it 00:49:09.680 |
I think a lot of other people will discover it. 00:49:11.300 |
I think that this, so I should also, full disclosure, 00:49:22.400 |
and the NBA, the music you find behind NBA clips right now 00:49:26.800 |
That's a company that had the right business model 00:49:32.620 |
And from day one, there was value brought to, 00:49:37.280 |
who suddenly their songs are on the NBA website, right? 00:49:45.400 |
- So you and I differ on the optimism of being able to 00:49:49.200 |
sort of change the direction of the Titanic, right? 00:50:11.440 |
And I would, there's a lot of times in the day 00:50:14.320 |
where something makes me either smile or think in a way 00:50:17.640 |
where I like consciously think this really gave me value. 00:50:25.720 |
way better than the New York Times themselves, by the way, 00:50:29.140 |
That's like real journalism is happening for some reason 00:50:31.660 |
in the podcast space, it doesn't make sense to me. 00:50:41.680 |
And how difficult, that's kind of what you're getting at, 00:50:47.800 |
How difficult is it to create a frictionless system 00:50:50.200 |
like Uber has, for example, for other things? 00:50:55.240 |
- So first of all, I pay little bits of money to, 00:51:00.360 |
I like Medium as a site, I don't pay there, but I would. 00:51:06.440 |
I would have loved to pay you a dollar and not others. 00:51:16.120 |
The goal is to actually have a broadcast channel 00:51:18.520 |
that I monetize in some other way if I chose to. 00:51:23.080 |
I could, I'm not doing it, but that's fine with me. 00:51:26.160 |
Also, the musicians who are making all this music, 00:51:28.360 |
I don't think the right model is that you pay 00:51:30.120 |
a little subscription fee to them, all right? 00:51:34.600 |
and it's just not that, that's not where the value is. 00:51:37.840 |
between real human beings, then you can follow up on that, 00:51:49.600 |
that recommend cool stuff to me, but it's pretty hard, 00:51:58.600 |
What's unknown about me is the most interesting. 00:52:00.760 |
- So this is the really interesting question. 00:52:13.480 |
- Yeah, but you don't, because, so for example, 00:52:15.640 |
this morning, I clicked on, I was pretty sleepy 00:52:24.400 |
I do not give a damn about the Queen of England. 00:52:30.320 |
"What the heck are they talking about there?" 00:52:31.560 |
I don't wanna have my life heading that direction. 00:52:38.400 |
will think that I care about the Queen of England. 00:52:44.320 |
that's been kind of the model, is if you collect 00:52:45.920 |
all this stuff, you're gonna figure all of us out. 00:52:48.880 |
Well, if you're trying to figure out one person, 00:52:50.420 |
like Trump or something, maybe you could figure him out, 00:52:52.600 |
but if you're trying to figure out 500 million people, 00:52:59.600 |
I think we are, humans are just amazingly rich 00:53:03.880 |
Every one of us has our little things that could intrigue us 00:53:09.880 |
but by God, there it comes, and you fall in love with it, 00:53:12.860 |
and I don't want a company trying to figure that out 00:53:18.480 |
a place that I kind of go, and by hook or by crook, 00:53:22.920 |
I'm walking down the street, and I hear some Chilean music 00:53:26.040 |
being played, and I never knew I liked Chilean music, 00:53:27.640 |
but wow, so there is that side, and I want them 00:53:30.360 |
to provide a limited, but interesting place to go, right? 00:53:34.760 |
And so don't try to use your AI to kind of figure me out 00:53:38.920 |
and then put me in a world where you figured me out. 00:53:45.040 |
where our creativity and our style will be enriched 00:53:48.440 |
and come forward, and it'll be a lot of more transparency. 00:53:55.240 |
based on stuff they know about me, facts that you know. 00:53:58.320 |
We are so broken right now, especially if you're a celebrity, 00:54:01.800 |
but it's about anybody that, anonymous people 00:54:04.920 |
are hurting lots and lots of people right now, 00:54:06.560 |
and that's part of this thing that Silicon Valley 00:54:08.560 |
is thinking that just collect all this information 00:54:12.400 |
So no, I'm not a pessimist, I'm very much an optimist 00:54:15.520 |
by nature, but I think that's just been the wrong path 00:54:23.480 |
Don't try to replace them, that's the AI mantra. 00:54:26.600 |
Don't try to anticipate them, don't try to predict them, 00:54:30.320 |
'cause you're not gonna be able to do those things, 00:54:34.600 |
- Okay, so right now, just give this a chance, 00:54:38.520 |
right now the recommender systems are the creepy people 00:54:52.920 |
the way they know you is by having conversation, 00:54:55.000 |
by actually having interactions back and forth. 00:54:57.320 |
Do you think there's a place for recommender systems, 00:55:00.680 |
sort of to step, 'cause you just emphasized the value 00:55:03.200 |
of human to human connection, but just give it a chance, 00:55:05.800 |
AI human connection, is there a role for an AI system 00:55:12.880 |
to try to figure out what kind of music you like, 00:55:23.040 |
maybe you're saying you have autism for Facebook, 00:55:25.120 |
so there I think it's misplaced, but I think that-- 00:55:34.840 |
- Yeah, no, good, human interaction on our daily, 00:55:37.440 |
the context around me in my own home is something 00:55:39.660 |
that I don't want some big company to know about at all, 00:55:49.240 |
I think Alexa's a research platform right now, 00:55:50.800 |
more than anything else, but Alexa done right, 00:55:53.640 |
could do things like, I leave the water running in my garden 00:55:56.600 |
and I say, "Hey, Alexa, the water's running in my garden," 00:55:59.120 |
and even have Alexa figure out that that means 00:56:00.640 |
when my wife comes home that she should be told about that. 00:56:03.520 |
That's a little bit of a reasoning, I would call that AI, 00:56:05.920 |
and by any kind of stretch, it's a little bit of reasoning, 00:56:13.400 |
but I kind of think that overall rises human happiness up 00:56:18.280 |
- But not when you're lonely, Alexa knowing loneliness-- 00:56:21.460 |
- No, no, I don't want Alexa to feel intrusive, 00:56:24.840 |
and I don't want just the designer of the system 00:56:32.340 |
and if a company can stand up and give me that 00:56:37.360 |
be way more successful than our current generation, 00:56:39.760 |
and like I said, I was mentioning Microsoft earlier, 00:56:45.060 |
but I think that they get that this is the way to go, 00:56:58.120 |
and that's the right business model going forward. 00:57:05.960 |
I mean, first of all, it should be an individual decision. 00:57:11.760 |
Privacy is not a zero one, it's not a legal thing, 00:57:16.880 |
it's not just about which data is available, which is not. 00:57:20.320 |
I like to recall to people that a couple hundred years ago, 00:57:25.960 |
everyone lived in on the countryside and villages, 00:57:28.720 |
and in villages, everybody knew everything about you, 00:57:30.840 |
very, you didn't have any privacy, is that bad? 00:57:34.880 |
Well, arguably no, because what did you get for that loss 00:58:01.200 |
I shouldn't just be adrift in a sea of technology 00:58:04.720 |
I don't wanna go reading things and checking boxes. 00:58:15.280 |
it's not just legal scholars meeting technologists, 00:58:18.680 |
there's gotta be kind of a whole layers around it. 00:58:20.840 |
And so when I alluded to this emerging engineering field, 00:58:29.640 |
but you just didn't plug electricity into walls 00:58:34.120 |
you don't have to have like underwriters laboratory 00:58:39.560 |
and that that machine will do this and that and everything, 00:58:41.680 |
there'll be whole people who can install things, 00:58:44.480 |
there'll be people who can watch the installers, 00:58:49.760 |
And for things as deeply interesting as privacy, 00:58:52.960 |
which is as least as interesting as electricity, 00:58:55.840 |
that's gonna take decades to kind of work out, 00:58:57.520 |
but it's gonna require a lot of new structures 00:59:02.160 |
- And you're saying there's a lot of money to be made 00:59:05.840 |
you should look at. - A lot of money to be made 00:59:07.000 |
and all these things that provide human services 00:59:08.800 |
and people recognize them as useful parts of their lives. 00:59:23.040 |
and newsrooms you see too much of this kind of thing. 00:59:29.080 |
are where we need to be having our conversations. 00:59:31.280 |
And actually there's not many forum for those. 00:59:39.000 |
Maybe I could go and I could read a comment section 00:59:49.120 |
because people are really hungry for conversation. 00:59:55.680 |
so comment sections of anything, including YouTube, 01:00:08.920 |
but it's a less anonymity, a little more locality, 01:00:15.520 |
and you trust the people there in those worlds 01:00:20.920 |
but it's not gonna be a total waste of your time, 01:00:24.800 |
A lot of us, I pulled out of Facebook early on 01:00:26.760 |
'cause it was clearly gonna waste a lot of my time, 01:00:31.000 |
And so yeah, worlds that are somehow you enter in 01:00:37.680 |
but you kind of have some trust in that world. 01:00:40.680 |
- And there's some deep, interesting, complex, 01:00:56.760 |
I was just, I didn't see any negatives there at all. 01:01:11.840 |
- So sorry for the big philosophical question, 01:01:13.760 |
but on that topic, do you think human beings, 01:01:19.140 |
do you think human beings are fundamentally good? 01:01:23.680 |
Like all of us have good intent that could be mind, 01:01:28.200 |
or is it, depending on context and environment, 01:01:41.240 |
We don't see the other person's pain that easily. 01:01:43.840 |
We don't see the other person's point of view that easily. 01:01:46.600 |
We're very much in our own head, in our own world. 01:01:54.000 |
and more less blinkered, and more understanding. 01:01:59.840 |
They didn't, they thought the other person was doing this, 01:02:17.520 |
cause them to do things they probably wouldn't want. 01:02:25.760 |
part of the progress of technology is to indeed allow it 01:02:28.520 |
to be a little easier to be the real good person 01:02:31.760 |
- Well, but do you think individual human life, 01:02:35.740 |
or society, could be modeled as an optimization problem? 01:02:44.300 |
one of the most complex phenomena in the whole, 01:02:47.520 |
- Which the individual human life, or society, 01:02:51.160 |
I mean, individual human life is amazingly complex. 01:03:12.200 |
what kind of properties does that surface have? 01:03:19.960 |
- Well, so optimization's just one piece of mathematics. 01:03:22.080 |
You know, there's like, just even in our era, 01:03:24.600 |
we're aware that, say, sampling is coming up, 01:03:28.040 |
examples of something, coming up with a distribution. 01:03:43.920 |
that is the optimum of a criterion function of some kind. 01:03:47.340 |
And sampling is trying to, from that same surface, 01:03:56.760 |
So I want the entire distribution in a sampling paradigm 01:04:01.360 |
and I want the single point that's the best point 01:04:11.100 |
the output of that could be a whole probability distribution. 01:04:13.280 |
So you can start to make these things the same. 01:04:25.920 |
- So, as a small tangent, what kind of world view 01:04:42.480 |
'Cause what you're gonna ask, what we're gonna do, 01:04:48.760 |
And so the best I can do is have kind of rough sense 01:04:53.160 |
and somehow use that in my reasoning about what to do now. 01:05:04.780 |
look like, so optimization can optimize sort of, 01:05:12.440 |
Sort of, at least from my, from a robotics perspective, 01:05:22.640 |
this game-theoretic concept starts popping up. 01:05:30.520 |
'Cause you've talked about markets and the scale. 01:05:47.700 |
is optimizing a functional called a Lagrangian. 01:05:55.640 |
So it's a description mathematically of something 01:05:57.440 |
that helps us understand as analysts what's happening. 01:06:00.760 |
And so the same thing will happen when we talk about 01:06:12.860 |
Now, at some point I may have set up a multi-agent 01:06:18.960 |
And I'm now thinking about an individual agent 01:06:24.800 |
They get certain signals and they have some utility. 01:06:33.240 |
So an optus could be embedded inside of an overall market. 01:06:45.100 |
there's just the, I don't know what you're gonna do. 01:06:51.460 |
And we kind of go back and forth in our own minds. 01:07:02.700 |
Maybe you can describe what saddle points are. 01:07:09.600 |
that you could try to explicitly look for saddle points 01:07:16.760 |
So there's all kinds of different equilibrium game theory. 01:07:20.320 |
And some of them are highly explanatory behavior. 01:07:42.380 |
Some of the simplest equilibria are saddle points. 01:07:53.980 |
you're trying to find algorithms that would find them. 01:07:57.740 |
I mean, so that's literally what you're trying to do. 01:08:17.940 |
So, you know, one example is a Stackelberg equilibrium. 01:08:21.060 |
So, you know, Nash, you and I are both playing this game 01:08:38.300 |
Now, since I know you're gonna look at my move, 01:08:43.660 |
But then I know that you are also anticipating me. 01:08:46.860 |
So we're kind of going back and forth in line. 01:08:51.440 |
And so those are different equilibria, all right? 01:09:04.460 |
You know, so some of these questions have answers, 01:09:14.420 |
there's just, you know, young people getting in this field 01:09:21.920 |
and they're really important and interesting. 01:09:31.060 |
So maybe I wanna push you in a part of the space 01:09:33.200 |
where I don't know much about you so I can get data. 01:09:35.580 |
And then later I'll realize that you'll never go there 01:09:43.780 |
- Even the game of poker is a fascinating space. 01:09:47.540 |
a lack of information, it's a super exciting space. 01:09:52.300 |
- Just lingering on optimization for a second. 01:10:04.100 |
that you see in the kinds of function surface 01:10:09.580 |
in the real world is trying to optimize over? 01:10:15.380 |
Is it just the usual kind of problems of optimization? 01:10:18.860 |
- I think from an optimization point of view, 01:10:28.060 |
there's kind of lots of paths down to reasonable optima. 01:10:31.340 |
And so kind of the getting downhill to an optima 01:10:34.500 |
is viewed as not as hard as you might have expected 01:10:38.380 |
The fact that some optima tend to be really good ones 01:10:51.940 |
from the particular generation of neural nets, 01:10:56.180 |
In 10 years, it will not be exactly those surfaces, 01:11:01.380 |
to why other surfaces or why other algorithms. 01:11:08.820 |
that's not, that didn't come from neuroscience per se. 01:11:10.980 |
I mean, maybe in the minds of some of the people 01:11:13.580 |
you know, about brains, but they were arithmetic circuits 01:11:20.340 |
And that layers of these could transform things 01:11:32.060 |
that it's working, it's able to work at this scale. 01:11:34.860 |
But I don't think that we're stuck with that, 01:11:44.020 |
sort of gradient descent, do you think we're stuck 01:11:56.940 |
these optimization spaces in more interesting ways? 01:12:01.820 |
and there are the architecture and the algorithm. 01:12:11.500 |
or matrix completion architectures and so on, 01:12:13.740 |
you know, I think we've kind of come to a place 01:12:16.540 |
where, yeah, a stochastic gradient algorithms 01:12:23.220 |
you know, that are a little better than others, 01:12:29.140 |
and there's ongoing research to kind of figure out 01:12:31.100 |
which is the best algorithm for which situation. 01:12:35.740 |
that that'll put pressure on the actual architecture, 01:12:37.780 |
and so we shouldn't do it in this particular way, 01:12:54.340 |
they have a lot of people who sort of study them 01:12:58.820 |
more deeply mathematically, are kind of shocked 01:13:07.140 |
if you move along the x-axis, you get, you know, 01:13:10.500 |
you go uphill in some objective by, you know, three units, 01:13:17.940 |
Now I'm gonna only allow you to move a certain, 01:13:30.140 |
so I'm gonna put all of it in the y-axis, right? 01:13:33.420 |
And why should I even take any of my strength, 01:13:37.140 |
my step size, and put any of it in the x-axis, 01:13:41.380 |
That seems like a completely, you know, clear argument, 01:13:45.100 |
and it's wrong, 'cause the gradient direction 01:13:51.660 |
And that, to understand that, you have to know some math, 01:13:58.340 |
so-called operator-like gradient is not trivial, 01:14:19.100 |
And we've come up with pretty favorable results 01:14:24.740 |
We know if you do this, we will give you a good guarantee. 01:14:29.980 |
that it must be done a certain way in general. 01:14:32.140 |
- So stochasticity, how much randomness to inject 01:14:46.700 |
but in some sense, again, it's kind of amazing. 01:14:52.660 |
particular features of a surface that could have hurt you 01:14:55.420 |
if you were doing one thing deterministically 01:15:01.380 |
you know, there's very little chance that you would get hurt. 01:15:03.260 |
And, you know, so here stochasticity, you know, 01:15:10.820 |
from some of the particular features of surfaces that, 01:15:14.260 |
you know, in fact, if you think about, you know, 01:15:17.220 |
surfaces that are discontinuous in a first derivative, 01:15:25.260 |
And if you're running a deterministic algorithm, 01:15:27.220 |
at that point, you can really do something bad, right? 01:15:30.060 |
Whereas stochasticity just means it's pretty unlikely 01:15:32.140 |
that's gonna happen, that you're gonna hit that point. 01:15:35.620 |
So, you know, it's again, non-trivial to analyze, 01:15:37.940 |
but especially in higher dimensions, also stochasticity, 01:15:43.260 |
but it has properties that kind of are very appealing 01:15:45.540 |
in high dimensions for kind of law of large number reasons. 01:15:52.600 |
is that you get to try to understand this mathematics. 01:15:58.580 |
partly empirically it was discovered stochastic gradient 01:16:05.060 |
but I don't see that we're getting clearly out of that. 01:16:15.580 |
- I don't know the most, but let me just say that, 01:16:18.580 |
you know, Nesterov's work on Nesterov acceleration to me 01:16:32.300 |
to move around into space, for the reasons I've alluded to, 01:16:41.740 |
you see this local person that can only sense 01:16:48.940 |
that's able to store all your previous gradients, 01:16:50.860 |
and so you start to learn something about the surface. 01:16:53.860 |
And I'm gonna restrict you to maybe move in the direction 01:17:02.740 |
So now we have a well-defined mathematical complexity model, 01:17:05.620 |
there's a certain classes of algorithms that can do that, 01:17:09.180 |
And we can ask for certain kinds of surfaces, 01:17:14.900 |
so for a smooth convex function, there's an answer, 01:17:19.460 |
which is one over the number of steps squared, 01:17:22.020 |
is that you will be within a ball of that size 01:17:27.140 |
Gradient descent in particular has a slower rate, 01:17:33.780 |
So you could ask, is gradient descent actually, 01:17:49.580 |
gradient is one over K, but is there something better? 01:17:58.100 |
that has got two pieces to it, it uses two gradients, 01:18:01.260 |
and puts those together in a certain kind of obscure way, 01:18:06.260 |
and the thing doesn't even move downhill all the time, 01:18:11.460 |
And if you're a physicist, that kind of makes some sense, 01:18:16.380 |
but that intuition is not enough to understand 01:18:29.660 |
and trying to explore that and understand it. 01:18:32.460 |
So there are lots of cool ideas in optimization, 01:18:35.100 |
but just kind of using gradients, I think, is number one, 01:18:37.380 |
that goes back 150 years, and then Nesterov, I think, 01:18:41.500 |
has made a major contribution with this idea. 01:18:51.900 |
- Coordinate descent is more of a trivial one, 01:18:55.300 |
- That's how we think, that's how our human minds-- 01:19:03.180 |
- An absurd question, but what is statistics? 01:19:09.780 |
it's somewhere between math and science and technology, 01:19:15.900 |
to make inferences that have got some reason to be believed, 01:19:18.820 |
and also principles that allow you to make decisions 01:19:31.180 |
but after you start making some of those assumptions, 01:19:37.900 |
I can guarantee that if you do this in this way, 01:19:40.700 |
your probability of making an error will be small. 01:19:43.460 |
Your probability of continuing to not make errors 01:19:47.580 |
and probability you found something that's real 01:20:06.420 |
because around that era, probability was developed 01:20:15.340 |
- So you would say, well, given the state of nature is this, 01:20:23.420 |
And especially if I do things long amounts of time, 01:20:27.660 |
And the physicists started to pay attention to this. 01:20:35.260 |
could I infer what the underlying mechanism was? 01:20:55.900 |
and he analyzed that data to determine policy, 01:21:24.340 |
there was game theory and decision theory developed nearby. 01:21:27.340 |
People in that era didn't think of themselves 01:21:34.380 |
And so, von Neumann is developing game theory, 01:21:36.540 |
but also thinking of that as decision theory. 01:21:39.180 |
Wald is an econometrician developing decision theory, 01:21:57.900 |
And to this day, in most advanced statistical curricula, 01:22:02.960 |
you teach decision theory as the starting point. 01:22:05.300 |
And then it branches out into the two branches 01:22:16.180 |
mysterious, maybe surprising idea that you've come across? 01:22:27.500 |
There's something that's way too technical for this thing, 01:22:33.500 |
and really takes time to wrap your head around. 01:22:43.460 |
wrote a really beautiful paper on James Stein estimation, 01:22:45.940 |
which just helps to, it's viewed as a paradox. 01:22:48.380 |
It kind of defeats the mind's attempts to understand it, 01:22:50.600 |
but you can, and Steve has a nice perspective on that. 01:22:58.360 |
is that it's like in physics, or in quantum physics, 01:23:02.420 |
There's a wave and particle duality in physics. 01:23:08.500 |
that you don't really quite understand the relationship. 01:23:11.660 |
The electron's a wave and electron's a particle. 01:23:16.660 |
There's Bayesian ways of thinking in Frequentist, 01:23:20.420 |
They sometimes become sort of the same in practice, 01:23:24.880 |
And then in some practice, they are not the same at all. 01:23:30.360 |
And so it is very much like wave and particle duality, 01:23:41.240 |
It's called Are You a Bayesian or a Frequentist? 01:23:43.200 |
And kind of help try to make it really clear. 01:23:47.060 |
So, decision theory, you're talking about loss functions, 01:23:51.300 |
which are a function of data X and parameter theta. 01:23:59.820 |
You don't know the data a priori, it's random, 01:24:04.380 |
So you have this function of two things you don't know, 01:24:06.220 |
and you're trying to say, I want that function to be small. 01:24:15.100 |
over these quantities or maximize over them or something 01:24:17.900 |
so that I turn that uncertainty into something certain. 01:24:21.940 |
So you could look at the first argument and average over it, 01:24:25.380 |
or you could look at the second argument, average over it. 01:24:27.900 |
So the Frequentist says, I'm gonna look at the X, the data, 01:24:42.020 |
And so it's looking at all the datasets you could get, 01:24:44.980 |
and saying how well will a certain procedure do 01:24:57.000 |
and people are using it on all kinds of datasets. 01:25:00.860 |
that has people running on many, many datasets 01:25:03.700 |
that 95% of the time it will do the right thing. 01:25:11.720 |
I'm gonna look at the other argument of the loss function, 01:25:17.500 |
So I could have my own personal probability for what it is. 01:25:22.140 |
I'm trying to infer the average height of the population. 01:25:24.020 |
Well, I have an idea of roughly what the height is. 01:25:32.060 |
So now that loss function has only now, again, 01:25:38.780 |
And that's what a Bayesian does, is they say, 01:25:40.340 |
well, let's just focus on a particular X we got, 01:25:44.940 |
Condition on the X, I say something about my loss. 01:25:50.300 |
And the Bayesian will argue that it's not relevant 01:25:53.220 |
to look at all the other datasets you could have gotten 01:25:55.840 |
and average over them, the frequentist approach. 01:25:58.640 |
It's really only the datasets you got, all right? 01:26:03.500 |
especially in situations where you're working 01:26:05.140 |
with a scientist, you can learn a lot about the domain, 01:26:07.540 |
and you're really only focused on certain kinds of data, 01:26:09.540 |
and you've gathered your data, and you make inferences. 01:26:12.280 |
I don't agree with it though, in the sense that 01:26:18.100 |
You're writing software, people are using it out there, 01:26:20.820 |
So these two things have got to fight each other a little bit 01:26:40.420 |
Write down a bunch of the math that kind of flows from that, 01:26:43.380 |
and then realize there's a bunch of things you don't know, 01:26:47.940 |
so you're uncertain about certain quantities. 01:26:56.500 |
there's quite a reasonable thing to do, to plug in. 01:26:59.780 |
There's a natural thing you can observe in the world 01:27:06.340 |
- So based on math or based on human expertise, 01:27:13.660 |
But the math kind of guides you along that path, 01:27:20.460 |
Under certain assumptions, this thing will work. 01:27:22.660 |
So you asked the question, what's my favorite, 01:27:29.620 |
which is you're making not just one hypothesis test, 01:27:39.300 |
you look at the ones where you made a discovery, 01:27:41.100 |
you announced that something interesting had happened. 01:27:43.740 |
All right, that's gonna be some subset of your big bag. 01:27:53.180 |
You'd like the fraction of your false discoveries 01:27:57.460 |
That's a different criterion than accuracy or precision 01:28:09.900 |
They say, given the truth is that the null hypothesis is true 01:28:25.780 |
And that's actually what false discovery rate is. 01:28:39.500 |
So I can't know that there's some priors needed in that. 01:28:42.440 |
And the empirical Bayesian goes ahead and plows forward 01:28:49.300 |
some of those things can actually be estimated 01:28:54.180 |
So this kind of line of argument has come out. 01:28:57.860 |
but it sort of came out from Robbins around 1960. 01:29:02.260 |
Brad Efron has written beautifully about this 01:29:06.260 |
And the FDR is, you know, Ben Yamini in Israel, 01:29:11.260 |
John Story did this Bayesian interpretation and so on. 01:29:14.700 |
So I've just absorbed these things over the years 01:29:16.940 |
and find it a very healthy way to think about statistics. 01:29:23.220 |
to jump slightly back out into philosophy, perhaps. 01:29:31.080 |
but you said that defining just even the question 01:29:33.980 |
of what is intelligence is a very difficult question. 01:29:41.900 |
the fundamentals of human intelligence and what it means? 01:29:44.800 |
You know, have good benchmarks for general intelligence 01:29:55.380 |
You're really asking a question for a psychologist, really. 01:29:58.500 |
And I studied some, but I don't consider myself 01:30:10.800 |
They might try to understand how a baby understands, 01:30:14.040 |
you know, whether something's a solid or liquid 01:30:18.400 |
And maybe how a child starts to learn the meaning 01:30:22.880 |
of certain words, what's a verb, what's a noun, 01:30:30.480 |
But humans' ability to take a really complicated 01:30:33.300 |
environment, reason about it, abstract about it, 01:30:35.860 |
find the right abstractions, communicate about it, 01:30:52.300 |
And certainly a psychologist doing experiments 01:30:54.040 |
with babies in the lab or with people talking 01:31:00.300 |
at our reasoning patterns, and they're not deeply 01:31:02.600 |
understanding all the how we do our reasoning, 01:31:04.660 |
but they're sort of saying, "Here's some oddities 01:31:06.500 |
"about the reasoning and some things you need 01:31:09.420 |
But also, as I emphasize in some things I've been writing 01:31:12.180 |
about, you know, AI, the revolution hasn't happened yet. 01:31:19.860 |
if you step back and look at intelligent systems 01:31:22.660 |
of any kind, whatever you mean by intelligence, 01:31:26.420 |
or, you know, the plants or whatever, you know. 01:31:29.400 |
So a market that brings goods into a city, you know, 01:31:31.880 |
food to restaurants or something every day is a system. 01:31:40.200 |
Every neuron is making its own little decisions, 01:31:44.340 |
And if you step back enough, every little part 01:31:46.460 |
of an economic system is making all of its decisions. 01:31:49.220 |
And just like with the brain, who knows what, 01:31:51.180 |
an individual neuron doesn't know what the overall goal is, 01:31:54.320 |
right, but something happens at some aggregate 01:32:01.240 |
It works at all scales, small villages to big cities. 01:32:09.160 |
So all the kind of, you know, those are adjectives 01:32:12.160 |
one tends to apply to intelligent systems, robust, 01:32:15.240 |
adaptive, you know, you don't need to keep adjusting it, 01:32:28.240 |
our humans are smart, but no markets are not. 01:32:50.280 |
The point, though, is that if you were to study humans 01:32:55.840 |
and come up with the theory of human intelligence, 01:32:57.440 |
you might have never discovered principles of markets, 01:33:00.440 |
you know, supply-demand curves and, you know, 01:33:11.400 |
There probably are third kinds of intelligence or fourth 01:33:20.440 |
Certainly the market one is relevant right now, 01:33:23.760 |
whereas understanding human intelligence is not so clear 01:33:31.960 |
or understanding intelligence in a deep sense and all that, 01:33:34.640 |
it definitely has to be not just human intelligence. 01:33:41.240 |
So, you know, it's definitely not just a philosophical stance 01:33:44.360 |
to say we gotta move beyond human intelligence. 01:33:54.000 |
it's some of the concepts you've just been describing. 01:33:59.000 |
if we see Earth, human civilization as a single organism, 01:34:02.600 |
do you think the intelligence of that organism, 01:34:05.160 |
when you think from the perspective of markets 01:34:07.040 |
and intelligence infrastructure is increasing? 01:34:14.080 |
What do you think the future of that intelligence? 01:34:25.000 |
Well, again, because you said it's so far in the future, 01:34:28.120 |
it's fun to ask and you'll probably, you know, 01:35:01.360 |
Is it possible to say anything interesting to that question 01:35:06.080 |
- It's not a stupid question, but it's science fiction. 01:35:09.040 |
- And so I'm totally happy to read science fiction 01:35:13.400 |
I love the, there was this like brain in a vat kind of, 01:35:16.200 |
you know, little thing that people were talking about 01:35:35.920 |
So the brain has got all the senders and receiver, 01:35:38.400 |
you know, on all of its exiting, you know, axons 01:35:49.680 |
And then you could do things like start killing off 01:35:56.720 |
You know, they thought they were out in the body 01:36:05.440 |
And I think every 18 year old should take philosophy classes 01:36:11.800 |
what could happen in society that's kind of bad 01:36:14.520 |
But I really don't think that's the right thing 01:36:15.800 |
for most of us that are my age group to be doing 01:36:19.500 |
I really think that we have so many more present, 01:36:32.040 |
on science fiction, at least in public for like this, 01:36:43.720 |
and I don't want a lot of people showing up there. 01:36:54.160 |
but maybe I don't have grounds for such optimism and hope. 01:36:59.000 |
Let me ask, you've mentored some of the brightest, 01:37:04.960 |
sort of some of the seminal figures in the field. 01:37:08.240 |
Can you give advice to people who are undergraduates today? 01:37:13.880 |
What does it take to take, you know, advice on their journey 01:37:16.760 |
if they're interested in machine learning and AI 01:37:19.080 |
and in the ideas of markets from economics to psychology 01:37:23.720 |
and all the kinds of things that you're exploring, 01:37:44.000 |
you spend a lot of time, you work on hard things, 01:37:46.880 |
you try and pull back and you be as broad as you can, 01:37:53.640 |
And it's like entering any kind of a creative community. 01:38:03.140 |
being a musician or being an artist or something, 01:38:05.960 |
you don't just, you know, immediately from day one, 01:38:08.480 |
you know, you're a genius and therefore you do it. 01:38:11.120 |
No, you, you know, practice really, really hard on basics 01:38:19.520 |
and then you realize you'll never be an expert 01:38:23.240 |
and there's a lot of randomness and a lot of kind of luck, 01:38:30.360 |
of the tree you go down, but you'll go down some branch. 01:38:40.640 |
It's very much about apprenticeship with an advisor. 01:38:43.080 |
It's very much about a group of people you belong to. 01:38:46.960 |
So it's plenty of time to start from kind of nothing 01:38:50.120 |
to come up to something, you know, more expertise 01:38:52.520 |
and then start to have your own creativity start to flower, 01:39:04.880 |
and I think in some other fields, it might be more so. 01:39:07.960 |
Here it's way more cooperative than you might imagine. 01:39:10.560 |
And people are always teaching each other something 01:39:13.480 |
and people are always more than happy to be clear. 01:39:16.760 |
So I feel I'm an expert on certain kinds of things, 01:39:19.300 |
but I'm very much not expert on lots of other things. 01:39:21.320 |
And a lot of them are relevant and a lot of them are, 01:39:23.520 |
I should know, but should in some sense, you know, you don't. 01:39:29.620 |
to people around me so they can teach me things. 01:39:31.840 |
And I think a lot of us feel that way about our field. 01:39:43.200 |
especially in the current era and everything, 01:39:44.520 |
is just at odds with the way that most of us think 01:39:47.960 |
where this is a human endeavor and we cooperate 01:40:01.360 |
And which language is more beautiful, English or French? 01:40:06.360 |
So first of all, I think Italian's actually more beautiful 01:40:22.960 |
And it is one of the great fun things to do in life 01:40:26.760 |
So in fact, when I kids or teens or college students 01:40:33.520 |
I say, well, do what your heart, where your heart is, 01:40:41.680 |
Throughout your life, you'll wanna be a thinking person. 01:40:53.480 |
I was living in the middle of the country in Kansas 01:40:59.880 |
And so my parents happened to have some French books 01:41:07.040 |
And I kind of learned the language by reading. 01:41:39.700 |
and kind of work on that language as part of that. 01:41:54.520 |
but it's all the creativity that went into it. 01:41:55.920 |
So I learned a lot of songs, read poems, read books. 01:42:11.920 |
So I just didn't have, I was kind of a bored person. 01:42:15.960 |
There's happened to be a lot of Italians at MIT, 01:42:22.440 |
I said, well, I should learn this language too. 01:42:39.240 |
And the people don't have any idea if you haven't traveled 01:42:42.640 |
kind of how amazingly rich and I love the diversity. 01:42:46.840 |
It's not just a buzzword to me, it really means something. 01:42:49.120 |
I love the, embed myself with other people's experiences. 01:42:53.060 |
And so yeah, learning language is a big part of that. 01:42:56.440 |
I think I've said in some interview at some point 01:43:01.380 |
what would you really work on if you really wanted to do AI? 01:43:03.720 |
And for me, that is natural language and really done right. 01:43:09.640 |
That's to me, amazingly interesting scientific challenge. 01:43:14.960 |
- One we're very far away, but good natural language people 01:43:19.160 |
I think a lot of them see that's where the core of AI is. 01:43:24.480 |
You understand something about the human mind, 01:43:26.080 |
the semantics that come out of the human mind. 01:43:28.440 |
And I agree, I think that will be such a long time. 01:43:32.320 |
just 'cause I kind of, I was behind in the early days. 01:43:54.280 |
- Jan was right, you truly are the Miles Davis 01:43:57.560 |
I don't think there's a better place than it. 01:43:59.440 |
Mike, it was a huge honor talking to you today. 01:44:08.480 |
And thank you to our presenting sponsor, Cash App. 01:44:11.480 |
Download it, use code LexPodcast, you'll get $10 01:44:15.680 |
and $10 will go to FIRST, an organization that inspires 01:44:23.800 |
If you enjoy this podcast, subscribe on YouTube, 01:44:26.640 |
give it five stars on Apple Podcast, support on Patreon 01:44:29.880 |
or simply connect with me on Twitter @LexFriedman. 01:44:33.480 |
And now let me leave you with some words of wisdom 01:44:37.120 |
from Michael I. Jordan from his blog post titled, 01:44:40.400 |
"Artificial Intelligence, the revolution hasn't happened yet 01:44:44.240 |
calling for broadening the scope of the AI field." 01:44:48.500 |
We should embrace the fact that what we are witnessing 01:44:51.480 |
is the creation of a new branch of engineering. 01:44:54.280 |
The term engineering is often invoked in a narrow sense 01:44:57.640 |
in academia and beyond, with overtones of cold, 01:45:01.720 |
effectless machinery and negative connotations 01:45:06.540 |
But an engineering discipline can be what we want it to be. 01:45:10.960 |
In the current era, we have a real opportunity 01:45:19.760 |
I'll resist giving this emerging discipline a name, 01:45:35.780 |
Thank you for listening and hope to see you next time.