back to indexGeorge Hotz: Comma.ai, OpenPilot, and Autonomous Vehicles | Lex Fridman Podcast #31
Chapters
0:0 George Hotz
5:52 Virtual Reality
7:21 Iphone Hack
14:50 Nuclear Weapons
17:28 How Was Your Evolution as a Programmer
19:4 Product Zero
35:35 System Specs
42:15 Driver Monitoring
43:46 Cancel in Autopilot
61:46 Natural Lane Change
69:51 What Does Success Look like for Comm Ai
78:32 Who Are the Competitors
104:56 Trolley Problem
108:6 Teleoperation
00:00:00.000 |
The following is a conversation with George Hotz. 00:00:04.500 |
a machine learning based vehicle automation company. 00:00:07.420 |
He is most certainly an outspoken personality 00:00:10.220 |
in the field of AI and technology in general. 00:00:13.180 |
He first gained recognition for being the first person 00:00:18.420 |
And since then, he's done quite a few interesting things 00:00:21.260 |
at the intersection of hardware and software. 00:00:29.580 |
give it five stars on iTunes, support it on Patreon, 00:00:39.080 |
And I'd like to give a special thank you to Jennifer 00:00:41.960 |
from Canada for her support of the podcast on Patreon. 00:00:47.660 |
She's been a friend and an engineering colleague 00:00:57.860 |
And now, here's my conversation with George Hotz. 00:01:12.420 |
- So if the simulation is designed in such a way 00:01:19.620 |
to show that no information can get in and out, 00:01:31.300 |
- So they've designed it such that it's a closed system, 00:01:38.740 |
We're either in a simulation which can be exploited, 00:01:41.340 |
we're in a simulation which not only can't be exploited, 00:01:48.100 |
you can't even detect if you're in a VM or not. 00:01:52.460 |
So we're, it's, yeah, so the simulation's running 00:01:56.500 |
- Yeah, but now in reality, all VMs have ways to detect. 00:02:00.260 |
I mean, is it, you've done quite a bit of hacking yourself, 00:02:04.540 |
and so you should know that really any complicated system 00:02:10.980 |
- So this isn't necessarily true going forward. 00:02:14.180 |
I spent my time away from Kama, I learned Coq. 00:02:24.300 |
And if you write code that compiles in a language like that, 00:02:34.940 |
is written in a language like this, in which case, you know. 00:02:39.580 |
- Yeah, but that can't be sufficiently expressive 00:02:48.940 |
- The simulation doesn't have to be Turing-complete 00:02:52.300 |
- Looks like it does, actually, with entropy. 00:02:58.540 |
that results in something as complicated as the universe 00:03:02.220 |
would have a formal proof of correctness, right? 00:03:22.140 |
- But the fundamental rules might be super simple. 00:03:37.980 |
Like, how would you approach hacking a simulation? 00:03:44.340 |
- By the way, I'm not familiar with the talk you gave. 00:03:46.660 |
I just read that you talked about escaping the simulation 00:04:05.100 |
If we continue on the path we're going with technology, 00:04:10.100 |
I think we're in big trouble, like, as a species, 00:04:19.460 |
So, if we could change rhetoric to be more like, 00:04:28.140 |
that we're in a simulation and how we could get out, 00:04:34.820 |
well, I assume I would have acquired way more intelligence 00:04:37.380 |
in the process of doing that, so I'll just ask that. 00:04:39.740 |
- So, the thinking upwards, what kind of ideas, 00:04:51.440 |
Are you thinking sort of exploration in all forms? 00:04:54.060 |
- The space narrative that held for the modernist generation 00:04:59.060 |
doesn't hold as well for the postmodern generation. 00:05:10.020 |
Like, we're gonna build rockets, we're gonna go to Mars, 00:05:21.820 |
It doesn't seem to hold the same weight in today's culture. 00:05:26.720 |
I'm hoping for good postmodern narratives that replace it. 00:05:32.180 |
- So, let's think, so you work a lot with AI. 00:05:39.100 |
There could be also, I don't know how much you do 00:05:45.160 |
but every time I play with it in our research, 00:05:58.780 |
The apartment I can rent in the cloud is way better 00:06:00.660 |
than the apartment I can rent in the real world. 00:06:04.780 |
Because others will have very nice apartments too, 00:06:07.300 |
so you'll be inferior in the virtual world as well. 00:06:09.260 |
- No, but that's not how I view the world, right? 00:06:11.340 |
I don't view the world, I mean, it's a very, like, 00:06:24.340 |
You just touch the dish and it's washed, right? 00:06:30.140 |
or if everybody had the apartment, I don't care. 00:06:34.780 |
The world first learned of Geohot, George Hotz, 00:06:50.520 |
What was the first system you discovered vulnerabilities for 00:06:55.040 |
- So, that was really kind of the first thing. 00:07:01.240 |
I had a book in 2006 called "Grey Hat Hacking," 00:07:25.000 |
- You had to open it up and pull an address line high. 00:07:36.540 |
how memory chips are connected to processors and stuff. 00:07:43.920 |
So, your view of the world and computers was physical, 00:07:49.200 |
- Actually, if you read the code that I released with that 00:07:58.300 |
- And in a broken sort of state machine-esque C, 00:08:10.040 |
You've live streamed some of your programming. 00:08:12.680 |
This chaotic, beautiful mess, how did you arrive at that? 00:08:18.640 |
I interned at Google the summer after the iPhone unlock. 00:08:29.020 |
and I wrote a software library to interact with it. 00:08:41.120 |
Now, of course, just getting that feedback is not enough. 00:08:50.440 |
was I wanted to write this thing that could emulate 00:09:09.420 |
And I built the tool once and it was terrible. 00:09:13.440 |
I built the tool a second time, it was terrible. 00:09:16.320 |
This was by the time I was at Facebook, it was kind of okay. 00:09:31.840 |
to visualize the change of state of the computer 00:09:41.760 |
So you can rewind just as easily as going forward. 00:09:56.500 |
Think about it like Git for your computers, the run log. 00:09:59.760 |
- So there's like a deep log of the state of the computer 00:10:07.820 |
Why isn't that, or maybe it is, maybe you can educate me, 00:10:11.460 |
why isn't that kind of debugging used more often? 00:10:25.460 |
So that's gonna be really hard to use for that. 00:10:30.180 |
and for boot ROMs and for small parts of code. 00:10:33.180 |
So it's hard if you're trying to debug like massive systems. 00:10:42.480 |
And CTF were these competitions that I played, 00:10:46.880 |
- Capture the Flag, I was gonna ask you about that. 00:10:49.360 |
Look, I watched a couple of videos on YouTube. 00:10:52.880 |
What have you learned about maybe at the high level 00:10:55.520 |
vulnerability of systems from these competitions? 00:11:04.180 |
you had all of the best security people in the world 00:11:17.140 |
there's like a file on the server called flag. 00:11:24.980 |
and then you cat flag, and then you type the flag 00:11:27.100 |
into like a web-based scoreboard and you get points. 00:11:29.460 |
So the goal is essentially to find an exploit 00:11:59.100 |
- Some video game like misspelled own way back in the day. 00:12:02.820 |
- Yeah, and it's just, I wonder if there's a definition. 00:12:09.780 |
- Okay, so what was the heyday of CTF, by the way? 00:12:12.740 |
But was it, what decade are we talking about? 00:12:37.000 |
you've turned it less into a game of generic computer skill 00:12:48.900 |
it didn't have like as much attention as it had. 00:12:53.640 |
I won $30,000 once in Korea for one of these competitions. 00:12:59.520 |
but that means there was probably good people there. 00:13:06.780 |
or are they grounded in some real flaws in real systems? 00:13:34.560 |
you don't need one exploit anymore, you need nine. 00:13:44.800 |
speaking higher level philosophically about hacking? 00:13:48.280 |
I mean, it sounds from everything I've seen about you, 00:13:55.100 |
You don't wanna bring that exploit out into the world 00:14:10.560 |
for why I don't do defense for why I don't do crime. 00:14:19.280 |
If you find nine of those holes as a defender, 00:14:23.220 |
because the attacker gets in through the last one. 00:14:30.820 |
if you log on with a VPN nine out of the 10 times, 00:14:47.100 |
'cause I was having a discussion with somebody 00:14:49.540 |
just at a high level about nuclear weapons actually, 00:14:56.240 |
And my feeling is all the smart people in the world, 00:14:59.800 |
if you look at the distribution of smart people, 00:15:07.620 |
I was talking to Sean Carroll, the physicist, 00:15:10.560 |
good and bad people are evenly distributed amongst everybody. 00:15:14.020 |
My sense was good hackers are in general good people 00:15:32.100 |
But if you're good and you have these skills, 00:15:41.080 |
is there a little voice in your head that says, 00:15:51.760 |
And you could earn a lot of money doing it though, 00:16:10.600 |
But when I was 19, I read "Crime and Punishment." 00:16:16.080 |
that talked me out of ever really doing crime. 00:16:21.680 |
I'd get away with it, but it would just run through my head. 00:16:32.640 |
- I wouldn't say I'm good, I would just say I'm not bad. 00:17:02.140 |
but you look at things for what they actually are. 00:17:07.900 |
- I understand that's what you tell me it is, 00:17:17.800 |
I'm a better programmer now than I was in 2014. 00:17:31.600 |
You started with C, at which point did you pick up Python? 00:17:59.620 |
- Operating systems, some of those classes you mentioned 00:18:05.620 |
At least, the circa 2012 operating systems and compilers 00:18:22.660 |
- Somehow, I picked up Python that semester as well. 00:18:28.080 |
That's when I really started to get into CTFs. 00:18:30.320 |
And CTFs, you're all to race against the clock, 00:18:36.260 |
so you really wanna use the programming language 00:18:38.980 |
- 48 hours, pwn as many of these challenges as you can. 00:18:45.380 |
- You were both at Facebook and Google for a brief stint. 00:18:51.100 |
- With Project Zero, actually, at Google for five months, 00:18:59.300 |
Just curious about the security efforts in these companies. 00:19:05.180 |
- Well, Project Zero started the same time I went there. 00:19:12.340 |
- 2015, so that was right at the beginning of Project Zero. 00:19:18.860 |
I'll try to give the best public-facing explanation 00:19:38.460 |
Sometimes people will find these vulnerabilities 00:19:43.460 |
and submit them in bug bounties to the companies. 00:19:47.820 |
But a lot of the companies don't really care. 00:19:51.140 |
It doesn't hurt for there to be a vulnerability. 00:19:53.820 |
So, Project Zero is like, "We're gonna do it different. 00:19:57.780 |
"and we're gonna give them 90 days to fix it. 00:20:04.140 |
- We're gonna drop the weapon on the exploits. 00:20:15.820 |
- I watched your coding sessions that you streamed online. 00:20:18.940 |
You code things up, basic projects, usually from scratch. 00:20:30.380 |
and your brain works in both brilliant and chaotic ways. 00:20:46.940 |
So, I'm curious, is there a method to your madness? 00:20:54.740 |
There's pros and cons to my programming style 00:21:02.700 |
get something up and working quickly with like an API 00:21:08.220 |
because I will throw things at it until it works. 00:21:10.260 |
If you ask me to take a vector and rotate it 90 degrees 00:21:17.460 |
I'll spam program for two hours and won't get it. 00:21:22.380 |
- Oh, because it's something that you could do 00:21:23.940 |
with a sheet of paper, think through, design, 00:21:26.300 |
and then just, you really just throw stuff at the wall 00:21:30.460 |
and you get so good at it that it usually works. 00:21:34.660 |
- I should become better at the other kind as well. 00:21:39.460 |
It's nowhere near as entertaining on the Twitch streams. 00:21:41.180 |
I do exaggerate it a bit on the Twitch streams as well. 00:21:43.540 |
The Twitch streams, I mean, what do you wanna see a gamer? 00:21:50.260 |
I think I watched, I watched probably several hours 00:21:53.380 |
of you, like I've actually left you programming 00:22:16.940 |
- So I'm twice as frantic on the live streams 00:22:23.700 |
So I wasn't even paying attention to what you were coding, 00:22:33.020 |
I've developed the workload Facebook and stuck with it. 00:22:39.460 |
What's your like methodology for learning new things? 00:22:45.940 |
the distributed file systems out in the world 00:22:50.740 |
Like if you want to install something like Ceph, 00:23:03.980 |
but these are all like 10,000 plus line projects. 00:23:06.860 |
I think some of them are even a hundred thousand line 00:23:18.860 |
and has this little master server that I wrote in Go. 00:23:23.540 |
- This, if I would say that I'm proud per line 00:23:27.260 |
maybe there's some exploits that I think are beautiful 00:23:35.580 |
is because that was the fourth version of it that I wrote. 00:23:49.500 |
It's some, it's like in a way it's C++ but easier. 00:24:01.700 |
this is like Python but it takes twice as long 00:24:12.740 |
for large code bases and why you want something like Go. 00:24:15.820 |
- Interesting, so why doesn't Python work for, 00:24:23.380 |
basically demo level work with autonomous vehicles 00:24:29.180 |
- Why doesn't Python work for large code bases? 00:24:32.380 |
- Because, well, lack of type checking is a big-- 00:24:41.900 |
the compiler can tell you like nothing, right? 00:24:52.980 |
There's like linters that can catch it some of the time. 00:24:55.740 |
There's no types, is really the biggest downside 00:25:00.540 |
and then well, Python's slow but that's not related to it. 00:25:02.660 |
Well, maybe it's kind of related to its lack of-- 00:25:10.340 |
My adventure into dependently typed languages, 00:25:37.020 |
which doesn't run on, so why do I have to compile 00:25:45.740 |
I've embraced JavaScript recently just because, 00:25:52.260 |
It seems that these worst possible languages live on 00:25:57.420 |
- Yeah, well, it's in the browser and it's fast. 00:26:02.540 |
- It's in the browser and compute might stay, 00:26:09.020 |
in terms of distributed computation in the future, so. 00:26:15.220 |
- Yeah, it's interesting if autonomous vehicles 00:26:19.460 |
I mean, you have to consider these possibilities. 00:26:30.100 |
which lets you analyze the can traffic from the car. 00:26:32.900 |
- So basically anytime you're visualizing something 00:26:37.740 |
- Well, the web is the best UI toolkit by far. 00:26:40.940 |
- So, and then, you know what, you're coding in JavaScript. 00:26:54.940 |
how did you get into the world of vehicle automation? 00:26:57.900 |
Can you also just, for people who don't know, 00:27:20.540 |
And he's looking for somebody to build a vision system 00:27:30.860 |
Elon back then was looking for a replacement. 00:27:33.700 |
And he brought me in and we talked about a contract 00:27:41.380 |
I would get paid $12 million if I could deliver it tomorrow 00:27:53.220 |
Even if it takes me 10 months, I get $2 million. 00:27:57.140 |
Maybe I don't finish it at all and I get paid nothing 00:28:04.260 |
because I've been working in robotics for a long time 00:28:06.300 |
and I'm curious to see a person like you just step in 00:28:08.300 |
and sort of somewhat naive, but brilliant, right? 00:28:13.980 |
'cause you basically full steam take on a problem. 00:28:25.880 |
- I remember I suggested to Elon in the meeting, 00:28:30.580 |
putting a GPU behind each camera to keep the compute local. 00:28:38.020 |
I leave the meeting 10 minutes later and I'm like, 00:28:41.460 |
"thinking about this problem before I went in." 00:28:44.180 |
- Oh, just send all your cameras to one big GPU. 00:28:51.800 |
I was like, "Oh, I'll put the first few layers 00:29:02.940 |
I left that meeting with Elon like beating myself up. 00:29:07.060 |
- Yeah, you haven't, like you haven't at least like 00:29:13.260 |
Like usually in life, I get away with saying stupid things 00:29:26.100 |
and I'll like correct it and bring the conversation back. 00:29:33.540 |
I was much more prepared the second time I met him. 00:29:35.540 |
- Yeah, but in general, how hard did you think it is? 00:29:45.700 |
I didn't think I'd solve level five self-driving 00:29:52.780 |
- I saw, my friend showed me the outputs from a Mobileye 00:29:55.500 |
and the outputs from a Mobileye was just basically two lanes 00:30:04.800 |
- Well, first time I tried the implementation of Mobileye 00:30:07.540 |
and the Tesla, I was really surprised how good it is. 00:30:13.380 |
just 'cause I've done a lot of computer vision, 00:30:14.820 |
I thought it'd be a lot harder to create a system 00:30:24.980 |
'cause I was kind of skeptical before trying it. 00:30:27.860 |
'Cause I thought it would go in and out a lot more. 00:30:36.180 |
So what, how hard is the problem when you tackled it? 00:30:44.500 |
Like Elon talked about disengagements on the 405 00:30:47.340 |
down in LA with like lane marks were kind of faded 00:30:57.260 |
that I would say was like the same quality in three months. 00:31:11.140 |
you're kind of going by feel 'cause you tested it out. 00:31:15.540 |
Like I would take, I borrowed my friend's Tesla. 00:31:20.740 |
And then I would take my system out for a drive. 00:31:24.420 |
So the 405, how hard is it to create something 00:31:30.440 |
that could actually be a product that's deployed? 00:31:36.940 |
this respondent said something about you saying 00:32:00.620 |
but Elon followed that up by saying it's going to take 00:32:04.900 |
And here I am four years later with a company of 12 people. 00:32:10.780 |
So yeah, so what do you think about how Tesla's progressing 00:32:19.180 |
- I think we've kept pace with them pretty well. 00:32:26.820 |
We had some demo features internally of the same stuff 00:32:37.380 |
- Consumer Reports does a great job of describing it. 00:32:43.340 |
You shouldn't ship things like autopilot, open pilot, 00:32:49.740 |
If you turn it on for a stretch of a highway, 00:32:53.420 |
like an hour long, it's never going to touch a lane line. 00:33:00.060 |
I don't know if you're grounded in data on that. 00:33:05.360 |
I wonder actually how often we touch lane lines 00:33:21.740 |
That's interesting 'cause every time you touch a lane, 00:33:26.780 |
and kind of lane keeping is removing that stress. 00:33:29.280 |
- That's ultimately the biggest value add, honestly, 00:33:32.320 |
is just removing the stress of having to stay in lane. 00:33:35.480 |
And I think, honestly, I don't think people fully realize, 00:33:44.960 |
- And that, not only, I find it a huge value add. 00:33:52.640 |
and I missed it, I missed having the system so much. 00:34:00.280 |
It is that lane centering that's the key feature. 00:34:15.860 |
Maybe someday it'll be this big cliff where it adds value, 00:34:28.260 |
I actually believe that really good lane keeping 00:34:33.260 |
is a reason to buy a car, will be a reason to buy a car, 00:34:39.680 |
I've never, until we just started talking about it, 00:34:43.840 |
that I've felt with Elon's chase of level four 00:34:52.320 |
It was, 'cause you should just say Tesla has the best, 00:35:00.580 |
Kamiya I should say Kamiya has the best lane keeping, 00:35:12.900 |
but ACC is much more forgiving than lane keep, 00:35:17.500 |
- By the way, are you Kamiya's camera only, correct? 00:35:23.700 |
- From the car, you're able to get the, okay. 00:35:29.640 |
but we leave the radar there as like a, it's Fusion now. 00:35:37.920 |
What's the hardware side of what you're providing? 00:35:46.780 |
- So OpenPilot, as the box that we sell that it runs on, 00:35:56.640 |
So you're like, you buy the phone, it's just easy. 00:35:59.320 |
It'll be easy set up, but it's sold with no software. 00:36:08.240 |
I think we'll be ready for a consumer product. 00:36:11.540 |
We're just gonna make the lane keeping really, really good. 00:36:44.320 |
And the other one is the proxy camera CAN bus. 00:37:03.320 |
- So we proxy, the vehicles that we work with 00:37:10.180 |
So lane keeping assist can mean a huge variety of things. 00:37:13.820 |
It can mean, it will apply a small torque to the wheel 00:37:17.800 |
after you've already crossed a lane line by a foot, 00:37:26.520 |
lane keeping assist, where it'll keep you perfectly 00:37:31.060 |
- You can control, like with a joystick, the car. 00:37:35.080 |
So these cars already have the capability of drive-by-wire. 00:37:37.920 |
So is it trivial to convert a car that it operates with? 00:37:49.720 |
so we have support now for 45 different makes of cars. 00:37:56.360 |
We support almost every Honda and Toyota made this year. 00:38:00.620 |
And then a bunch of GMs, bunch of Subarus, bunch of Chevrolets. 00:38:08.120 |
- The 2020 Corolla is the best car with OpenPILOT. 00:38:11.720 |
The actuator has less lag than the older Corolla. 00:38:14.160 |
- I think I started watching a video with your, 00:38:21.920 |
You're just literally at the dealership streaming. 00:38:27.520 |
- Yeah, and basically, like if stuff goes a little wrong, 00:38:39.760 |
to the way other companies would put together 00:38:46.920 |
- And if you become super rich one day and successful, 00:38:50.800 |
because I think that's actually what people love, 00:38:56.960 |
- Money has no, if I sell out to like make money, 00:39:03.400 |
- And I think Tesla's actually has a small inkling 00:39:11.340 |
They did reveal more than, I mean, of course, 00:39:14.080 |
there's marketing communications, you can tell, 00:39:15.760 |
but it's more than most companies would reveal, 00:39:17.720 |
which is, I hope they go towards that direction more, 00:39:27.840 |
You think, you're focused on level two currently. 00:39:43.800 |
I'm not making any timeline on when Tesla's gonna be. 00:39:48.400 |
if you think you're gonna have an autonomous robo-taxi fleet 00:39:59.840 |
are kind of just doing their usual safety driver, 00:40:25.240 |
They're putting real users behind the wheel of the cars. 00:40:34.480 |
- Well, so there's a few components to Tesla approach 00:40:38.800 |
It's what you spoke with is the ones, the software, 00:40:47.280 |
Those differentiate them from the automakers. 00:40:52.000 |
no cars with lane keeping system have that except Tesla. 00:40:55.760 |
- And the other one is the data, the other direction, 00:41:04.520 |
but the ability to turn on collection and turn it off. 00:41:15.040 |
Many people believe that level two autonomy is problematic 00:41:26.040 |
You start to fall asleep, you start to become complacent, 00:41:51.660 |
You can cheat right now, our driver monitoring system. 00:41:59.720 |
Before we ship a consumer product that can drive cars, 00:42:02.580 |
I want to make sure that I have driver monitoring 00:42:05.480 |
- What's like a successful driver monitoring system 00:42:08.200 |
Is it all about just keeping your eyes on the road? 00:42:12.760 |
So that's what we went with at first for driver monitoring. 00:42:24.640 |
And actually a lot of it, just psychology wise, 00:42:30.760 |
it reminds you that you have to be paying attention. 00:42:52.000 |
'cause we're doing just data collection, not real time. 00:42:59.080 |
And the smartphone is really like the biggest problem. 00:43:03.320 |
- I'll show you, I can show you one of the pictures 00:43:09.680 |
the driver monitoring will be the answer to that. 00:43:13.160 |
- I think the other point that you raised in your paper 00:43:17.000 |
You're not asking a human to supervise a machine 00:43:20.480 |
without giving them the, they can take over at any time. 00:43:28.000 |
We don't disengage on steering, I don't feel you have to. 00:43:49.040 |
which no one knows what that is, so they press the brake. 00:43:51.040 |
But a lot of times you don't actually wanna press the brake. 00:43:53.360 |
You wanna press the gas, so you should cancel on gas. 00:43:55.920 |
Or wiggle the steering wheel, which is bad as well. 00:43:58.920 |
I haven't heard anyone articulate that point. 00:44:12.940 |
But you just described that it could be even better. 00:44:15.540 |
- I love Super Cruise as an experience once it's engaged. 00:44:24.040 |
- Yeah, I've used the, I've driven Super Cruise a lot. 00:44:27.520 |
So what's your thoughts on the Super Cruise system in general? 00:44:29.480 |
- You disengage Super Cruise and it falls back to ACC. 00:44:36.280 |
Otherwise, when you actually have Super Cruise engaged 00:44:42.320 |
We just sold it, but we bought it just to experience this. 00:44:45.600 |
And I wanted everyone in the office to be like, 00:44:56.400 |
If there's a sun shining back here, it'll be blind to you. 00:45:03.320 |
- That's so cool that you know all this stuff. 00:45:08.480 |
'cause it's such a rare car, unfortunately, currently. 00:45:34.560 |
and not some weird Cadillac that nobody bought. 00:46:08.240 |
that they're only messing up once every thousand miles, 00:46:17.320 |
One possibility is that without driver monitoring, 00:46:30.640 |
- You've seen all the people sleeping in Teslas? 00:46:51.640 |
are really sleeping in Teslas versus sleeping. 00:47:08.520 |
was I ran into a Model 3 in March and drove it around. 00:47:18.720 |
For Tesla, you have to touch the wheel just enough 00:47:43.720 |
I mean, I think Super Cruise is a better experience 00:47:50.920 |
to engagement and disengagement are significantly worse. 00:48:00.800 |
And I think it's like six seconds or something. 00:48:03.640 |
If you look off road, it'll start warning you. 00:48:19.880 |
It needs to communicate what it sees about you more. 00:48:24.440 |
- If Tesla shows what it sees about the external world, 00:48:27.160 |
it would be nice if Super Cruise would tell us 00:48:33.320 |
and it just says, "Super Cruise unavailable." 00:48:46.280 |
- We have car state packet, which has the state of the car, 00:48:57.320 |
- You think that's possible with computer vision? 00:49:06.520 |
Actually, I quite seriously looked at the literature. 00:49:08.480 |
It's not obvious to me that from the eyes and so on, 00:49:13.480 |
You might need how they're controlling the car, right? 00:49:15.800 |
And that's fundamentally, at the end of the day, 00:49:18.680 |
But I think, especially when people are really drunk, 00:49:21.680 |
they're not controlling the car nearly as smoothly 00:49:35.520 |
So what are the current limitations of OpenPILOT? 00:49:38.280 |
What are the main problems that still need to be solved? 00:49:41.800 |
- We're hopefully fixing a few of them in 0.6. 00:49:49.480 |
So if you're coming up to a red light at like 55, 00:49:56.920 |
which is responsible for two Autopilot accidents, 00:50:07.520 |
There's no way from the radar data to tell the difference. 00:50:10.680 |
but I don't really believe in mapping at all anymore. 00:50:16.920 |
- So you're basically, the OpenPILOT solution is saying, 00:50:24.400 |
- And then eventually when you want to do navigate 00:50:26.200 |
on OpenPILOT, I'll train the net to look at ways. 00:50:35.920 |
We use it to very carefully ground truth the paths. 00:50:45.640 |
in that local part of the environment, but it's all local. 00:50:48.720 |
- How are you testing in general, just for yourself, 00:51:13.680 |
- We can load in a drive and simulate what the system 00:51:26.080 |
but as soon as we start using it for training, that's it. 00:51:30.880 |
What's your feeling about the real world versus simulation? 00:51:35.720 |
- So we have to distinguish two types of simulators, right? 00:51:46.760 |
I feel that this kind of simulator is useless. 00:51:57.000 |
You're not solving the computer vision problem, 00:51:59.940 |
but you're solving the computer graphics problem. 00:52:02.440 |
- Right, and you don't think you can get very far 00:52:08.040 |
- No, because you can create ultra-realistic graphics 00:52:10.360 |
of the road, now create ultra-realistic behavioral models 00:52:18.320 |
You need real, you need actual human behavior 00:52:25.880 |
The definition of driving is what humans do when they drive. 00:52:29.080 |
Whatever Waymo does, I don't think it's driving. 00:52:32.800 |
- Right, well, I think actually Waymo and others, 00:52:36.400 |
if there's any use for reinforcement learning, 00:52:50.080 |
- By that point, you've already gone so many layers. 00:52:55.680 |
Did you hand code the feature vector of their state? 00:53:00.240 |
Did you guys learn anything from computer vision 00:53:07.160 |
- So perception to you is the sticking point. 00:53:10.880 |
I mean, what's the hardest part of the stack here? 00:53:13.800 |
- There is no human understandable feature vector 00:53:32.960 |
there is a human understandable feature vector 00:53:36.000 |
I mean, okay, so I have like three degrees position, 00:53:38.720 |
three degrees orientation and those derivatives, 00:53:42.000 |
That's human understandable, that's physical. 00:53:48.560 |
So like Waymo has a perception stack and then a planner. 00:53:55.600 |
is they have a simulator that can separate those two. 00:54:04.880 |
about like the two different kinds of simulators. 00:54:08.240 |
and there's the kind that can't work on real data. 00:54:28.080 |
- And you're saying that's too difficult to hand engineer. 00:54:40.640 |
All you have to do is define the state vector 00:54:43.960 |
- I'm missing the state vector that separates those two. 00:54:49.420 |
- So what is the output of your perception system? 00:54:57.360 |
there's, okay, well, there's several ways to do it. 00:55:06.680 |
- And then there's the different objects in the scene. 00:55:09.720 |
- And different objects in the scene over time, 00:55:23.180 |
- I can give you a concrete example of something you missed. 00:55:28.600 |
Humans understand that when they see this bush 00:55:30.880 |
that there may or may not be a car behind that bush. 00:55:34.640 |
Drivable area and a list of objects does not include that. 00:55:44.680 |
- Right, but even that, what do you mean by occluded? 00:55:49.600 |
Well, if it's the other side of a house, I don't care. 00:55:58.040 |
I can come up with 10 more examples that you can't add. 00:56:05.520 |
because it's simulating the entire, you know, 00:56:12.640 |
But what I'm saying is if you have a hand-engineered, 00:56:16.580 |
if your perception system output can be written 00:56:22.240 |
- Yeah, I mean, certainly it's hard to argue with that 00:56:33.480 |
- It's a 1024 dimensional vector, trained by a neural net. 00:56:38.840 |
- No, that's the 1024 dimensions of who knows what. 00:56:54.800 |
and you say it has 256 dimensions in the middle, 00:57:02.900 |
Can you hand label all 256 of those dimensions? 00:57:05.460 |
- Well, no, but those are generated automatically. 00:57:23.700 |
if you could design a face reconstructor system, 00:57:28.160 |
- No, but I think we're missing here a little bit. 00:57:35.120 |
about expressing a fundamental problem of simulators, 00:57:47.120 |
from in the simulated world than the real world. 00:57:57.140 |
between perception and planning, even in your system. 00:58:08.780 |
like saying, "Here's a list of all the cars in the scene. 00:58:13.060 |
"Here's a vector of drivable areas," it's insufficient. 00:58:20.900 |
- Currently, what we're doing is the perception problem 00:58:33.460 |
And you're saying that really there's a lot missing there. 00:59:16.700 |
Is that this whole problem, the right way to do it 00:59:21.660 |
- I'll argue that not only is it the right way, 00:59:23.740 |
it's the only way that's gonna exceed human performance. 00:59:33.540 |
And then someone came along and wrote some 10,000 line thing 00:59:36.260 |
that doesn't know anything about Go that beat everybody. 00:59:41.180 |
- True, in that sense, the open question then 00:59:55.420 |
So how, 'cause I think the Elon Musk approach here 01:00:08.300 |
but really do formulate it as a learning problem 01:00:15.780 |
how many years would it take to solve this problem 01:00:52.980 |
You mentioned that you didn't navigate an autopilot. 01:00:59.220 |
if it's done really well, it's a good feature? 01:01:02.360 |
- I think that it's too reliant on hand-coded hacks 01:01:07.360 |
for how does navigate an autopilot do a lane change? 01:01:18.900 |
Navigate an autopilot, at least every time I use it, 01:01:24.260 |
I mean, this is a fundamental thing, actually, 01:01:31.900 |
Tesla probably does it better than most cars, 01:01:36.780 |
of creating a comfortable, natural experience. 01:01:39.940 |
And navigate on autopilot is just lane changes 01:01:44.100 |
So how do you learn to do a natural lane change? 01:01:49.100 |
- So we have it and I can talk about how it works. 01:01:53.020 |
So I feel that we have the solution for lateral, 01:01:58.020 |
we don't yet have the solution for longitudinal. 01:02:00.740 |
There's a few reasons longitudinal is harder than lateral. 01:02:16.460 |
we hand label all the lane changes, 'cause you have to. 01:02:19.620 |
I've struggled a long time about not wanting to do that, 01:02:26.580 |
Oh, we actually, we have an automatic ground truther, 01:02:28.420 |
which automatically labels all the lane changes. 01:02:33.660 |
- Yeah, detect the lane, I see when it crosses it, right? 01:02:34.860 |
And I don't have to get that high percent accuracy, 01:02:39.020 |
- Now I set the bit when it's doing the lane change 01:02:44.900 |
And then I set it to zero when it's not doing a lane change. 01:02:47.980 |
So now if I wanted to do a lane change at test time, 01:02:49.780 |
I just put the bit to a one and it'll do a lane change. 01:02:52.420 |
- Yeah, but so if you look at the space of lane change, 01:02:54.700 |
you know, some percentage, not 100% that we make as humans 01:03:06.980 |
- How do we label the ones that are natural and feel good? 01:03:13.420 |
the current Navigator and Autopilot just doesn't feel good. 01:03:18.500 |
is a hand-coded policy written by an engineer in a room 01:03:21.700 |
who probably went out and tested it a few times on the 280. 01:03:25.060 |
- Probably a more, a better version of that, but yes. 01:03:29.460 |
- That's how we would have written it at the company, yeah. 01:03:36.460 |
Um, no, but, so if you learn the lane change, 01:03:40.060 |
if you learn how to do a lane change from data, 01:03:42.420 |
just like you have a label that says lane change 01:03:44.660 |
and then you put it in when you want it to do the lane change 01:03:55.900 |
we haven't worked too much on this problem yet. 01:04:06.140 |
and all bad drivers are bad in different ways. 01:04:12.380 |
So you just basically, if that's true, hypothesis, 01:04:16.540 |
then your task is to discover the good drivers. 01:04:19.860 |
- The good drivers stand out because they're in one cluster 01:04:23.300 |
and the bad drivers are scattered all over the place 01:04:27.220 |
- Yeah, that's, so you just learn from the good drivers 01:04:36.900 |
But we'll eventually probably have to filter them out. 01:04:41.580 |
'Cause the counter theory is there is many clusters, 01:04:46.460 |
maybe arbitrarily many clusters of good drivers. 01:04:53.700 |
'Cause if there's one cluster of good drivers, 01:05:04.580 |
And you're saying that there is some evidence that-- 01:05:12.060 |
- I would argue that all four of those are good clusters. 01:05:15.260 |
All the things that are random are noise and probably bad. 01:05:25.020 |
- And the hope is it's not too dependent on the driver. 01:05:34.240 |
The hope is that the distribution's not bimodal. 01:05:41.640 |
how to fix, how to improve navigating an autopilot? 01:05:45.000 |
That's the lessons that you've learned from CalmAI. 01:05:50.560 |
is please put driver monitoring in your cars. 01:05:59.280 |
of many of hundreds of thousands of cars being produced 01:06:18.760 |
- You can hope that you can kind of scrape up 01:06:29.560 |
- Every time I've heard Karpathy talk about the problem 01:06:33.280 |
and how the machine learning is gobbling up everything, 01:06:35.240 |
I think this is absolutely the right strategy. 01:06:37.440 |
I think that he didn't write and have it get on autopilot. 01:06:51.080 |
- Well, that's Andre's job is to turn everything 01:06:54.400 |
into a learning problem and collect a huge amount of data. 01:07:01.160 |
can be turned into a learning problem in the short term. 01:07:04.140 |
In the end, everything will be a learning problem. 01:07:07.320 |
The reality is, like if you wanna build L5 vehicles today, 01:07:20.400 |
It's the crutch statement that LIDAR is a crutch. 01:07:29.960 |
classification on driving is the question still. 01:07:35.840 |
I'll say this, I'm here to play for 10 years. 01:07:40.360 |
I'm here to play for 10 years and make money along the way. 01:07:50.600 |
- What do you think was the motivation behind saying that? 01:07:53.560 |
Other companies are also promising L5 vehicles 01:07:56.720 |
with very different approaches in 2020, 2021, 2022. 01:08:03.720 |
that those things do not pan out, I will bet you. 01:08:06.920 |
Even money, even money, I'll bet you as much as you want. 01:08:10.800 |
So are you worried about what's going to happen? 01:08:16.160 |
What's going to happen when 2022, '21 come around 01:08:19.200 |
and nobody has fleets of autonomous vehicles? 01:08:36.120 |
I think not many have declared as like definitively 01:08:45.080 |
Do I think that it's possible for Waymo to continue 01:08:51.000 |
until it gets to level four in Chandler, Arizona? 01:08:58.180 |
- But by, sorry, which year are we talking about? 01:09:02.520 |
- Oh, I even think that's possible by like 2020, 2021. 01:09:10.340 |
- Level four, meaning some very defined streets 01:09:18.320 |
And then practically these streets are pretty empty. 01:09:20.720 |
If most of the streets are covered in Waymos, 01:09:28.920 |
If your self-driving network is the majority of cars 01:09:34.640 |
with respect to each other and all the humans 01:09:42.200 |
- I mean, already you can talk about autonomy 01:09:47.120 |
because you can really just follow the GPS line. 01:09:50.560 |
- So what does success look like for Calm.ai? 01:09:55.560 |
What are the milestones like where you can sit back 01:09:59.040 |
with some champagne and say, "We did it, boys and girls." 01:10:06.120 |
- Yeah, but you must drink champagne every year. 01:10:13.160 |
- A big milestone that we're hoping for by mid next year 01:10:35.320 |
OpenPilot has gotten so much better in the last two years. 01:10:43.780 |
We're gonna disable no safety features in the car. 01:10:49.120 |
Our project this week is we're analyzing the data set 01:10:54.700 |
We have a better data set on that than the manufacturers. 01:10:59.400 |
How much, just how many, does Toyota have 10 million miles 01:11:05.320 |
- So let me give you, 'cause you asked, right, 01:11:29.080 |
and not selling in a way where here at Automaker, 01:11:31.800 |
but creating, we've done this actually at MIT, 01:11:35.480 |
but you could do it for significant money purposes 01:11:37.760 |
and make the world a better place by creating a consortia 01:11:44.240 |
and then they get to have free access to the data. 01:11:46.960 |
And I think a lot of people are really hungry for that 01:11:51.460 |
and would pay a significant amount of money for it. 01:11:56.880 |
It'd be very easy for me to give them access to my servers 01:12:12.560 |
and he's not gonna know most of the words I'm saying. 01:12:21.120 |
I'm the same way, but you just tell them the terms 01:12:43.300 |
I'll give you access to this data subscription. 01:12:46.820 |
Came up with that number off the top of my head. 01:12:50.200 |
where it's like, we would like to pay 100K a month 01:13:01.220 |
The email is gonna be, hey, do you have some time 01:13:03.880 |
in the next month where we can sit down and we can, 01:13:07.880 |
- Yeah, you could politely respond to that email 01:13:10.080 |
by not saying, I don't have any time for your bullshit. 01:13:13.280 |
You say, oh, well, unfortunately, these are the terms, 01:13:19.840 |
in order to minimize the friction, the communication. 01:13:24.520 |
one, two million dollars a year, and you have access. 01:13:32.720 |
Am I gonna hire a business development person 01:13:38.760 |
- If they reached into me, I'm not gonna ignore the email. 01:13:42.480 |
- I'm willing to pay 100K a month for access to the data. 01:13:56.280 |
and they want it, but there'll need to be some communication. 01:14:00.640 |
- Mobileye circa 2015 had the lowest R&D spend 01:14:08.360 |
and you look at all the people who work for them, 01:14:12.120 |
because the car companies are impossible to work with. 01:14:32.260 |
if I showed off the actual learning tech that we have, 01:14:40.140 |
It's years and years ahead of everybody else's. 01:14:43.860 |
I think Tesla has similar stuff to us, actually. 01:15:06.220 |
because you basically don't care about a lot of things 01:15:22.300 |
What I don't necessarily fully can't trust you on, 01:15:25.980 |
with all due respect, is how good it is, right? 01:15:28.420 |
I can only, but I also know how bad others are. 01:15:42.340 |
and try OpenPILOT 0.6 when it comes out next month. 01:15:48.420 |
and you'll be like, this is already really good. 01:15:51.220 |
And then I could be doing that all with hand labelers 01:15:54.260 |
and all with like the same approach that like Mobileye uses. 01:15:57.980 |
When we release a model that no longer has the lanes in it, 01:16:04.980 |
then think about how we did that machine learning. 01:16:14.100 |
you'll know that everything I'm saying is true 01:16:17.660 |
- You know what I'm saying is true about the simulator. 01:16:22.700 |
- But like, I listened to your talk with Kyle 01:16:25.740 |
and Kyle was originally building the aftermarket system 01:16:30.460 |
and he gave up on it because of technical challenges. 01:16:35.140 |
- Because of the fact that he's going to have to support 01:16:43.460 |
We have alerts and warnings to deal with all of that 01:17:10.180 |
You could work with Argo AI, working with Ford. 01:17:13.220 |
It's interesting because chances that you fail, 01:17:17.580 |
business-wise, like bankrupt, are pretty high. 01:17:28.340 |
- Well, I have a long-term way for Commodore to make money too. 01:17:36.740 |
is things you don't expect, ways to make money 01:17:41.020 |
or create value that you don't expect will pop up. 01:17:48.500 |
- Well, which part to know how to do which part? 01:17:50.460 |
- Our long-term plan is to be a car insurance company. 01:17:57.740 |
to know who statistically is the safest drivers. 01:17:59.820 |
And oh, oh, we see you, we see you driving unsafely, 01:18:08.940 |
because the only people who can't get Comm insurance 01:18:11.700 |
Geico can insure them, their premiums are crazy high, 01:18:15.300 |
We'll win car insurance, take over that whole market. 01:18:29.940 |
Who else is thinking like this and working like this 01:18:34.740 |
Are there people seriously, I don't think anyone 01:18:37.540 |
that I'm aware of is seriously taking on lane keeping, 01:18:47.140 |
that then creates, yeah, like that creates other businesses 01:19:07.740 |
Like Ford, once the autonomy doesn't fall through. 01:19:20.860 |
if it was not for those conversations with Elon 01:19:36.060 |
What are your thoughts on his provocative statement 01:19:44.020 |
but sometimes he'll say absolutely, completely, 01:19:58.140 |
If you have to localize your car to centimeters 01:20:06.260 |
on top of LIDAR data, meaning like to help you 01:20:12.820 |
- The main goal of those LIDARs on those cars, 01:20:15.300 |
I think is actually localization more than perception. 01:20:20.900 |
- If you wanna localize to centimeters, you can't use GPS. 01:20:25.140 |
especially if you're under tree cover and stuff. 01:20:30.220 |
I mean in some research they're using it for perception, 01:20:40.500 |
I'm not saying they don't use it for perception, 01:20:54.020 |
You wanna get centimeter accurate localization, 01:21:20.820 |
I think Waymo will be the only one left standing at the end 01:21:24.180 |
with the, given the financial backing that they have. 01:21:28.820 |
I'll say nice things about both Waymo and Cruz. 01:21:34.460 |
- Waymo is by far the furthest along with technology. 01:21:39.380 |
Waymo has a three to five year lead on all the competitors. 01:21:42.940 |
If the Waymo-looking stack works, maybe three year lead. 01:21:52.860 |
Now I argue that Waymo has spent too much money 01:22:00.220 |
Also, self-driving cars have no network effect like that. 01:22:04.860 |
You have a market, you have drivers and you have riders. 01:22:07.180 |
Self-driving cars, you have capital and you have riders. 01:22:11.460 |
If I wanna blanket a new city in self-driving cars, 01:22:13.860 |
I buy the off the shelf Chinese knockoff self-driving cars 01:22:18.380 |
And that's why Uber has a first mover advantage 01:22:32.460 |
I open for business in Austin, Texas, let's say. 01:22:42.820 |
And I need to keep them both sufficiently happy, right? 01:22:46.580 |
if it takes more than five minutes for an Uber to show up. 01:22:50.220 |
if they have to sit around all day and there's no riders. 01:22:54.420 |
And whenever you have to carefully balance a market, 01:22:58.380 |
because there's a switching cost for everybody, right? 01:23:04.180 |
Let's even say that, let's say a Luber shows up 01:23:08.980 |
and Luber somehow agrees to do things at a bigger, 01:23:13.780 |
we're just gonna, we've done it more efficiently, right? 01:23:22.840 |
because the switching cost is higher than that 5%. 01:23:28.980 |
Autonomous vehicles of the level five variety 01:23:36.840 |
say I wanna go to a new city, look at the scooters. 01:23:45.800 |
And that's why you have 10 different scooter companies. 01:23:56.060 |
don't get a say in whether they wanna be bought 01:24:00.120 |
- We're gonna entice the scooters with subsidies and deals. 01:24:03.080 |
- So whenever you have to invest that capital, 01:24:08.720 |
That can't be your main criticism of the Waymo approach. 01:24:12.480 |
- Oh, I'm saying even if it does technically work. 01:24:15.000 |
Even if it does technically work, that's a problem. 01:24:40.820 |
the human-driven cars, in terms of financially. 01:25:01.220 |
because they're gonna stop for every little thing everywhere. 01:25:09.260 |
- What was the nice, oh, 'cause there are three-- 01:25:10.740 |
- Three years technically ahead of everybody. 01:25:19.140 |
For $1 billion, GM bought an insurance policy against Waymo. 01:25:25.620 |
They put, cruise is three years behind Waymo. 01:25:30.000 |
That means Google will get a monopoly on the technology 01:25:39.060 |
you might not even be right about the three years. 01:25:42.700 |
Cruise actually might not be that far behind. 01:25:44.300 |
I don't know how much Waymo has waffled around 01:25:47.340 |
or how much of it actually is just that long tail. 01:25:50.620 |
If that's the best you could say in terms of nice things, 01:26:05.220 |
For cruise to leapfrog Waymo would really surprise me. 01:26:08.520 |
- Yeah, so let's talk about the underlying assumptions 01:26:15.320 |
Tesla would have to seriously mess up for us. 01:26:37.520 |
- I don't know Aurora, Zooks is the same stack as well. 01:26:41.760 |
They're all the same DARPA Urban Challenge code base. 01:26:46.820 |
do you think there's a room for brilliance and innovation 01:26:53.900 |
It could be if revolution and mapping, for example, 01:27:13.080 |
to where all the way you said before becomes incorrect. 01:27:45.720 |
If all you have to deal with is the static problem 01:27:48.420 |
and you can statically schedule your machines, 01:27:50.160 |
it's the same as like statically scheduling processes. 01:28:00.160 |
Maps only helps you with the static driving problem. 01:28:06.960 |
You've just made it sound like it's really easy. 01:28:18.760 |
it's failing on the fundamental static driving problem. 01:28:24.480 |
The static driving problem is not easy for the world. 01:28:27.740 |
The static driving problem is easy for one route. 01:28:37.960 |
and like no deterioration, no cracks in the road. 01:28:45.640 |
- But that's the problem is how do you have a perfect-- 01:28:55.080 |
- With LIDAR, yeah, but you use LIDAR, right? 01:29:07.240 |
- I'm not even concerned about the one or 10 centimeters, 01:29:09.280 |
I'm concerned if every once in a while you're just way off. 01:29:15.960 |
carefully make sure you're always tracking your position. 01:29:21.720 |
but you can get the reliability of that system 01:29:29.680 |
where it's not that bad if you're way off, right? 01:29:33.720 |
it's like as will be that you're never in a case 01:29:42.240 |
- We can, especially with LIDAR and good HD maps, 01:29:51.880 |
- Very typical for you to say something's easy, I got it. 01:29:54.440 |
It's not as challenging as the other ones, okay. 01:29:56.880 |
- Well, it's, okay, maybe it's obvious how to solve it. 01:30:00.680 |
and a lot of people don't even think about the third one 01:30:01.920 |
and even see it as different from the second one. 01:30:05.760 |
The second one is like, say there's an obvious example, 01:30:08.600 |
it's like a car stopped at a red light, right? 01:30:18.000 |
and then you have to do the appropriate action, right? 01:30:20.700 |
Also, that car is not a fixed object, that car may move 01:30:25.740 |
and you have to predict what that car will do, right? 01:30:32.840 |
This involves, again, like you're gonna need models 01:30:42.440 |
Are you including in that your influence on people? 01:31:04.720 |
will scale completely to the static and dynamic. 01:31:07.560 |
The counterfactual, the only way I have to do it yet, 01:31:11.320 |
the thing that I wanna do once we have all of these cars 01:31:13.960 |
is I wanna do reinforcement learning on the world. 01:31:16.760 |
I'm always gonna turn the exploiter up to max, 01:31:20.420 |
but the only real way to get at the counterfactual 01:31:27.760 |
- So that's fascinating that you break it down like that. 01:31:31.720 |
- I've spent my life thinking about this problem. 01:31:34.440 |
Part of it 'cause you're slightly insane is that, 01:31:44.800 |
some non-zero percent of your brain has a madman in it, 01:31:55.920 |
that I think when there's sort of counterfactuals and so on 01:32:00.240 |
How do you even start to think about, just in general, 01:32:03.320 |
I mean, you've had some friction with NHTSA and so on. 01:32:31.560 |
of how do you interact with dynamic objects and so on, 01:32:34.840 |
- You can do reinforcement learning without ever exploring. 01:32:38.080 |
And I said that, so you can think about your, 01:32:48.040 |
I could always set that to zero and still learn, 01:32:50.680 |
and I feel that you'd always want that set to zero 01:32:54.040 |
- Gotcha, but the problem is you first don't know very much, 01:32:59.560 |
So the learning, the exploration happens through mistakes. 01:33:06.080 |
OpenPilot and Autopilot are making mistakes left and right. 01:33:14.080 |
OpenPilot makes tens of thousands of mistakes a week. 01:33:21.160 |
These mistakes are, oh, I wanted to take this exit, 01:33:26.840 |
so I'm just going to carefully touch the wheel. 01:33:36.200 |
- So driver monitoring is the way you ensure they keep-- 01:33:47.840 |
- Oh, I couldn't scale with any amount of money. 01:33:57.000 |
- Better, better. - And I don't know how to-- 01:34:02.600 |
and he basically said that the human factor doesn't matter. 01:34:07.600 |
The human doesn't matter 'cause the system will perform. 01:34:12.320 |
There'll be sort of a, sorry to use the term, 01:34:20.880 |
But it seems like that human catching the system 01:34:29.440 |
which will make something like reinforcement learning work. 01:34:35.680 |
for you, for the industry in general should change? 01:34:39.120 |
- I think our messaging's pretty clear, at least. 01:34:41.880 |
Our messaging wasn't that clear in the beginning, 01:34:45.240 |
We are proud right now to be a level two system. 01:34:53.240 |
It's not gonna be just a magical OTA upgrade. 01:35:03.400 |
I mean, not like, okay, rigorous, who knows what that means, 01:35:31.160 |
Would you like to have an intelligent conversation 01:35:41.600 |
Python not being real-time never causes disengagements. 01:35:44.320 |
Disengagements are caused by, you know, the model is QM. 01:35:50.680 |
the driver must be paying attention at all times. 01:35:53.080 |
I still consider the software to be alpha software 01:35:57.760 |
until we can actually enforce that statement, 01:36:00.120 |
but I feel it's very well-communicated to our users. 01:36:04.560 |
One is the user must be able to easily take control 01:36:10.920 |
So if you step on the gas or brake with OpenPilot, 01:36:14.480 |
it gives full manual control back to the user, 01:36:17.800 |
Step two, the car will never react so quickly, 01:36:32.360 |
like, our torque limit's way lower than Tesla's. 01:36:47.080 |
We limit, and all this code is open source, readable, 01:36:52.080 |
and I believe now it's all MISRA C-compliant. 01:36:55.880 |
- MISRA is like the automotive coding standard. 01:37:05.880 |
They're actually written by very smart people. 01:37:21.140 |
They talk about whether certain conditions in MISRA 01:37:33.120 |
and we want to make our code compliant with that. 01:37:38.260 |
So you were the founder and I think CEO of CalmAI. 01:37:55.240 |
- Oh, so you're like Putin when he was like the, 01:38:07.520 |
Are you just focused on getting things right now, 01:38:11.680 |
making some small amount of money in the meantime, 01:38:14.960 |
and then when it works, it works and you scale? 01:38:24.960 |
But we haven't like tried very hard at that yet. 01:38:28.240 |
- And the revenue is basically selling stuff online. 01:38:42.680 |
We're honest and transparent about what they are. 01:38:49.000 |
'Cause you could easily start blowing up like smoke, 01:39:03.360 |
Yeah, well, I can trade my social capital for more money. 01:39:06.680 |
I did it once, I regret it doing it the first time. 01:39:24.600 |
Have you had some introspection, some soul searching? 01:39:29.280 |
I've come to a pretty stable position on that. 01:39:33.920 |
I realized that I don't want attention from the masses. 01:39:44.000 |
- So are these like Elon Musk type characters? 01:40:24.760 |
I mean, I think about all the times in my life 01:40:32.640 |
I accept that there's a decent chance that I'm wrong. 01:40:37.480 |
having talked to Chris Armisen, Sterling Anderson, 01:40:49.000 |
when everybody thinks it's impossible, that's legit. 01:40:53.880 |
who are like writing the infrastructure of the world, 01:40:55.720 |
like the Linus Torvalds and the Chris Latimer. 01:41:23.360 |
- Do you think he might long-term become a competitor? 01:41:28.880 |
Well, so I think that he has the other right approach. 01:41:32.440 |
I think that right now there's two right approaches. 01:41:35.360 |
One is what we're doing and one is what he's doing. 01:41:37.720 |
- Can you describe, I think it's called Pronto AI. 01:41:43.240 |
- Embark is also doing the same sort of thing. 01:41:47.360 |
so if you're, I can't partner with Honda and Toyota. 01:41:51.880 |
Honda and Toyota are like 400,000 person companies. 01:41:59.480 |
Like I don't think of it like, I don't personify it. 01:42:07.160 |
maybe that has like, some truckers are independent. 01:42:10.320 |
Some truckers drive for fleets with a hundred trucks. 01:42:12.120 |
There are tons of independent trucking companies out there. 01:42:14.980 |
Start a trucking company and drive your costs down 01:42:18.160 |
or figure out how to drive down the cost of trucking. 01:42:23.160 |
Another company that I really respect is Notto. 01:42:33.920 |
If I owned a fleet of cars and I could pay 40 bucks a month 01:42:48.980 |
that is like the business model that I like most respect 01:42:57.880 |
How do we create value today with some of this? 01:43:03.860 |
or full steam ahead on the driver monitoring too, 01:43:06.760 |
which I think actually where the short-term value, 01:43:12.880 |
that everything has to have driver monitoring. 01:43:18.760 |
that to get the experience of it right is not trivial. 01:43:25.420 |
I think the timeout depends on speed of the car, 01:43:35.500 |
it's very different if you don't pay attention 01:43:40.660 |
- And long-term, it should probably learn from the driver 01:43:45.780 |
because that's to do, I watched a lot of video. 01:43:49.540 |
just to analyze how people are using smartphones 01:43:57.940 |
- We haven't watched nearly enough of the videos. 01:44:00.300 |
We have, I got millions of miles of people driving cars. 01:44:02.940 |
- In this moment, I spend a large fraction of my time 01:44:05.900 |
just watching videos because it never fails to learn. 01:44:10.900 |
I've never failed from a video watching session 01:44:15.380 |
In fact, I usually, like when I eat lunch, I'll sit, 01:44:20.660 |
and just watch pedestrians with an eye to understand, 01:44:24.540 |
like from a computer vision eye, just to see, 01:44:30.460 |
And there's so many things that we don't understand. 01:44:34.740 |
- Yeah, it's, I'm trying to always think like, 01:44:46.960 |
- So something that's, from a hacking perspective, 01:44:51.780 |
this is always comes up, especially with folks. 01:45:01.940 |
There are some ethical questions, I think, that arise. 01:45:05.020 |
Maybe you want to, do you think there's any ethical, 01:45:11.300 |
- We have a solution to the trolley problem at ComAI. 01:45:14.060 |
Well, so there is actually an alert in our code, 01:45:19.180 |
We don't know how yet to detect the ethical dilemmas, 01:45:26.660 |
- No, but the trolley problem deserves to be trolled. 01:45:28.740 |
- Yeah, you're, that's a beautiful answer, actually. 01:45:36.660 |
like you can have a kind of discussion about it, 01:45:38.060 |
like when you get someone who's like really like earnest 01:45:40.140 |
about it, because it's the kind of thing where, 01:45:45.580 |
whether we should use a SQL stack or no SQL stack, 01:45:48.320 |
if they're not that technical, they have no opinion, 01:45:50.600 |
but if you ask them what color they want to paint the office, 01:46:43.540 |
Like there's tons of human adversarial examples too. 01:47:01.820 |
- At least that's all they could do was take an exit. 01:47:14.560 |
V2V now relies on both parties getting communication right. 01:47:24.800 |
security is like a special case of safety, right? 01:47:28.440 |
Safety is like we put a little piece of caution tape 01:47:33.480 |
so that people won't walk into it by accident. 01:47:35.520 |
Security is I put a 10 foot fence around the hole 01:47:38.220 |
so you actually physically cannot climb into it 01:47:40.100 |
with barbed wire on the top and stuff, right? 01:47:42.360 |
So like if you're designing systems that are like unreliable 01:47:53.800 |
- And then the local sensor should be hardwired 01:47:55.240 |
and then could somebody hack into your CAN bus 01:47:58.640 |
Yes, but they could do it before Comet AI too, so. 01:48:04.680 |
So do you think teleoperation has a role in any of this? 01:48:09.400 |
So remotely stepping in and controlling the cars? 01:48:27.560 |
- So that's the same argument you're using for V2I, V2V? 01:48:31.100 |
- Well, there's a lot of non-safety critical stuff 01:48:40.840 |
There's a lot of great stuff you can do with V2I. 01:48:43.240 |
Like for example, you can, well, I already have V2, 01:48:56.940 |
but it's not a fundamental fallback for safety. 01:49:03.560 |
that require wireless communication are more than QM, 01:49:20.760 |
What do you think it takes to accomplish great things? 01:49:28.120 |
You know, you need to, in order to accomplish great things, 01:49:33.720 |
And then some people are just insanely working, 01:49:41.400 |
- I think I was trolling in the Siraj interview 01:49:44.880 |
Off camera, right before, I smoked a little bit of weed. 01:49:47.280 |
Like, you know, come on, this is a joke, right? 01:50:09.800 |
I think that if you're working 12 hours a day, 01:50:12.040 |
you should be working smarter and not harder. 01:50:27.200 |
or the trolling, mentioned that one of the things 01:50:30.720 |
you look forward to in the future is AI girlfriends. 01:50:34.320 |
- So that's a topic that I'm very much fascinated by. 01:50:42.000 |
What kind of system do you imagine when you say 01:50:45.440 |
AI girlfriend, whether you were trolling or not? 01:50:49.680 |
And I'm serious about that on both a shallow level 01:51:38.040 |
- I mean, I'll see how nice of the software maps 01:51:41.640 |
I mean, yeah, they sweep a lot of the fascinating, 01:51:44.480 |
really difficult technical challenges under the rug. 01:51:46.480 |
Like assuming it's possible to do the mapping 01:51:52.400 |
- You did it with the little like stim unit on the head. 01:52:37.720 |
When I was 15, I read Eliezer Yudkowsky's early writings 01:52:49.160 |
is going to surpass human intelligence massively. 01:53:01.360 |
Like, why should I even bother to learn anything? 01:53:03.360 |
It's all gonna be meaningless when the machines show up. 01:53:13.160 |
And then I'm like, well, the machines ain't here yet, 01:53:14.720 |
you know, and I seem to be pretty good at this stuff. 01:53:28.800 |
is in a long-term monogamous relationship with a machine. 01:53:35.720 |
if you really truly merge with another machine? 01:53:51.640 |
I've seen couples who've been together for a long time 01:53:56.000 |
and like, I almost think of them as one person, 01:53:58.480 |
like couples who spend all their time together and. 01:54:08.200 |
Like a lot of people imagine it's just an efficient link, 01:54:15.280 |
But it's more, you're saying that there's the same kind 01:54:27.760 |
so you're like, oh, I'm gonna download Wikipedia 01:54:33.320 |
My reading speed is limited by my inner processing loop. 01:54:36.760 |
And to like bootstrap that sounds kind of unclear 01:54:54.040 |
whether the simulations are conscious or not. 01:54:58.200 |
But using those simulations to play out hypothetical futures 01:55:08.680 |
- So on that path to AI of superhuman level intelligence, 01:55:14.680 |
you've mentioned that you believe in the singularity, 01:55:18.640 |
Again, could be trolling, could be not, could be part. 01:55:27.680 |
How many years do you think before the singularity, 01:55:32.120 |
Does that mean fundamental shifts in capabilities of AI? 01:55:41.400 |
So I can buy a human being's worth of compute 01:55:47.720 |
I want like, I think they claim a hundred paid of flops. 01:55:58.600 |
I could buy a stack of like, I buy 1080 TI's, 01:56:03.800 |
And for a million bucks, I can get a human worth of compute. 01:56:08.080 |
But when you look at the total number of flops in the world, 01:56:14.440 |
which goes up very, very slowly with the population, 01:56:17.080 |
and machine flops, which goes up exponentially, 01:56:25.900 |
When most flops in the world are silicon and not biological, 01:56:32.320 |
Like they're now the dominant species on the planet. 01:56:35.520 |
- And just looking at how technology is progressing, 01:56:38.720 |
when do you think that could possibly happen? 01:56:44.440 |
I like 2038 because it's the Unix timestamp rollover. 01:56:51.840 |
So you've said that the meaning of life is to win. 01:57:14.640 |
It may not mean, I was criticized for that in the comments. 01:57:18.280 |
Like, doesn't this guy wanna like save the penguins 01:57:24.920 |
I'm not talking about like I have a yacht or something. 01:57:38.880 |
if you're an intelligent agent and you're put into a world, 01:57:44.840 |
you can go back to like Schmidhuber theories about this, 01:57:47.120 |
is to build a compressive model of the world, 01:57:50.520 |
to build a maximally compressive, to explore the world 01:57:55.680 |
maximizes the derivative of compression of the past. 01:58:10.240 |
that in the future, I might be given a real purpose 01:58:14.680 |
And then at that point, now I know what the game is 01:58:18.280 |
I think right now I'm still just trying to figure out 01:58:26.440 |
you have a lot of uncertainty about the reward function 01:58:32.220 |
- So the purpose is to maximize it while you have 01:58:44.240 |
- What is the, if you believe in the universal prior, 01:58:57.280 |
I wonder what that reward function is for you. 01:59:01.920 |
And I look forward to seeing that in five years 01:59:09.840 |
So I'm happy you exist and I wish you the best of luck.