back to indexKarl Iagnemma & Oscar Beijbom (Aptiv Autonomous Mobility) - MIT Self-Driving Cars
Chapters
0:0 Introduction to Karl Iagnemma and Oscar Beijbom
1:0 Karl - Aptiv Background
10:18 Dimensions of Safety for AVs
12:47 Trusting neural networks behind the wheel
15:7 Validation of black-box systems
17:50 Trusting the data
19:27 Trusting the algorithms
22:27 Safety architecture for neural networks
25:20 Engineering is inching closer to the natural sciences
25:57 Oscar - DL for 3D Detection
30:6 PointPillars
39:51 nuScenes - a dataset for multimodal 3d object detection
43:17 Q&A
00:00:12.160 |
Carl is the president of Aptiv Autonomous Mobility, 00:00:18.080 |
Carl founded Neutonomy, as many of you know, in 2013. 00:00:22.280 |
It's a Boston-based autonomous vehicle company, 00:00:33.000 |
in autonomous vehicle development and deployment, 00:00:36.080 |
with cars on roads all over the United States, 00:00:39.440 |
But most importantly, Carl is MIT through and through, 00:00:43.180 |
as also some of you may know, getting his PhD here. 00:01:05.160 |
Very impressed that you guys are here during IAP. 00:01:08.040 |
My course load during IAP was usually ice skating, 00:01:13.080 |
and sometimes there was a wine tasting course. 00:01:22.660 |
so I'm gonna do my best and try something radical, actually. 00:01:25.520 |
Since I'm president now of Aptiv Autonomous Driving, 00:01:53.200 |
Some of the work that he and his outstanding team 00:01:55.180 |
have been doing around machine learning-based detectors 00:02:03.560 |
So let me first introduce Aptiv a little bit, 00:02:11.000 |
Aptiv's actually been around for a long time, 00:02:34.160 |
Essentially, they take software and hardware, 00:02:39.880 |
so it can run for many, many hundreds of thousands of miles 00:02:48.080 |
they develop what they say is safer, greener, 00:02:55.520 |
autonomous driving systems of the type that we're building. 00:03:03.720 |
And then more connected connectivity solutions, 00:03:23.580 |
when you think about the future of autonomous driving. 00:03:32.780 |
The biggest my research group ever was at MIT 00:03:51.140 |
so of which Oscar is one very important person. 00:03:53.640 |
We're about 700 working on autonomous driving. 00:04:01.480 |
But first, let me take a trip down memory lane 00:04:10.400 |
kind of as a community, but also me personally. 00:04:17.880 |
The fact is 2007, there were groups driving around 00:04:22.140 |
with cars like running blade servers in the trunk 00:04:35.740 |
But people did enough algorithmically, computationally, 00:04:53.920 |
that given enough devotion of thought and resources 00:04:57.960 |
that this might actually become a real thing someday. 00:05:00.860 |
So I was one of those people that got convinced. 00:05:03.940 |
2010, this is now, I'm gonna crib from my co-founder Emilio 00:05:09.560 |
who was a former MIT faculty member in AeroAstro. 00:05:12.320 |
Emilio started up an operation in Singapore through Smart, 00:05:19.000 |
That's James, who looks really young in that picture. 00:05:35.300 |
2014, they did a demo where they let people of Singapore 00:05:39.020 |
come and ride around in these carts in a garden, 00:05:42.240 |
and that worked great over the course of a weekend. 00:05:48.080 |
We'd actually started a commercial enterprise. 00:06:05.920 |
so I couldn't actually accompany people on rides. 00:06:09.820 |
We ended up switching cars to a Renault Zoe platform, 00:06:17.520 |
open to the public rides in our cars in Singapore 00:06:21.040 |
in the part of the city that we were allowed to operate in. 00:06:28.840 |
the evolution of these systems has come a long way 00:06:31.480 |
in a short time, and we're just a point example 00:06:34.340 |
of this phenomenon, which is kind of, broadly speaking, 00:06:39.720 |
But 2017, we joined Aptiv, and we were excited by that 00:06:43.840 |
because we, as primarily scientists and technologists, 00:06:48.560 |
how we were gonna industrialize this technology 00:06:51.760 |
and make it reliable and robust and make it safe, 00:06:55.120 |
which is what I'm gonna talk about a little bit here today. 00:06:57.600 |
So we joined Aptiv with its global footprint. 00:07:04.960 |
and we've got connectivity to Aptiv's other sites 00:07:28.340 |
We've got 30 of those cars on the Lyft network. 00:07:33.720 |
So if you go to Vegas and you open your Lyft app, 00:07:36.320 |
it'll ask you, do you wanna take a ride in an autonomous car? 00:07:39.640 |
You can opt in, you can opt out, it's up to you. 00:07:43.520 |
one of our cars will pick you up if you call for a ride. 00:07:55.600 |
If you take a ride, when you get out of the car, 00:07:58.800 |
you gotta give us a star rating, one through five. 00:08:00.920 |
And that, to us, is actually really interesting 00:08:13.800 |
and the efficiency of getting to where you wanted to go. 00:08:16.720 |
Our star rating today is 4.95, which is pretty good. 00:08:24.280 |
over 30,000 rides to more than 50,000 passengers. 00:08:30.160 |
and a little bit additional, but primarily there. 00:08:46.240 |
You'll see a sped up, slightly sped up view of a run from, 00:09:04.600 |
But to give you an example of some of the types of problems 00:09:11.660 |
And you'll see as this car is cruising down the road, 00:09:26.340 |
where other road users are maybe not perfectly behaving 00:09:33.100 |
Construction in Singapore, like everywhere else, 00:09:51.140 |
So typical day, a route that any one of us as humans 00:09:55.060 |
would drive through without batting an eye, no problem, 00:09:58.340 |
is actually presents some really, really complex problems 00:10:05.780 |
These are the things you have to do if you want to be 00:10:07.300 |
on the road, and certainly if you want to drive 00:10:17.500 |
So let me talk about, we're gonna talk about learning 00:10:54.700 |
And the question of safety, convincing ourselves 00:10:59.660 |
even if we could train it to accurately approximate 00:11:13.140 |
And I'll raise some of the issues around why that is. 00:11:19.860 |
are not incredibly useful for autonomous driving 00:11:23.580 |
And Oscar will show you examples of why that is 00:11:25.900 |
and how Aptiv is using some learning methods today. 00:11:34.620 |
One is the actual technical safety of the system, 00:11:36.860 |
which is to say, can we build a system that's safe, 00:11:41.860 |
that we can validate, which we can convince ourselves, 00:11:49.580 |
that adheres to whatever regulatory requirements 00:11:52.980 |
might be imposed on our jurisdictions that we're operating. 00:11:56.220 |
And there's a whole longer list related to technical safety. 00:12:08.300 |
even if it's safe, do you believe that it's safe? 00:12:11.900 |
And therefore, will you wanna take another trip? 00:12:16.300 |
and as engineers, we're typically uncomfortable 00:12:24.860 |
And quite obviously, we gotta sit up here, right? 00:12:32.060 |
but one that feels safe, that inspires confidence 00:12:34.940 |
in riders, in regulators, and in everybody else. 00:12:40.940 |
of elements of this system that may be black boxes, 00:13:07.140 |
the operational requirements in the domain of interest 00:13:12.660 |
There's three dimensions to this key question 00:13:19.820 |
some topics of interest around each of these. 00:13:32.260 |
I mean, everybody knows garbage in, garbage out. 00:13:35.020 |
There's various ways that we can make this garbage. 00:13:38.340 |
We can have data which is insufficiently covering our domain, 00:13:47.300 |
who we've trusted to label certain things of interest. 00:14:15.740 |
we're trying to approximate really complicated functions. 00:14:19.020 |
I don't think we typically use neural networks 00:14:35.180 |
So there's these events that happen very, very infrequently 00:14:48.660 |
So these are the sorts of things that we think about 00:14:51.740 |
and that we have to answer in an intelligent way 00:15:00.060 |
Okay, let me just step through each of these topics 00:15:11.300 |
There's a number of different dimensions here. 00:15:18.860 |
The underlying phenomena is really complicated. 00:15:21.580 |
Again, if it weren't, we'd probably possibly be modeling it 00:15:25.940 |
We'd write a closed-form equation to describe it. 00:15:47.060 |
If you believe Rand, and they're pretty smart folks, 00:16:01.820 |
Can we think about using some correlated incident, 00:16:08.180 |
as a proxy for accidents, which may be more frequent, 00:16:14.020 |
which I won't say we don't have any answers to, 00:16:24.580 |
The regulatory dimension is one of these known unknowns. 00:16:31.700 |
if the requirements that may be imposed upon us 00:16:34.100 |
from outside regulatory bodies are still to be written? 00:16:42.680 |
on what the safety target should be for these systems. 00:16:56.020 |
or anywhere in between, what this target needs to be. 00:16:58.820 |
And then lastly, and this is a really interesting one, 00:17:03.880 |
we can get through a validation process for a bill to code. 00:17:08.360 |
Well, what happens when we wanna update the code? 00:17:11.860 |
Does that mean we have to start that validation process 00:17:14.820 |
which will unavoidably be expensive and lengthy? 00:17:18.520 |
Well, what if we only change a little bit of the code? 00:17:22.020 |
But what if that one line is the most important line of code 00:17:33.260 |
And then not even, again, keep that code base fixed. 00:17:39.620 |
to your previous city, but not exactly the same. 00:17:46.280 |
So this continuous development issue is a challenge. 00:17:50.780 |
All right, let me move on to talking about the data. 00:17:59.360 |
But there's a couple of obvious questions, I would say, 00:18:02.780 |
that we think about when we think about data. 00:18:09.940 |
for one reason or another, we won't have a great output. 00:18:12.800 |
So one thing we think about is the sufficiency, 00:18:19.100 |
and the bias that may be inherent in the data 00:18:26.900 |
and we only train on data collected during daytime, 00:18:31.580 |
Annotating the data is another dimension of the problem. 00:18:38.420 |
that covers our space, but when we annotate it, 00:18:49.340 |
but we put a lot of trust in that third party, 00:18:58.880 |
that they're gonna catch all the interesting parts 00:19:02.640 |
So this annotation part, which seems very mundane, 00:19:06.500 |
very easy to manage, and kind of like low-hanging fruit, 00:19:16.020 |
Okay, and this reference just kind of points to the fact 00:19:31.060 |
So moving on from the data to the actual algorithm, 00:19:34.800 |
how do we convince ourselves that that algorithm, 00:19:38.420 |
that like any kind of learning-based algorithm, 00:20:06.260 |
If we make small perturbations to this function, 00:20:17.640 |
You know, I'm sure you've all heard of examples 00:20:31.060 |
You show it a turtle, it says, "Well, that's a turtle." 00:20:33.460 |
And then you show it a turtle that's maybe fuzzed 00:20:35.180 |
with a little bit of noise that the human eye can't perceive. 00:20:44.860 |
we want a stop sign to be correctly identified 00:20:51.380 |
if somebody goes up and puts a piece of duct tape 00:20:54.220 |
to be interpreted as a yield sign, for example. 00:20:58.000 |
So this question of the properties of the algorithm, 00:21:08.460 |
And then lastly, to add one more point to this, 00:21:14.620 |
So interpretability, understanding why an algorithm 00:21:19.500 |
This is the sort of thing that may not be a nice-to-have, 00:21:34.580 |
your trajectory generator was a data-driven system, 00:21:38.620 |
was a deep-learning-based trajectory generator. 00:21:56.580 |
and doing active research into this specific question 00:22:18.580 |
be able to explain why algorithm X made decision Y. 00:22:34.180 |
You think, well, man, why are we taking this course 00:22:35.820 |
with Lex, 'cause we're never gonna really use this stuff. 00:22:49.820 |
within the context of a broader safety argument. 00:22:53.940 |
We tend not to think about using a neural network 00:23:07.980 |
maybe make more rigorous claims about their performance, 00:23:14.900 |
holistic safety argument that this end-to-end system is safe. 00:23:31.460 |
we're basically asking ourselves the question, 00:23:34.100 |
is this overall function doing what it's intended to do? 00:23:48.860 |
around functional safety and SOTIF affirmatively, 00:23:52.720 |
even when we have neural network-based elements 00:23:57.020 |
in order to eventually put this car on the road. 00:24:00.260 |
All right, so I mentioned that we need to do some embedding. 00:24:02.940 |
This is an example of what it might look like. 00:24:14.360 |
And in this case, it's up there at the top in red. 00:24:17.340 |
That might be that trajectory proposer I was talking about. 00:24:21.360 |
So let's say we've got a powerful trajectory proposer. 00:24:24.740 |
We've got it on what we call our performance compute, 00:24:31.460 |
but it's generally speaking, good performance. 00:24:35.300 |
And we've got our neural network-based generator on it, 00:24:41.280 |
Well, we make the argument that if we can surround that, 00:25:08.140 |
So I'm gonna conclude my part of the talk here, 00:25:10.420 |
hand it over to Oscar, with kind of a quote, an assertion. 00:25:20.460 |
Engineering is inching closer to the natural sciences. 00:25:24.920 |
We're creating things that we don't fully understand, 00:25:27.460 |
and then we're investigating the properties of our creation. 00:25:30.320 |
We're not writing down closed-form functions. 00:25:38.100 |
functional approximators, and then we're just poking at 'em 00:25:42.020 |
what does this thing do under these situations? 00:26:02.120 |
I run the machine learning team at Aptiv Neutronomy. 00:26:08.920 |
You know, not long ago, image classification was, 00:26:28.660 |
And it was, you know, it's funny 'cause it's true. 00:26:36.340 |
and Lex, you know, previous introductory slides 00:26:54.580 |
end-to-end learning for image net classification 00:27:14.700 |
when the visual community really started seeing, 00:27:28.620 |
that was sort of the final argument for like, 00:27:51.180 |
you know, performance is no longer a joke, right? 00:27:54.980 |
So this is a network that we developed in my group. 00:27:59.460 |
So it's a joint image classification segmentation network. 00:28:04.460 |
This thing, we can run this at 200 hertz on a single GPU. 00:28:20.300 |
And you can see that we can model several different classes, 00:28:23.300 |
you know, both boxes and the surfaces at the same time. 00:28:29.480 |
Here's my cartoon drawing of a perception system 00:28:33.660 |
So you have the three different main sensibilities. 00:28:38.000 |
Typically have some module that does detection and tracking. 00:28:41.300 |
You know, there's tons of variations of this, of course, 00:28:46.600 |
and then in the end, you have a tracking and fusion step. 00:28:55.140 |
but it's like going from the camera to detections. 00:29:04.240 |
learning community, so when I started looking 00:29:06.980 |
at this pipeline, I'm like, why are there so many steps? 00:29:15.540 |
It's a very well-defined input/output function. 00:29:18.620 |
And like Carl alluded to, it's one that can be verified 00:29:22.460 |
quite well, assuming you have the right data. 00:29:49.380 |
it's input, and we're gonna have a neural network 00:29:54.980 |
3D bounding boxes that are in a well-coordinated system. 00:30:01.780 |
this rotation and this orientation and so on. 00:30:13.100 |
and new scenes, which is a benchmark data that we released. 00:30:31.620 |
by a similar architecture that you would use in an image. 00:30:45.860 |
And there's a pre-printout and some code available 00:31:06.880 |
you have this encoder, and that's where we introduce 00:31:10.860 |
I'll show you guys, you can have various types of encoders. 00:31:14.540 |
And then after that, that feeds into a backbone, 00:31:16.600 |
which is now a standard convolutional 2D backbone. 00:31:19.780 |
You have a detection head, and you might have, 00:31:22.220 |
you may or may not have a segmentation head on that. 00:31:26.620 |
everything looks just like, the architecture's 00:31:32.540 |
So let's go into a little bit more detail, right? 00:31:55.900 |
or in the pillar here is a vertical column, right? 00:31:59.100 |
So you have N, M of those that are non-empty in the space. 00:32:03.060 |
And you say a pillar P contains all the points, 00:32:05.500 |
which are a lot of point X, Y, C, and intensity. 00:32:09.100 |
And there's N sub, M indexed by M points in each pillar, 00:32:16.940 |
So it could be one single point at a particular location, 00:32:23.220 |
And the goal here is to produce a tensor as a fixed size. 00:32:29.780 |
range of a resolution, width, range of a resolution, 00:32:41.400 |
We call it a pseudo-image, but it's the same thing. 00:32:47.140 |
Yeah, so here's the same thing without math, right? 00:32:52.860 |
So you have a lot of points, and then you have this space 00:32:55.140 |
where you just grid it up in these pillars, right? 00:33:02.900 |
let me give a little bit of a literature review. 00:33:05.900 |
What people tend to do is you take each pillar, 00:33:14.460 |
For example, how many points are in this voxel? 00:33:20.740 |
Then you extract features for the whole pillar, right? 00:33:23.580 |
What is the max intensity across all the points 00:33:33.820 |
So what you can do is you can now concatenate them, 00:33:40.940 |
So then, VoxelNet came around, I'd say, a year or so ago. 00:33:51.740 |
So they do the first, the first step is similar, right? 00:33:56.520 |
and then you take, you map the points in each voxels. 00:34:10.300 |
And I'm not gonna get into the details of a PointNet, 00:34:18.820 |
and map it to, again, a fixed length representation. 00:34:21.920 |
So it's a series of 1D convolutions and max pooling layers. 00:34:32.340 |
but now I end up with this awkward four-dimensional tensor 00:34:42.900 |
So then they have to consolidate the Z dimension 00:34:52.060 |
So it's very nice in the sense that it's end-to-end method. 00:35:05.780 |
It's much, much slower than a standard 2D convolution. 00:35:12.800 |
We basically said, let's just forget about voxels. 00:35:22.540 |
So just that single change gave a 10- to 100-fold speedup 00:35:37.780 |
So we simplified it to a single 1D convolution 00:35:41.400 |
And then we showed you can get a really fast implementation 00:35:45.380 |
by taking all your pillars that are not empty, 00:35:48.140 |
stack them together into a nice, dense tensor 00:35:52.620 |
And then you can run the forward pass with a single, 00:36:00.460 |
So the final encoder runtime is now 1.3 milliseconds, 00:36:14.300 |
you have this pillar feature net, which is the encoder. 00:36:28.700 |
but of course the key is that none of the steps are, 00:36:34.780 |
And we can back propagate through the whole thing 00:36:43.500 |
these were the results we got on the Qt Benchmark. 00:36:53.580 |
so this is I think the bird's eye view metric. 00:37:00.420 |
And we did that running at a little bit over 60 hertz. 00:37:05.780 |
And this is, like I said, this is in terms of bird's eye view 00:37:17.780 |
and we get the same, very similar performance. 00:37:23.580 |
Yeah, so, you know, car did well, cyclist did well, 00:38:04.660 |
You have the, you know, just for visibilization 00:38:08.460 |
So you see the gray boxes are the ground truth 00:38:22.700 |
So we have, for example, the person on the right there, 00:38:38.260 |
Here's a child on a bicycle that didn't get detected. 00:39:08.700 |
Input is still alive, you know, a few lighter sweeps, 00:39:12.180 |
but just projected into the images for visualization. 00:39:17.660 |
And again, no tracking or smoothing applied here. 00:39:19.740 |
So it's every single frame is analyzed independently. 00:39:32.260 |
yeah, you can actually accumulate multiple point clouds 00:39:37.940 |
and now you can start reasoning about velocity as well. 00:39:42.980 |
So the second part I want to talk about is NuScenes, 00:39:52.660 |
which is a new benchmark data set that we have published. 00:40:01.820 |
that we collected with our development platform. 00:40:05.140 |
So it's a full, it's the same platform that Carl showed, 00:40:08.740 |
or a sort of previous generation platform, the Zoe vehicle. 00:40:12.060 |
So it's full, you know, the full automotive sensor suite, 00:40:15.660 |
data is registered and synced in 360 degree view. 00:40:20.620 |
And it's also fully annotated with 3D bounding boxes. 00:40:22.860 |
I think there's over one million 3D bounding boxes. 00:40:27.020 |
And we actually make this freely available for research. 00:40:32.820 |
and download a teaser release, which is 100 scenes, 00:40:40.380 |
And of course the motivation is straightforward, right? 00:40:44.940 |
So, you know, the whole field is driven by benchmark, 00:40:48.060 |
and you know, without image, I don't think none of it, 00:41:08.580 |
They don't have full 3D view, they don't have any radar. 00:41:15.980 |
to sort of push the field forward a little bit. 00:41:25.980 |
And really the only one that you can really compare to 00:41:30.860 |
But so there's other data sets that have maybe LIDAR only, 00:41:35.540 |
tons of data sets that have image only, of course. 00:41:46.180 |
So you see the layout with the radars along the edge, 00:41:51.100 |
all the cameras on the roof and the top LIDAR, 00:41:59.160 |
The taxonomy, so we model several different subcategories 00:42:16.640 |
So this is one of the thousand scenes, right? 00:42:18.820 |
So all I'm showing here is just playing the frames 00:42:48.060 |
so you guys can play around with it how you like. 00:42:59.780 |
So here I'm showing just colored by distance. 00:43:02.540 |
So now you have some sort of sparse density measurement 00:43:29.580 |
that your model is going to generalize to unseen data 00:43:33.980 |
or do you have other validation that you need to do? 00:43:36.940 |
so the new scenes effort is purely an academic effort. 00:43:40.100 |
So we wanna share our data with academic community 00:44:01.200 |
One of the hardest things was always collecting data 00:44:28.120 |
So not totally, a little bit of self-interest there. 00:44:34.380 |
To give you a sense of the scale of validation, 00:44:39.020 |
saying you gotta drive 275 million miles or more, 00:44:56.900 |
under over hundreds of different bills of code 00:45:02.100 |
you're supposed to drive hundreds of millions of miles 00:45:04.560 |
in a particular environment on a single bill of code, 00:45:08.980 |
Now obviously we're probably not gonna do that. 00:45:11.020 |
What we'll end up doing is supplementing the driving 00:45:15.820 |
and then other methodologies to convince ourselves 00:45:20.260 |
ultimately a statistical argument for safety. 00:45:30.140 |
and other kind of morally equivalent versions 00:45:46.900 |
So the safety case is really quite a bit broader, 00:45:56.340 |
what do you think can 5G offer for autonomous vehicles? 00:46:08.980 |
Certainly when you think about operating them as a fleet. 00:46:12.620 |
When the day comes when you have an autonomous vehicle 00:46:16.900 |
and that day will come in some point in the future, 00:46:24.380 |
and you wanna coordinate the activity of that fleet 00:46:27.380 |
in a way to maximize the efficiency of that network, 00:46:33.180 |
The requirements of that connectivity is fairly relaxed 00:46:35.740 |
if you're talking about just passing back and forth 00:46:37.380 |
the position of the car and maybe some status indicators. 00:46:40.500 |
You know, are you in autonomous mode, manual mode, 00:46:42.280 |
are all systems go, or do you have a fault code, 00:46:49.580 |
if you think about what we call teleoperation 00:46:53.900 |
The case where if the car encounters a situation 00:47:09.380 |
There may be a demand of high bandwidth, low latency, 00:47:21.420 |
Broadly speaking, we see it as a very nice to have, 00:47:39.700 |
such that we don't rely on kind of the coming 5G wave, 00:47:43.260 |
but we'll certainly welcome it when it arrives. 00:47:45.220 |
- So you said you have presence in 45 countries. 00:47:48.260 |
So did you observe any interesting patterns from that? 00:47:51.140 |
Like your car, your same self-driving car model 00:47:55.500 |
that is deployed in Vegas as well as Singapore 00:48:01.180 |
or the model was able to perform very well in Singapore 00:48:15.540 |
I mean, you're on the other side of the road for starters, 00:48:20.900 |
and it's sort of underappreciated people drive differently. 00:48:38.580 |
which tries to admit in a general and fairly flexible way 00:48:44.020 |
the ability to reprioritize rules, reassign rules, 00:48:47.300 |
change weights on rules to enable us to drive 00:48:55.020 |
when we wanted to get on the road in Singapore, 00:49:02.100 |
who was tasked with writing the decision-making engine 00:49:03.820 |
and you decide I'm gonna do a finite state architecture, 00:49:07.860 |
I'm gonna do them by hand, it's gonna be great. 00:49:09.820 |
And then you did that for the right-hand driving 00:49:13.100 |
"Oh yeah, next Monday we're gonna be left-hand driving, 00:49:15.020 |
"so just flip all that and get it ready to go." 00:49:21.620 |
'cause it's generally speaking you're doing it manually 00:49:33.580 |
we actually quite carefully designed the system 00:49:36.940 |
such that we can scale to different cities and countries. 00:49:41.780 |
And one of the ways you do that is by thinking carefully 00:49:51.860 |
There's four cities I mentioned which are our primary sites, 00:49:58.760 |
I mean, everybody knows Boston, which is pretty bad. 00:50:02.700 |
Vegas is warm weather, mid-density urban, but it's Vegas. 00:50:21.860 |
So that exposure to this different spectrum of data, 00:50:25.340 |
I think, I'll speak for Oscar, maybe, is pretty valuable. 00:50:27.900 |
I know for other parts of the development team, 00:50:34.820 |
So every time we drive out, there's a new construction zone. 00:50:47.680 |
People don't break the rules, but they jaywalk. 00:50:55.880 |
Well, it's interesting because there's other dimensions. 00:50:59.120 |
So when we look at which countries are interesting to us 00:51:21.560 |
So it's really all of these things put together. 00:51:28.420 |
and assign them scores and then try to understand 00:51:36.020 |
but there's no one using mobility services there. 00:51:38.780 |
There's no opportunity to actually generate revenue 00:51:45.160 |
- Yeah, and I think, I mean, one thing to keep in mind 00:51:48.100 |
that it's always the first thing I tell candidates 00:51:54.100 |
to the business model we're proposing, right? 00:51:57.260 |
So we can choose, even if we commit to a certain city, 00:52:00.180 |
we can still select the routes that we feel comfortable 00:52:03.540 |
and we can roll it out sort of piece by piece. 00:52:37.100 |
And in the Voxels, you had a four-dimensional tensor 00:52:40.900 |
that you were starting with, and your point pillar, 00:52:43.900 |
You're throwing away the Z, as I understood it. 00:52:49.100 |
that you're losing information about potential occlusions 00:52:57.380 |
So I may have been a little bit sloppy there. 00:53:15.900 |
I felt the need to spoon-feed the network a little bit 00:53:26.940 |
where we learn to consolidate that into a single vector. 00:53:30.820 |
We just said, why not just learn those things together? 00:53:38.060 |
You mentioned that if people make change to the code, 00:53:51.180 |
So when we make any change to our simulation code, 00:54:04.660 |
So in your opinion, do you think for self-driving, 00:54:08.260 |
we need another third-party validation committee or not? 00:54:14.140 |
Or should that be a third party, or is just self-check? 00:54:24.140 |
I wouldn't be surprised, let me put it this way. 00:54:29.340 |
with third-party regulatory oversight, or it didn't. 00:54:34.700 |
There's great precedence for what you just described. 00:54:45.780 |
they can impose strict regulation, or advise regulation, 00:54:58.880 |
The automotive industry has largely been self-certifying. 00:55:01.680 |
There's an argument, which is certainly not unreasonable, 00:55:28.580 |
where the federal government is actually moving 00:55:38.660 |
that has never generated a dollar of revenue. 00:55:43.620 |
But if you would have told me a few years ago 00:55:45.680 |
that there would have been very thoughtfully defined 00:55:55.540 |
around this industry, I probably wouldn't have believed you. 00:55:59.180 |
There's a third version that was released this summer 00:56:04.300 |
So there's intense interest on the regulatory side. 00:56:18.740 |
Looking at this slide, I'm wondering how easy 00:56:29.300 |
if it is snowing, do you need specific trainings 00:56:32.780 |
for specifically for your lidars to work effectively 00:56:41.700 |
to this method as any other machine learning-based method. 00:56:54.120 |
I do like, one thing I like after having worked 00:56:57.680 |
so much with vision though is that the lidar point cloud 00:57:00.220 |
is really easy to augment and play around with. 00:57:07.720 |
you wanna be robust in really rare events, right? 00:57:15.560 |
But it's hard because I have very few examples 00:57:19.200 |
Now if you think about augmenting your visual dataset 00:57:27.480 |
But it is quite easy to do that in your lidar data, right? 00:57:37.040 |
fairly realistic point cloud return from that, right? 00:57:40.520 |
So I like that part about working with lidar. 00:57:42.080 |
You can augment, you can play around with it. 00:57:44.360 |
In fact, one of the things we do when we train this model 00:57:59.420 |
you can just paste it into your current lidar sweep. 00:58:08.500 |
And we found that that was a really useful data. 00:58:14.420 |
And it speaks to the ability to do that with lidar point cloud.