back to indexMIT 6.S094: Deep Learning
Chapters
0:0 Introduction
8:14 Self-Driving Cars
14:20 Deep Learning
00:00:00.000 |
Thank you everyone for braving the cold and the snow to be here. 00:00:06.200 |
This is 6S094, Deep Learning for Self-Driving Cars. 00:00:12.600 |
And it's a course where we cover the topic of deep learning, 00:00:18.800 |
which is a set of techniques that have taken a leap in the last decade 00:00:24.400 |
for our understanding of what artificial intelligence systems are capable of doing. 00:00:30.800 |
And self-driving cars, which is systems that can take these techniques 00:00:36.600 |
and integrate them in a meaningful profound way 00:00:41.200 |
into our daily lives in a way that transforms society. 00:00:45.800 |
So that's why both of these topics are extremely important and extremely exciting. 00:00:52.400 |
My name is Lex Friedman and I'm joined by an amazing team of engineers 00:00:57.600 |
in Jack Terwilliger, Julia Kindlesberger, Dan Brown, Michael Glazer, 00:01:04.000 |
Lee Ding, Spencer Dodd, and Benedict Jenick, among many others. 00:01:14.400 |
Not just ones that perceive and move about the environment, 00:01:19.600 |
but ones that interact, communicate, and earn the trust and understanding 00:01:26.000 |
of human beings inside the car, the drivers and the passengers, 00:01:32.600 |
the pedestrians and other drivers and cyclists. 00:01:39.200 |
The website for this course, selfdrivingcars.mit.edu 00:01:43.600 |
If you have questions, email deepcars@mit.edu 00:01:51.800 |
For registered MIT students, you have to register on the website. 00:01:57.200 |
And by midnight, Friday, January 19th, build a neural network 00:02:03.800 |
and submit it to the competition that achieves the speed of 65 miles per hour 00:02:11.600 |
It's much harder and much more interesting than last year's 00:02:24.600 |
There's guest speakers that come from Waymo, Google, Tesla. 00:02:31.800 |
And those are starting new autonomous vehicle startups 00:02:50.800 |
For those of you who brave the snow and continue to do so, 00:02:54.600 |
towards the end of the class, there will be free shirts. 00:02:57.800 |
Yes, I said free and shirts in the same sentence. 00:03:07.800 |
There's a lot of updates and we'll cover those on Wednesday. 00:03:11.200 |
It's a deep reinforcement learning competition. 00:03:13.800 |
Last year, we received over 18,000 submissions. 00:03:23.000 |
Not only can you control one car within your network, 00:03:27.800 |
This is multi-agent deep reinforcement learning. 00:03:34.200 |
Second, SegFuse, Dynamic Driving Scene Segmentation competition. 00:03:44.800 |
the kinematics of the vehicle, so the movement of the vehicle, 00:03:52.200 |
For the training set, you're given ground truth labels, 00:04:05.200 |
than the state-of-the-art in image-based segmentation. 00:04:21.800 |
in the physical space, not only must interpret, 00:04:26.600 |
the spatial visual characteristics of a scene. 00:04:29.200 |
They must also interpret, understand, and track 00:04:34.000 |
This competition is about temporal propagation of information, 00:05:01.400 |
where a car knowing nothing is using a monocular camera 00:05:05.000 |
as a single input, driving over 30 miles an hour, 00:05:08.400 |
through a scene it has very little control through, 00:05:24.600 |
This competition will result in four submissions, 00:05:50.800 |
which is using the large-scale naturalistic driving data set, 00:05:54.000 |
we have to train a neural network to do end-to-end steering. 00:05:57.800 |
That takes in monocular video from the forward roadway, 00:06:10.000 |
Tomorrow we'll talk about autonomous vehicles. 00:06:15.800 |
Driving scene understanding, so segmentation. 00:06:29.400 |
that's truly taking huge strides in fully autonomous vehicles. 00:06:33.600 |
They're taking the fully L4, L5 autonomous vehicle approach, 00:07:12.000 |
we are going to talk about the topic of our research, 00:07:16.000 |
and my personal fascination is deep learning, 00:07:32.600 |
He is now the CEO of autonomous vehicle startup Voyage, 00:08:06.600 |
the self-driving car startup that I mentioned, 00:08:11.200 |
that has now partnered with NVIDIA and many others. 00:08:40.200 |
Wide-reaching, because there's one billion cars on the road, 00:09:13.000 |
into the hands of an artificial intelligence system. 00:09:20.200 |
you can Google, first time with Tesla autopilot, 00:09:46.600 |
and the profound, the life-critical nature of it, 00:10:00.400 |
that we cannot escape, considering the human being. 00:10:03.600 |
That autonomous vehicle, must not only perceive, 00:10:06.400 |
and control its movement through the environment, 00:10:27.000 |
an autonomous vehicle is more of a personal robot, 00:10:31.000 |
than it is a perfect perception control system. 00:10:58.800 |
that effectively transfer control to human beings, 00:11:26.200 |
navigating through streets of Boston, is easy. 00:11:33.400 |
and you're late, or you're sick of the person in front of you, 00:11:37.200 |
that you want to go in the opposing lane, and speed up. 00:11:47.400 |
can't escape human nature, they must work with it. 00:11:53.200 |
we'll talk about next week, for cognitive load. 00:11:56.400 |
Where we take the raw, 3D convolutional neural networks, 00:12:00.600 |
take in the eye region, the blinking, and the pupil movement, 00:12:04.600 |
to determine the cognitive load of the driver. 00:12:06.800 |
We'll see how we can detect everything about the driver, 00:12:09.800 |
where they're looking, emotion, cognitive load, 00:12:25.600 |
that it almost requires human level intelligence. 00:12:30.000 |
That the, as I said, two, three, four decade out, 00:12:34.000 |
journey for artificial intelligence researchers, 00:12:37.400 |
to achieve full autonomy, will require achieving, 00:12:40.400 |
solving some of the problems, fundamental problems, 00:12:46.600 |
And, that's something we'll discuss in much more depth, 00:12:53.200 |
for the artificial general intelligence course. 00:12:58.200 |
Ray Kurzweil, Mark Rybert, from Boston Dynamics, 00:13:17.000 |
the human-centered artificial intelligence approach, 00:13:20.200 |
in every algorithmic design, considers the human. 00:13:37.000 |
can handle 90, and increasing percent of the cases. 00:13:59.600 |
Thank you, I didn't know it last year, I know now. 00:14:38.400 |
that learn from data, learn from real world data. 00:15:02.000 |
with the human, and the human robot interaction. 00:15:53.800 |
to be able to do something interesting with it. 00:15:59.400 |
which is most capable and focused on this task. 00:16:26.400 |
is the basic example of image classification. 00:16:39.800 |
higher and higher order representations are formed. 00:16:48.200 |
semantic classification of what's in the image. 00:17:38.600 |
that separates green triangles and blue circles, 00:17:43.000 |
on the left, the task is much more difficult, 00:18:38.200 |
is what we're able to achieve with deep learning. 00:24:48.800 |
from multiple neurons being connected together, 00:25:26.800 |
that essentially memorize patterns in the data. 00:26:36.800 |
And with the unsupervised learning approaches, 00:26:41.800 |
the possibilities of artificial intelligence lie. 00:27:46.800 |
And then there's general purpose intelligence, 00:30:10.800 |
We'll discuss that in a little bit more detail, 00:36:07.800 |
in order, and then to not be able to generalize, 00:38:40.800 |
as opposed to putting the weight on one of the edges. 00:39:00.800 |
because they're useful to some of the competitions, 00:39:07.800 |
to play around with some of these parameters. 00:39:38.800 |
dominating the artificial intelligence community. 00:40:32.800 |
as opposed to having to implement stuff from scratch, 00:41:06.800 |
even as in deep reinforcement learning formulation, 00:41:27.800 |
That's several orders of magnitude less data. 00:42:59.800 |
there is a lot of different kinds of objects, 00:43:01.800 |
for cats, dogs, cars, bicyclists, pedestrians.