back to indexGreg Brockman: OpenAI and AGI | Lex Fridman Podcast #17
Chapters
0:0
0:1 Greg Brockman
4:45 Technological Determinism
7:0 Wikipedia
13:55 Technical Safety
16:18 Policy Team
18:45 History of Ai
21:44 Generality
22:3 Competence
24:58 Formation of Open Ai
31:29 The Startup Mindset
40:28 Is this Idea that It'D Be Great if You Could Try To Describe or Untangle Switching from Competition to Collaboration and Late-Stage Agi Development It Was Really Interesting this Dance between Competition and Collaboration How Do You Think about that Yeah Assuming You Can Actually Do the Technical Side of Agi Development I Think There's Going To Be Two Key Problems with Figuring Out How Do You Actually Deploy It Make It Go Well the First One of these Is the Run-Up to Building the First Agi You Look at How Self-Driving Cars Are Being Developed
47:1 Then There Was a Bunch of Conversation Where Various People Said It's So Obvious that You Should Have Just Released It There Other People Said It's So Obvious You Should Not Have Released It and I Think that that Almost Definitionally Means that Holding It Back Was the Correct Decision Right if It's Contra if There's if It's Not Obvious whether Something Is Beneficial or Not You Should Probably Default to Caution and So I Think that the Overall Landscape for How We Think about It Is that this Decision Could Have Gone either Way There Are Great Arguments in both Directions but for Future Models down the Road and Possibly Sooner than You'D Expect because You Know Scaling these Things Up Doesn't Have To Take that Long those Ones but You'Re Definitely Not Going To Want To Release into the Wild
77:33 The Reasoning Team
80:28 Simulation for Self-Driving Cars
00:00:00.000 |
The following is a conversation with Greg Brockman. 00:00:07.440 |
developing ideas in AI with the goal of eventually 00:00:24.480 |
Their mission is a catalyst for an important public discourse 00:00:35.820 |
Artificial Intelligence Podcast at MIT and beyond. 00:00:39.520 |
If you enjoy it, subscribe on YouTube, iTunes, 00:00:48.060 |
And now, here's my conversation with Greg Brockman. 00:00:51.820 |
So in high school, and right after you wrote a draft 00:01:01.420 |
So it's clear you have an intuition and a passion 00:01:11.420 |
with AI, deep learning, reinforcement learning, and so on. 00:01:15.400 |
Do you see the physical world and the digital world 00:01:17.380 |
as different, and what do you think is the gap? 00:01:20.540 |
- A lot of it actually boils down to iteration speed. 00:01:23.380 |
That I think that a lot of what really motivates me 00:01:26.540 |
Just the, you know, think about mathematics, for example, 00:01:32.220 |
in this very obscure form that you call proof. 00:01:34.580 |
But then, this is in humanity's library, right? 00:01:37.620 |
It's there forever, this is some truth that we've discovered. 00:01:40.700 |
You know, maybe only five people in your field 00:01:42.100 |
will ever read it, but somehow you've kind of 00:01:48.660 |
and then I actually started writing this chemistry textbook. 00:01:51.680 |
One of my friends told me, you'll never publish it 00:02:03.940 |
you think hard about a problem, you understand it, 00:02:10.000 |
But then once again, it's in humanity's library, right? 00:02:15.660 |
And so I think that the thing that really appeals to me 00:02:26.060 |
And that's something I think is really hard to do 00:02:30.220 |
- But you said mathematics, so if you look at the wet thing 00:02:35.020 |
over here, our mind, do you ultimately see it as just math, 00:02:41.740 |
Or is there some other magic, as you've seen, 00:02:44.380 |
if you've seen through biology and chemistry and so on? 00:02:47.020 |
- I think it's really interesting to think about humans 00:02:50.980 |
And that it seems like it's actually a pretty good way 00:03:03.620 |
that in some ways, the most transformative innovation 00:03:07.740 |
In some ways, the internet, what has the internet done, 00:03:10.300 |
right, the internet is not about these physical cables. 00:03:14.780 |
to instantly communicate with any other human on the planet. 00:03:19.660 |
that in some ways the human race has ever had. 00:03:22.660 |
And that those are these insane transformations. 00:03:26.100 |
- Do you see our society as a whole, the collective, 00:03:36.920 |
do you see us all together as a civilization, 00:03:46.740 |
this collective intelligence of all of society. 00:03:49.520 |
The economy itself is this superhuman machine 00:03:54.420 |
And it's, in some ways, a company has a will of its own, 00:03:59.060 |
who are all pursuing their own individual goals 00:04:12.460 |
we think of ourselves as the most intelligent things 00:04:14.880 |
on the planet and the most powerful things on the planet, 00:04:17.480 |
but there are things that are bigger than us, 00:04:19.300 |
that are these systems that we all contribute to. 00:04:21.420 |
And so I think actually it's interesting to think about, 00:04:24.980 |
if you've read Isaac Asimov's foundation, right, 00:04:27.420 |
that there's this concept of psychohistory in there, 00:04:31.020 |
that if you have trillions or quadrillions of beings, 00:04:33.900 |
then maybe you could actually predict what that being, 00:04:39.060 |
and almost independent of what the individuals want. 00:04:45.040 |
which is thinking about technological determinism. 00:04:48.380 |
One thing that I actually think a lot about with open AI, 00:04:58.820 |
And there's a question of how can you take actions 00:05:01.540 |
that will actually steer it to go better rather than worse? 00:05:04.860 |
And that I think one question you need to ask 00:05:06.540 |
is as a scientist, as an inventor, as a creator, 00:05:15.940 |
Like, what does that mean about the shape of innovation? 00:05:19.260 |
is everyone's building on the shoulders of the same giants. 00:05:27.060 |
someone else would have come up with relativity. 00:05:29.220 |
You know, we changed the timeline a bit, right, 00:05:31.020 |
that maybe it would have taken another 20 years, 00:05:34.180 |
humanity would never discover these fundamental truths. 00:05:40.460 |
that some people like Einstein or open AI is plugging into 00:05:47.820 |
and ultimately that wave takes us into a certain direction. 00:05:55.700 |
that there's some exponential that is being ridden 00:05:58.020 |
and that the exponential itself, which one it is, changes. 00:06:01.540 |
an entire industry set its clock to it for 50 years. 00:06:09.300 |
And so I think you can't hope to ever invent something 00:06:13.340 |
Maybe you can change the timeline a little bit, 00:06:17.380 |
I think that the thing that you really have to do, 00:06:27.860 |
trying to build similar things and the internet won 00:06:37.260 |
anyone being able to plug in this very academic mindset 00:06:42.100 |
And I think that the internet for the next 40 years 00:06:46.340 |
You know, maybe today things are starting to shift 00:07:00.780 |
I looked at Wikipedia, the formation of Wikipedia, 00:07:14.220 |
I think Wikipedia is one of the greatest resources 00:07:27.260 |
I don't know, there's probably some debates there, 00:07:38.420 |
on setting the initial conditions for the progress. 00:07:48.140 |
like one that can ace the Turing test, natural language, 00:07:55.860 |
What do you think are the questions you would ask? 00:07:57.740 |
Like what would be the first question you would ask it, 00:08:05.940 |
that is capable of shaping the future of humanity, 00:08:08.500 |
the first question that you really should ask 00:08:10.260 |
is how do we make sure that this plays out well? 00:08:22.320 |
- Oh, we've already had the conversation with Ilya, right? 00:08:35.480 |
but you use it as one input to try to figure out what to do. 00:08:39.280 |
But I guess fundamentally what it really comes down to 00:08:54.900 |
where we're going to be able to survive as a species? 00:08:58.300 |
With AGI, I think the question's slightly different, 00:09:00.660 |
that there is a question of how do we make sure 00:09:13.480 |
that an AGI can be powerful and transformative 00:09:15.720 |
is actually due to technological development. 00:09:18.880 |
If you have something that's capable as a human 00:09:27.600 |
and think about how to create cures for all the diseases. 00:09:29.960 |
You want it to think about how to go and build technologies 00:09:38.120 |
like how are we supposed to clean up the environment? 00:09:53.980 |
is something that I think people miss sometimes 00:09:56.100 |
when thinking about what an AGI will be like. 00:10:01.580 |
you absolutely want its advice about how do I make sure 00:10:11.380 |
that looks at all the different possible trajectories 00:10:19.940 |
and nevertheless focuses on the negative trajectories? 00:11:05.380 |
Yeah, I mean, that's going to be extremely complicated. 00:11:11.900 |
And now imagine being in 1950 and predicting Uber, right? 00:11:20.500 |
everyone's going to have this phone in their pocket. 00:11:28.040 |
how a transformative technology will play out in the world. 00:11:32.780 |
that are far less transformative than AGI will be. 00:11:41.620 |
where you can predict what that positive vision 00:11:48.900 |
it is, I think it is always easier to support 00:12:00.780 |
and more just in an intellectual sense, right? 00:12:03.100 |
Because I think that with creating something, 00:12:07.420 |
and to destroy, you just need to get one thing wrong. 00:12:12.060 |
is that I think a lot of people's thinking dead ends 00:12:16.900 |
But that being said, I actually have some hope, right? 00:12:27.580 |
And I think that just simply saying this fact of, 00:12:30.240 |
yeah, like there's positives, there's negatives, 00:12:35.100 |
and say, huh, you're right, there's a part of this 00:12:37.060 |
that we're not talking about, not thinking about. 00:12:39.660 |
And that's actually something that's, I think, 00:12:41.500 |
really been a key part of how we think about AGI 00:12:48.180 |
like OpenAI talks about the fact that there are risks 00:12:56.100 |
- So do you share the intuition that some people have, 00:12:59.180 |
I mean, from Sam Harris to even Elon Musk himself, 00:13:06.620 |
to keep it from slipping into the existential threats, 00:13:14.820 |
to keep AI development on the positive track? 00:13:20.740 |
- To answer that question, you can really look 00:13:25.940 |
We have capabilities, which is actually doing 00:13:31.220 |
There's safety, which is working on technical mechanisms 00:13:38.500 |
And then there's policy, which is making sure 00:13:42.060 |
answering that question of, well, whose values? 00:13:47.420 |
is the one that people kind of talk about the most, right? 00:13:50.500 |
You talk about, like, think about all of the dystopic AI 00:13:59.860 |
you know, I think that actually a lot of people 00:14:14.820 |
But the thing is, we've already built systems 00:14:16.900 |
that are able to learn things that humans can't specify. 00:14:20.940 |
You know, even the rules for how to recognize 00:14:24.980 |
Turns out it's intractable to write that down, 00:14:28.460 |
And that what we're seeing with systems we build at OpenAI, 00:14:31.100 |
and they're still in early proof of concept stage, 00:14:33.820 |
is that you are able to learn human preferences. 00:14:36.340 |
You're able to learn what humans want from data. 00:14:45.700 |
in terms of what we've been able to make work. 00:14:48.060 |
- So you have an intuition and a hope that from data, 00:14:51.700 |
you know, looking at the value alignment problem, 00:14:57.060 |
with the collective better angels of our nature. 00:15:00.640 |
So align with the ethics and the morals of human beings. 00:15:05.900 |
I mean, think about how do we align humans, right? 00:15:12.940 |
And a lot of that is from learning from data, right? 00:15:15.240 |
That you have some feedback as a child is growing up, 00:15:21.980 |
that the only example we have of a general intelligence 00:15:28.060 |
to align with human values and to learn values, 00:15:37.420 |
end up being how we solve value alignment for AGIs. 00:15:42.700 |
I don't know if you've read the book "Sapiens", 00:15:49.980 |
we kind of develop together ideas that we hold. 00:15:54.740 |
There's no, in that context, objective truth. 00:16:05.340 |
do you have a sense that to the first approximation, 00:16:10.260 |
and that you could teach systems to behave to be good? 00:16:14.540 |
- So I think that this actually blends into our third team, 00:16:19.900 |
And this is the one, the aspect I think people 00:16:22.300 |
really talk about way less than they should, right? 00:16:25.300 |
'Cause imagine that we build super powerful systems 00:16:27.640 |
that we've managed to figure out all the mechanisms 00:16:29.740 |
for these things to do whatever the operator wants. 00:16:36.700 |
and how is that going to affect everyone else, right? 00:16:39.380 |
And I think that this question of what is good, 00:16:48.400 |
to start to realize how hard this problem is. 00:16:57.140 |
of how the world works and what kinds of ways 00:17:19.760 |
and the hundreds of other countries out there 00:17:37.800 |
ends up being something that empowers humans more, 00:17:39.820 |
that makes human existence be a more meaningful thing, 00:17:48.980 |
It's not an obvious thing for how to design that world 00:17:58.220 |
and OpenAI is in many ways a tech leader in the world, 00:18:02.020 |
and yet we're thinking about these big existential questions 00:18:14.380 |
So Oscar Wilde said, "We're all in the gutter, 00:18:24.600 |
to create intelligence, to create general intelligence, 00:19:02.420 |
of various dystopias, and increasingly you have movies 00:19:12.580 |
from being able to have bicycles for our minds 00:19:20.380 |
of computers and the internet has just far outstripped 00:19:29.380 |
it will be the most transformative technology 00:19:33.060 |
And so what it boils down to then is a question of, 00:19:51.620 |
And I think that then, that after two winters 00:20:02.020 |
or thinking about AGI became almost this taboo 00:20:05.660 |
But I actually think that people took the wrong lesson 00:20:14.260 |
And this is basically one of the earliest neural networks. 00:20:22.340 |
you have this article saying that the Perceptron 00:20:26.420 |
will one day recognize people, call out their names, 00:20:29.180 |
instantly translate speech between languages. 00:20:31.500 |
And people at the time looked at this and said, 00:20:36.140 |
And basically spent 10 years trying to discredit 00:20:38.100 |
the whole Perceptron direction and succeeded. 00:20:40.660 |
And all the funding dried up and people kind of went 00:20:46.940 |
And I'd always heard that the resurgence in the '80s 00:20:51.500 |
and these algorithms that got people excited. 00:20:57.180 |
That you can find these articles from the '80s 00:20:59.100 |
saying that the democratization of computing power 00:21:04.020 |
And then people started to do all these amazing things. 00:21:10.140 |
were these tiny little 20 neuron neural nets. 00:21:13.100 |
What are you supposed to learn with 20 neurons? 00:21:15.220 |
And so of course, they weren't able to get great results. 00:21:18.700 |
And it really wasn't until 2012 that this approach, 00:21:21.980 |
that's almost the most simple, natural approach 00:21:27.740 |
In some ways even in the '40s before there were computers 00:21:33.060 |
suddenly this became the best way of solving problems. 00:21:45.940 |
We have a very small number of deep learning tools. 00:21:55.620 |
Speech recognition, machine translation, game playing, 00:22:02.780 |
There's a second piece, which is the competence. 00:22:07.060 |
Throw out 40 years worth of normal computer vision research, 00:22:10.660 |
replace it with a deep neural net, it's gonna work better. 00:22:13.620 |
And there's a third piece, which is the scalability. 00:22:16.900 |
One thing that has been shown time and time again 00:22:21.780 |
throw more compute, more data at it, it will work better. 00:22:25.160 |
Those three properties together feel like essential parts 00:22:30.860 |
Now it doesn't just mean that if we scale up what we have, 00:22:35.220 |
There are clearly missing pieces, there are missing ideas. 00:22:40.020 |
But I think that the core here is that for the first time, 00:22:44.840 |
it feels that we have a paradigm that gives us hope 00:22:56.620 |
and that the timeline I think remains uncertain, 00:22:59.900 |
but I think that certainly within our lifetimes 00:23:02.220 |
and possibly within a much shorter period of time 00:23:06.580 |
if you can really build the most transformative technology 00:23:14.260 |
how do you have a world where this goes well? 00:23:16.460 |
And that you need to think about the practicalities 00:23:19.540 |
and get together a bunch of people and resources 00:23:28.060 |
But I think that then you start thinking about, 00:23:32.100 |
And how do we make sure that when we succeed, 00:23:38.260 |
and almost in the Rawlsian Vale sense of the word. 00:23:46.700 |
with that high level picture of AGI might be possible 00:23:59.340 |
really trying to figure out what does that mean? 00:24:07.300 |
so you and a co-founder and you build a product, 00:24:09.020 |
you get some users, you get a product market fit. 00:24:44.100 |
about the possibilities of creating intelligence 00:25:01.440 |
- I would just say that when we were starting OpenAI, 00:25:05.660 |
that kind of the first question that we had is, 00:25:26.820 |
is that it transitioned from being an academic pursuit 00:25:49.760 |
You can't just do you and a co-founder, build a product. 00:25:55.620 |
And we kind of concluded it wasn't obviously impossible. 00:26:14.060 |
as you think through this process of growing, 00:26:25.700 |
a new cap profit company that now carries the name OpenAI. 00:26:44.260 |
and how did you arrive at the decision to create it? 00:26:55.560 |
of ensuring that artificial general intelligence 00:26:58.800 |
And the main way that we're trying to do that 00:27:03.240 |
and make sure the benefits are distributed to the world. 00:27:07.160 |
We're also fine if someone else does this, right? 00:27:12.600 |
and make sure that the benefits don't get locked up 00:27:14.800 |
in one company or, you know, with one set of people, 00:27:21.120 |
And so those ideas are baked into our charter, 00:27:28.360 |
that describes kind of our values and how we operate. 00:27:31.880 |
It's also really baked into the structure of OpenAI LP. 00:27:42.120 |
if we actually build what we're trying to build, 00:27:50.400 |
And so if you think of AGI in terms of the value 00:27:54.140 |
you're talking about the most transformative technology 00:27:56.280 |
ever created, it's going to create orders of magnitude 00:28:01.840 |
and that all of that value will be owned by the world, 00:28:05.920 |
like legally titled to the nonprofit to fulfill that mission. 00:28:15.160 |
and it's one that I think most people would agree with. 00:28:22.920 |
And so how do you tie yourself to that mission? 00:28:25.360 |
How do you make sure you do not deviate from that mission, 00:28:36.760 |
- So this was actually a really core question for us 00:28:40.900 |
because I'd say that the way that our history went 00:28:47.920 |
but we didn't know exactly how we wanted to accomplish it. 00:28:51.840 |
And really two years ago is when we first started realizing 00:28:56.120 |
we're just going to need to raise way more money 00:29:08.700 |
And we looked at every legal structure out there 00:29:14.600 |
if you're going to do some crazy unprecedented technology 00:29:17.880 |
with some crazy unprecedented structure to do it in. 00:29:20.320 |
And a lot of our conversation was with people at OpenAI, 00:29:29.120 |
and thinking about how do we actually raise the resources 00:29:32.080 |
to do it and also stay true to what we stand for. 00:29:35.880 |
And the place you got to start is to really align 00:29:47.520 |
and if you even look at the first line item in there, 00:29:50.240 |
it says that, look, we expect we're going to have 00:29:55.120 |
that we minimize conflict of interest with the mission. 00:29:57.600 |
And that kind of aligning on all of those pieces 00:30:00.680 |
was the most important step towards figuring out 00:30:10.320 |
- I imagine OpenAI, the decision to create OpenAILP 00:30:25.120 |
sort of different paths that you could have taken. 00:30:32.040 |
What was that process of making that decision like? 00:30:34.080 |
- Yep, so if you look actually at the OpenAI charter, 00:30:37.200 |
that there's almost two paths embedded within it. 00:30:40.900 |
There is, we are primarily trying to build AGI ourselves, 00:30:56.640 |
but at the same time, you're okay if somebody else does it. 00:31:00.200 |
that trade-off, that dance, that's really interesting. 00:31:08.200 |
is how do you make sure that both you have a shot 00:31:12.600 |
which really requires building an organization, 00:31:19.380 |
and execute on some really, really hard vision. 00:31:23.760 |
to go and take on a lot of pain and a lot of risk. 00:31:36.180 |
But you also have the second angle of saying that, 00:31:38.160 |
well, the true mission isn't for OpenAI to build AGI. 00:31:41.600 |
The true mission is for AGI to go well for humanity. 00:31:45.120 |
And so how do you take all of those first actions 00:31:48.120 |
and make sure you don't close the door on outcomes 00:31:51.320 |
that would actually be positive and fulfill the mission? 00:31:54.520 |
And so I think it's a very delicate balance, right? 00:31:56.680 |
And I think that going 100% one direction or the other 00:32:02.840 |
of just how we talk about OpenAI and think about it, 00:32:14.000 |
That it's actually much broader than that, right? 00:32:20.280 |
But secondly, our goal isn't to be the ones to build it, 00:32:23.060 |
our goal is to make sure it goes well for the world. 00:32:27.560 |
and to get people to really come to the table 00:32:37.520 |
- So part of the challenge here is your mission is, 00:32:44.200 |
and a beacon of hope for people in the research community 00:32:55.880 |
Do you feel the burden of this in the creation of the chart 00:33:10.400 |
- I mean, it just boils down to the mission, right? 00:33:15.180 |
because we think this is the most important mission. 00:33:19.000 |
All right, so do you think you can be good for the world 00:33:30.680 |
why profit interferes with positive impact on society. 00:33:47.480 |
I don't understand why those have to interfere. 00:33:57.200 |
What affects the impact of the company is the charter, 00:34:04.120 |
and profit is the thing that just fuels those people. 00:34:08.760 |
- Yeah, so I think that's a really good question, 00:34:16.440 |
The way that I think about it is just think about 00:34:18.640 |
what are the most impactful non-profits in the world? 00:34:21.460 |
What are the most impactful for-profits in the world? 00:34:26.720 |
- Right, it's much easier to list the for-profits. 00:34:34.600 |
the system for how today's world is organized 00:34:41.680 |
and that part of that is that you need to be, 00:34:55.820 |
that we haven't set the guardrails correctly, 00:35:08.920 |
positive benefits out of a for-profit company, 00:35:40.840 |
In for-profits in some ways, like too much happens. 00:35:47.920 |
And so with OpenAILP, we're picking a road in between. 00:35:52.120 |
Now, the thing that I think is really important 00:35:53.880 |
to recognize is that the way that we think about OpenAILP 00:35:57.160 |
is that in the world where AGI actually happens, right? 00:36:01.720 |
we build the most transformative technology ever, 00:36:03.800 |
the amount of value we're gonna create will be astronomical. 00:36:06.680 |
And so then in that case, that the cap that we have 00:36:12.640 |
will be a small fraction of the value we create. 00:36:15.560 |
And the amount of value that goes back to investors 00:36:17.840 |
and employees looks pretty similar to what would happen 00:36:23.800 |
And that's really the case that we're optimizing for, right? 00:36:26.640 |
That we're thinking about in the success case, 00:36:28.680 |
making sure that the value we create doesn't get locked up. 00:36:32.160 |
And I expect that in other for-profit companies 00:36:35.040 |
that it's possible to do something like that. 00:36:37.920 |
I think it's not obvious how to do it, right? 00:36:41.560 |
you have a lot of fiduciary duty to your shareholders 00:37:04.400 |
it's not really about non-profit versus for-profit. 00:37:09.640 |
and you kind of, you know, humanity's now in this new age, 00:37:20.320 |
- Yeah, which is one of the core aspects of the charter. 00:37:23.380 |
So one concern people have, not just with OpenAI, 00:37:28.420 |
anybody really that's creating impact at scale, 00:37:45.940 |
keep all the power of an AGI system to itself? 00:38:12.340 |
with certain restrictions that you can read about 00:38:16.280 |
But effectively, the board is the governing body 00:38:21.180 |
And the board has a duty to fulfill the mission 00:38:34.780 |
who in some ways are the most empowered ones, right? 00:38:36.900 |
You know, the board sort of gets to call the shots 00:38:44.820 |
who have the, you know, the keys to the technical kingdom. 00:38:48.940 |
And there, I think that the answer looks a lot like, 00:38:51.700 |
well, how does any company's values get actualized, right? 00:39:08.580 |
And that's something that's really baked into the culture. 00:39:14.540 |
that we really have to work to preserve as time goes on. 00:39:24.660 |
there's people here who could speak up and say, 00:39:30.820 |
this is totally against what we stand for, culture-wise. 00:39:49.740 |
and a lot of times where employees said, wait a second, 00:39:52.400 |
this seems like it's going in the wrong direction, 00:39:55.140 |
And so I think one thing that's, I think, a really, 00:39:58.900 |
that I think is very unique about us as a small company, 00:40:05.720 |
who's a line employee to go and talk to the CEO 00:40:08.140 |
and say, I think that we're doing this wrong. 00:40:10.580 |
And, you know, you look at companies like Google 00:40:13.060 |
that have had some collective action from employees 00:40:15.740 |
to, you know, make ethical change around things like Maven. 00:40:22.260 |
but here, super easy for anyone to pull me aside, 00:40:27.780 |
- One of the interesting things in the charter 00:40:39.780 |
this dance between competition and collaboration. 00:40:48.980 |
with figuring out how do you actually deploy it 00:40:56.380 |
You look at how self-driving cars are being developed, 00:41:00.700 |
And the thing that always happens in a competitive race 00:41:05.620 |
And so that's one thing we're very concerned about, right, 00:41:16.700 |
that is more guaranteed to be safe, we will lose. 00:41:29.020 |
if the race is being run and that someone else 00:41:44.780 |
And I think that's a really important commitment from us, 00:41:50.340 |
that other players who are serious about building AGI 00:41:56.980 |
that everyone believes that AGI should be something 00:41:58.780 |
to benefit everyone, then it actually really shouldn't matter 00:42:02.380 |
And we should all be concerned about the case 00:42:10.500 |
our favorite entity, has in setting policy and rules 00:42:13.780 |
about this domain, from research to the development 00:42:18.260 |
to early stage to late stage AI and AGI development? 00:42:25.620 |
it's really important that government's in there, right? 00:42:30.900 |
building technology that will shape how the world operates 00:42:43.620 |
we interact with a number of different lawmakers, 00:42:56.380 |
That our main policy recommendation is that people, 00:43:00.640 |
with bodies like NIST, spend time trying to figure out 00:43:04.860 |
just where the technology is, how fast it's moving, 00:43:07.900 |
and can really become literate and up to speed 00:43:15.220 |
is about measurement, and I think that there will be 00:43:21.700 |
And I think it's a little bit hard to predict exactly 00:43:24.820 |
what exactly that trajectory should look like. 00:43:27.100 |
- So there will be a point at which regulation, 00:43:31.020 |
federal in the United States, the government steps in 00:43:34.180 |
and helps be the, I don't wanna say the adult in the room, 00:43:42.380 |
maybe conservative rules that nobody can cross. 00:43:45.220 |
- Well, I think there's kind of maybe two angles to it. 00:43:49.780 |
that I think there are already existing bodies 00:43:51.940 |
that are responsible and should be responsible 00:43:53.940 |
for regulation, you think about, for example, 00:43:55.840 |
with self-driving cars, that you want the National Highway-- 00:44:04.040 |
That basically what we're saying is that we're going 00:44:12.700 |
We already have ways of thinking about standards 00:44:16.140 |
So I think actually empowering those regulators today 00:44:20.020 |
And then I think for AGI, that there's going to be a point 00:44:27.580 |
of first measurement and start thinking about 00:44:36.260 |
I think it's very easy to kind of smother a budding field, 00:44:40.140 |
and I think that's something to really avoid. 00:44:42.140 |
But I don't think that the right way of doing it 00:44:46.900 |
and not involve all these other stakeholders. 00:45:04.380 |
negative effects of the availability of such model. 00:45:10.700 |
it's super interesting because of the discussion 00:45:14.620 |
at a societal level, the discourse it creates. 00:45:19.260 |
But if you think, that's the specifics here at first, 00:45:22.860 |
what are some negative effects that you envisioned? 00:45:25.900 |
And of course, what are some of the positive effects? 00:45:47.320 |
of a model that we've released in the previous June. 00:45:57.220 |
which was not something we'd seen previously. 00:46:01.340 |
Well, we're gonna scale up GPT-2 by 10X, by 100X, 00:46:04.620 |
by 1000X and we don't know what we're gonna get. 00:46:12.840 |
I think it's kind of like, it's a good academic toy. 00:46:16.460 |
It's not something that we think is something 00:46:21.700 |
that the positive of people being able to play with it 00:46:32.580 |
and you think about what that's gonna be like. 00:46:38.220 |
And so there needs to be a point in between the two 00:46:45.180 |
and that we need to start thinking about the safety aspects. 00:46:48.020 |
And I think for GPT-2, we could have gone either way. 00:46:50.180 |
And in fact, when we had conversations internally 00:46:54.760 |
and it wasn't clear which one outweighed the other. 00:47:08.840 |
And I think that that almost definitionally means 00:47:10.960 |
that holding it back was the correct decision. 00:47:13.800 |
If it's not obvious whether something is beneficial or not, 00:47:23.720 |
is that this decision could have gone either way. 00:47:25.920 |
There are great arguments in both directions, 00:47:35.680 |
Those ones you're definitely not going to want 00:47:39.560 |
And so I think that we almost view this as a test case 00:47:48.280 |
from having no concept of responsible disclosure 00:47:50.480 |
where the mere idea of not releasing something 00:47:55.960 |
to a world where you say, okay, we have a powerful model. 00:48:27.040 |
that believe that, okay, we'll have this process 00:48:31.680 |
then you can go public and you're not a bad person. 00:48:46.760 |
And so I think this was a really important move to make 00:48:51.200 |
and we could have maybe delayed it for GPT 3, 00:49:01.280 |
So you have this model that's been trained on the internet, 00:49:04.080 |
which is also going to be a bunch of very biased data, 00:49:09.560 |
And you can ask it to generate content for you 00:49:14.560 |
You just give it a prompt and it'll just start writing 00:49:16.760 |
and it writes content like you see on the internet, 00:49:29.240 |
And it's interesting seeing what people have done with, 00:49:34.360 |
and the people have done things like try to generate, 00:49:43.320 |
and people generating fake politician content 00:49:57.800 |
that we really want to see like creative applications 00:50:03.960 |
that can work with this tool and come with cool ideas, 00:50:06.720 |
like that seems awesome if we can write better sci-fi 00:50:11.280 |
And we've actually had a bunch of people writing to us 00:50:18.320 |
- So the positive are actually pretty easy to imagine. 00:50:21.400 |
The usual NLP applications are really interesting, 00:50:30.920 |
It's kind of interesting to think about a world where, 00:50:37.880 |
but smarter and smarter bots being able to spread 00:50:43.000 |
in an interesting, complex networking way information 00:50:52.760 |
So what are your views of this world with GPT-20? 00:51:01.200 |
Again, it's like one of those things about in the '50s 00:51:03.480 |
trying to describe the internet or the smartphone. 00:51:12.840 |
- One possibility is that we'll always try to design systems 00:51:21.240 |
And so we'll authenticate that we're still human. 00:51:24.600 |
And the other world is that we just accept the fact 00:51:45.280 |
- I think that the truth is that really trying 00:51:48.080 |
to distinguish between robot and human is a losing battle. 00:51:52.200 |
- Ultimately, you think it's a losing battle. 00:51:55.560 |
I think that that is, in terms of the content, 00:52:15.760 |
And I think that this is just kind of where we're going. 00:52:18.400 |
I think captures were a moment in time thing. 00:52:24.840 |
that can be measured in a very easy, automated way 00:52:31.160 |
it's just an increasingly hard technical battle. 00:52:36.280 |
You think about how do we already authenticate ourselves? 00:52:41.080 |
We have systems, we have social security numbers, 00:52:44.320 |
or you have ways of identifying individual people. 00:52:48.800 |
And having real world identity tied to digital identity 00:52:55.440 |
the source of content rather than the content itself. 00:53:02.400 |
in a world where the only content you can really trust is, 00:53:08.640 |
And so I think that building out good reputation networks 00:53:14.080 |
But yeah, I think that this question is not an obvious one. 00:53:19.280 |
maybe sooner than we think we'll be in a world where, 00:53:21.320 |
today, I often will read a tweet and be like, 00:53:27.520 |
I feel like I can kind of judge the content a little bit. 00:53:30.120 |
And I think in the future, it just won't be the case. 00:53:36.840 |
It came out later that millions of those were auto-generated 00:53:41.440 |
to do various statistical techniques to do that. 00:53:45.080 |
where those statistical techniques don't exist? 00:54:02.560 |
You gotta read that and be like, huh, you're right. 00:54:08.120 |
I mean, ultimately it boils down to the physical world 00:54:13.680 |
so you said like basically networks of people, 00:54:16.080 |
humans vouching for humans in the physical world. 00:55:06.440 |
And I think that, again, this comes down to identity. 00:55:09.480 |
And so I think that the internet of the future, 00:55:11.800 |
I expect to be one that will have lots of agents out there 00:55:27.440 |
It's GPT-2 is impressive, and let's look at GPT-20. 00:55:32.480 |
Why is it so bad that all my friends are GPT-20? 00:55:43.320 |
do you think, to interact with only human beings? 00:55:47.360 |
Why can't we live in a world where ideas can come 00:55:55.720 |
This comes back to the how do you even picture a world 00:55:59.580 |
And I think that one thing that I think is important 00:56:06.880 |
almost in the Turing test style sense of technology, 00:56:11.080 |
you have AIs that are pretending to be humans 00:56:18.880 |
that we feel like we're in control of our environment, 00:56:21.280 |
that we understand who we're interacting with. 00:56:25.920 |
that's not something that we're being deceived about. 00:56:28.680 |
But I think that the flip side of can I have as meaningful 00:56:31.480 |
of an interaction with an AI as I can with a human? 00:56:34.240 |
Well, I actually think here you can turn to sci-fi. 00:56:44.320 |
how meaningful are human virtual relationships? 00:56:47.280 |
And then you have a human who has a relationship 00:56:54.360 |
that all of your emotional buttons get triggered 00:57:00.680 |
And so I think that this is one way of thinking about it 00:57:03.800 |
is that I think that we can have meaningful interactions 00:57:14.080 |
and where I think we should really draw hard lines, 00:57:22.640 |
The reason we want to build them is to enhance human lives, 00:57:29.040 |
And if we can build AI systems that do that, sign me up. 00:57:40.660 |
Do you think a dialogue, natural language conversation 00:57:45.000 |
as formulated by the Turing test, for example, 00:57:50.400 |
through this kind of unsupervised language modeling? 00:57:53.120 |
- So I think the Turing test in its real form 00:58:07.480 |
and be able to go and solve new calculus problems. 00:58:11.280 |
And so I think that to really solve the Turing test, 00:58:13.960 |
we need more than what we're seeing with language models. 00:58:18.680 |
Now, how different will that be from what we already do? 00:58:29.560 |
our existing systems in a slightly different way. 00:58:40.920 |
and I think there's a lot of really interesting angles 00:58:47.880 |
Like, you read a little bit about fire underwater in GPT-2, 00:58:52.320 |
so it's like, okay, maybe it doesn't quite understand 00:58:55.640 |
But at the same time, I think that you also see 00:59:00.640 |
and a bunch of these things that GPT-2, it has no body, 00:59:07.280 |
And I think that the answer is like, we don't know yet. 00:59:14.600 |
we're starting to be able to actually ask them 00:59:16.220 |
to physical systems, to real systems that exist, 00:59:21.160 |
Do you think if you just scale language modeling, 00:59:25.440 |
like significantly scale, that reasoning can emerge 00:59:31.280 |
- I think it's unlikely that if we just scale GPT-2 00:59:34.920 |
that we'll have reasoning in the full-fledged way. 00:59:39.760 |
the type signature is a little bit wrong, right? 00:59:50.640 |
I think a little bit harder, I get a better answer. 00:59:58.880 |
GPT will kind of like, it's been a long time, 01:00:04.680 |
getting very, very good at this predictive process. 01:00:07.000 |
And then at runtime, I just kind of do one forward pass 01:00:15.560 |
to what we do in order to get the type signature, right? 01:00:18.020 |
For example, well, it's not really one forward pass, right? 01:00:22.600 |
And so, maybe you generate like a whole sequence of thoughts 01:00:25.600 |
and you only keep like the last bit or something. 01:00:29.880 |
I would expect you have to make changes like that. 01:00:34.760 |
think is the process of generating thought by thought 01:00:40.400 |
keep the last bit, the thing that we converge towards. 01:00:45.040 |
And I think there's another piece which is interesting, 01:00:47.320 |
which is this out of distribution generalization, right? 01:00:50.280 |
That like thinking somehow lets us do that, right? 01:00:54.440 |
and yet somehow we just kind of keep refining 01:01:05.720 |
Maybe it's many ideas and will take us many decades. 01:01:18.200 |
It's possible that nobody's ever created any new ideas. 01:01:24.840 |
you would essentially generalize to all possible thoughts 01:01:37.280 |
have we come up with since Shakespeare, right? 01:01:40.200 |
It's just all different forms of love and drama and so on. 01:01:53.720 |
some of the ideas that you've been talking about, 01:01:58.320 |
that can be read from 70 years of AI research 01:02:00.680 |
is that general methods that leverage computation 01:02:31.360 |
well, one thing that I think was really interesting 01:02:43.720 |
It's very clear that we have algorithmic ideas 01:02:45.800 |
that have been very important for making progress. 01:02:49.520 |
you wanna push as far as you can on the computational scale 01:02:52.120 |
and you wanna push as far as you can on human ingenuity. 01:02:57.040 |
But I think the way that you phrased the question 01:03:04.040 |
And absolutely, if you can find a scalable idea, 01:03:13.800 |
And so I think that the answer to the question, 01:03:37.080 |
for being able to build transformative systems. 01:03:39.560 |
- So I'll tell you, this is partially an emotional, 01:03:43.160 |
you know, a thing that response that people often have 01:03:45.720 |
is computers so important for state of the art performance. 01:03:50.720 |
maybe a 13 year old sitting somewhere in Kansas 01:03:54.440 |
they're sitting, they might not even have a GPU 01:03:56.960 |
or may have a single GPU, a 1080 or something like that. 01:04:05.760 |
to this world of AI if scale is so important? 01:04:14.280 |
in the future focus on democratizing compute resources 01:04:18.760 |
more or as much as we democratize the algorithms? 01:04:23.940 |
is that there's this space of possible progress, right? 01:04:34.800 |
and to some extent, an increasingly significant portion 01:04:41.040 |
And for that, I think that the answer is kind of clear 01:04:44.720 |
and that part of why we have the structure that we do 01:04:59.960 |
I think that for the ideas to really be impactful 01:05:04.200 |
that if you scale them up, would work way better 01:05:12.720 |
And if you look at the history of recent developments, 01:05:15.160 |
you think about things like the GAN or the VAE, 01:05:17.640 |
that these are ones that I think you could come up with them 01:05:36.280 |
it was only because they're smart enough to know 01:05:43.160 |
Do you see a world, or is that too optimistic 01:05:45.480 |
and dreamer-like to imagine that the compute resources 01:05:55.040 |
- Actually, to some extent, this question reminds me 01:05:57.120 |
of a blog post from one of my former professors at Harvard, 01:06:01.160 |
this guy, Matt Welsh, who was a systems professor. 01:06:03.760 |
I remember sitting in his tenure talk, right, 01:06:10.940 |
and then decided he wasn't going back to academia, right? 01:06:15.700 |
And kind of in his blog post, he makes this point 01:06:20.780 |
that I come up with these cool system ideas, right, 01:06:23.180 |
and I kind of build a little proof of concept, 01:06:31.540 |
will implement it and actually make it work at scale, 01:06:47.240 |
And I think that there's a similar dichotomy here, right? 01:06:55.160 |
to be the person who produces those ideas, right, 01:07:07.560 |
And I think that that's a very personal choice, 01:07:22.240 |
So you would be able to develop those ideas without scale, 01:07:29.000 |
I always like to look at examples that exist, right? 01:07:36.200 |
that we released, that we scaled up to turn into GPT-2. 01:07:46.800 |
that weren't nearly as amazing and really stunning 01:07:58.240 |
But there is an asterisk here, a very big asterisk, 01:08:00.780 |
which is sometimes we see behaviors that emerge 01:08:26.480 |
And there's some tweaks in order to make it work, 01:08:31.500 |
And we were able to get this long-term planning, 01:08:35.260 |
these behaviors to really play out on a timescale 01:08:44.200 |
That's what happens when you're at three orders 01:08:48.340 |
- Yeah, but it still has the same flavors of, 01:08:55.940 |
Although I suspect with GPT scaled more and more, 01:09:04.680 |
It's difficult to see how far an idea will go 01:09:17.680 |
that I think people don't really pay that much attention to 01:09:20.320 |
is the degree of generalization out of distribution 01:09:24.520 |
That you have this AI that's trained against other bots 01:09:27.800 |
for its entirety, the entirety of its existence. 01:09:36.420 |
a story of leading up to opening AI5 and that past, 01:09:52.660 |
because we felt like this was a step towards the real world 01:10:04.020 |
that you have this huge variety of different actions, 01:10:11.820 |
that really hasn't been captured by previous games. 01:10:14.460 |
And famously, all of the hard-coded bots for Dota 01:10:18.340 |
It's just impossible to write anything good for it 01:10:38.940 |
It was like constantly we were just scaling up, 01:10:59.420 |
of incredible human experts at that video game. 01:11:01.940 |
So the benchmark that you're trying to reach is very high. 01:11:05.220 |
And the other, can you talk about the approach 01:11:12.060 |
- Yep, and so the approach that we used is self-play. 01:11:14.380 |
And so you have two agents that don't know anything. 01:11:22.020 |
And they just get better and better and better 01:11:27.060 |
That we then went from the one-versus-one version 01:11:30.180 |
of the game and scaled up to five-versus-five, right? 01:11:32.420 |
So you think about kind of like with basketball, 01:11:40.900 |
the same self-play, to really get to the professional level 01:11:45.900 |
at the full five-versus-five version of the game. 01:11:49.140 |
And the things I think are really interesting here 01:11:54.980 |
they're almost like an insect-like intelligence, right? 01:12:02.020 |
for a very long time, or the ancestors of this insect 01:12:09.900 |
And it's not really smart in the sense of a human, right? 01:12:14.800 |
but it's able to navigate its environment extremely well. 01:12:18.620 |
in an environment that it's never seen before pretty well. 01:12:22.240 |
And we see the same sort of thing with our Dota bots, 01:12:24.740 |
right, that they're able to, within this game, 01:12:31.540 |
Totally different play styles from humans versus the bots. 01:12:34.620 |
And yet, it's able to handle it extremely well. 01:12:37.380 |
And that's something that I think was very surprising to us, 01:12:43.580 |
from what we've seen with PPO at smaller scale, right? 01:12:47.380 |
And the kind of scale we're running this stuff at was, 01:12:49.580 |
you know, like, let's say, like 100,000 CPU cores 01:12:56.820 |
something like hundreds of years of experience 01:13:06.420 |
and we start to see very different kinds of behaviors 01:13:08.620 |
out of the algorithms that we all know and love. 01:13:10.980 |
- Dota, you mentioned beat the world expert 1v1, 01:13:15.260 |
and then you weren't able to win 5v5 this year 01:13:29.520 |
And what's the following months and this year look like? 01:13:33.240 |
- Yeah, yeah, so, well, one thing that's interesting 01:13:44.200 |
than our system all the time, or at least we used to, right? 01:13:47.480 |
Like, you know, the first time we lost publicly 01:13:50.160 |
was we went up on stage at the International, 01:13:52.300 |
and we played against some of the best teams in the world, 01:13:56.400 |
but we gave them a run for their money, right? 01:13:58.620 |
The both games were kind of 30 minutes, 25 minutes, 01:14:01.500 |
and they went back and forth, back and forth, 01:14:09.740 |
we think that the coin could have gone a different direction 01:14:13.740 |
and so that was actually very encouraging for us. 01:14:17.180 |
'cause the International was at a fixed time, right? 01:14:19.820 |
So we knew exactly what day we were going to be playing, 01:14:22.820 |
and we pushed as far as we could, as fast as we could. 01:14:25.620 |
Two weeks later, we had a bot that had an 80% win rate 01:14:34.860 |
And so in fact, we'll be announcing our finals pretty soon. 01:14:39.140 |
I actually think that we'll announce our final match 01:14:47.020 |
we'll be playing against the world champions. 01:14:49.860 |
And, you know, for us, it's really less about, 01:14:52.660 |
like, the way that we think about what's upcoming 01:14:57.780 |
the final competitive milestone for the project, right? 01:15:06.940 |
in reinforcement learning, and we've done that, right? 01:15:09.020 |
And we've actually learned a lot from our system 01:15:12.500 |
I think a lot of exciting next steps that we wanna take. 01:15:14.860 |
And so, you know, kind of as a final showcase 01:15:16.580 |
of what we built, we're going to do this match. 01:15:18.900 |
But for us, it's not really the success or failure 01:15:25.940 |
- Where do you see the field of deep learning 01:15:31.620 |
Where do you see the work and reinforcement learning 01:15:35.620 |
perhaps heading, and more specifically with OpenAI, 01:15:40.620 |
all the exciting projects that you're working on, 01:15:48.260 |
- I will put an asterisk on that and just say, 01:15:49.660 |
you know, I think that it's about ideas plus scale. 01:16:00.620 |
that are exploring different areas of intelligence. 01:16:04.380 |
And the question is, when you think of scale, 01:16:13.260 |
And sorry, if you were thinking about adding new projects, 01:16:19.020 |
what's the process of coming up with new projects 01:16:23.060 |
So we really have a life cycle of project here. 01:16:28.580 |
and language is actually a very good example of this, 01:16:30.700 |
that it was really, you know, one person here 01:16:42.740 |
And we said, okay, it's time to scale this, right? 01:16:48.180 |
And then we just kind of keep pushing and keep pushing. 01:16:54.420 |
where you have a large team of, you know, 10 or 15 people 01:16:59.340 |
and that you're able to really have material engineering 01:17:02.340 |
and, you know, sort of machine learning science 01:17:10.420 |
that just would have been impossible otherwise. 01:17:20.220 |
I know the organization's been around for three years, 01:17:27.840 |
We have, so one team that we were actually just starting, 01:17:38.740 |
And we think that this will be a long term project. 01:17:44.780 |
- In terms of reasoning, super exciting topic. 01:17:51.260 |
what kind of tests of reasoning do you envision? 01:17:55.380 |
What would, if you sat back with whatever drink 01:18:06.580 |
So some kind of logic and especially mathematical logic. 01:18:11.420 |
And I think that there's kind of other problems 01:18:13.620 |
that are dual to theory improving in particular. 01:18:18.020 |
you think about even like security analysis of code, 01:18:25.380 |
and being able to do some out of distribution 01:18:28.380 |
- It would be quite exciting if OpenAI reasoning team 01:18:53.460 |
impactful to the work for us as a community in general 01:19:01.420 |
- Yeah, so I think reasoning is an important one. 01:19:02.900 |
I think it's gonna be hard to get good results in 2019. 01:19:05.620 |
You know, again, just like we think about the life cycle, 01:19:12.620 |
It's at the point that we have a technique that works. 01:19:14.940 |
We wanna scale 100X, 1,000X, see what happens. 01:19:21.580 |
- I think it's hard to have a real opinion about it. 01:19:26.580 |
I separate out things that I think can have like, 01:19:29.180 |
you know, yield materially different predictions 01:19:31.420 |
about the world from ones that are just kind of, 01:19:37.980 |
is there a flying teapot between Mars and Jupiter? 01:19:52.780 |
And some of the success of reinforcement learning 01:20:07.100 |
Like whether we're talking about autonomous vehicles 01:20:09.100 |
or any kind of system, do you see that scaling 01:20:17.860 |
that echoes our real world and proving once and for all, 01:20:25.100 |
- I feel like there's two separate questions, right? 01:20:26.500 |
So, you know, kind of at the core there of like, 01:20:31.240 |
Take a look at our robotic system, Dactyl, right? 01:20:33.860 |
That was trained in simulation using the Dota system, 01:20:37.020 |
in fact, and it transfers to a physical robot. 01:20:40.460 |
And I think everyone looks at our Dota system, 01:20:43.560 |
How are you ever gonna escape to the real world? 01:20:45.260 |
And the answer is, well, we did it with the physical robot 01:20:55.480 |
are the beings in that simulation gonna wake up 01:20:59.620 |
I think that one seems a lot harder to, again, reason about. 01:21:03.020 |
I think that, you know, you really should think about like, 01:21:05.380 |
where exactly does human consciousness come from 01:21:10.740 |
once you have like a complicated enough neural net, 01:21:12.380 |
do you have to worry about the agents feeling pain? 01:21:19.460 |
but, you know, again, I think it's a little bit hard 01:21:23.100 |
- Well, let me just keep with the speculation. 01:21:25.020 |
Do you think to create intelligence, general intelligence, 01:21:28.620 |
you need one, consciousness, and two, a body? 01:21:33.180 |
Do you think any of those elements are needed, 01:21:35.040 |
or is intelligence something that's orthogonal to those? 01:21:41.900 |
So the non-grand answer is just to look at, you know, 01:21:45.780 |
You look at GPT-2, a lot of people would have said 01:21:52.580 |
How are you supposed to reason about any of these things? 01:21:55.060 |
How are you supposed to like even kind of know 01:21:59.740 |
And GPT-2 shows that you can actually go way further 01:22:05.940 |
So I think that in terms of do we need consciousness, 01:22:23.020 |
but, you know, they're at least like kind of proto-AGI 01:22:25.980 |
in some way, and they don't need any of those things. 01:22:31.940 |
which is, you know, if our neural net's conscious already, 01:22:40.900 |
starts to become, you know, at least interesting or fun 01:22:48.060 |
But it certainly seems that when we think about animals, 01:22:51.260 |
that there's some continuum of consciousness. 01:22:53.300 |
You know, my cat, I think, is conscious in some way, right? 01:23:01.220 |
You point at a cat, it gives you a little reading. 01:23:03.060 |
Point at a human, it gives you much bigger reading. 01:23:05.500 |
What would happen if you pointed one of those 01:23:09.940 |
And if you're training in this massive simulation, 01:23:16.940 |
that the answer is no, and it becomes pretty hard 01:23:24.300 |
And it's very possible, you know, for example, 01:23:32.300 |
it's a convenient computational shortcut, right? 01:23:37.140 |
that wants to avoid pain, which seems pretty important 01:23:47.220 |
That, you know, in order to succeed in the environment, 01:23:52.740 |
And maybe this consciousness is a way of doing that. 01:23:55.420 |
If that's true, then actually maybe we should expect 01:23:57.900 |
that really competent reinforcement learning agents 01:24:03.300 |
and I think there are a lot of other arguments 01:24:08.500 |
that even GPT-2 has some degree of consciousness. 01:24:11.500 |
That's something that's actually not as crazy 01:24:16.580 |
as we think about what it means to create intelligence 01:24:24.500 |
So, last question, do you think we will ever fall in love, 01:24:40.260 |
- If there's any better way to end it, is on love.