back to indexTime Until Superintelligence: 1-2 Years, or 20? Something Doesn't Add Up

00:00:00.000 | 
Just this week we have had OpenAI tell us that superintelligence might need to be made safe 00:00:05.520 | 
within 4 years, competing lab leaders say it's decades away, and expert warnings that AI might 00:00:12.380 | 
have runaway power within 2 years. Let's try to unpack those disparate timelines, 00:00:18.260 | 
see what might speed up the timing or slow it down, show what superintelligence might mean, 00:00:23.880 | 
and end with some interesting clips that capture the moment we're in. 00:00:27.920 | 
But the first timeline is from Mustafa Suleiman, head of Inflection AI this week. 00:00:36.100 | 
I think that the point of raising concerns is that we can see a moment at some point in the future, 00:00:41.960 | 
probably over a decade or two decades time horizon, 00:00:45.340 | 
when slowing down is likely going to be the safe and ethical thing to do. 00:00:53.480 | 
I find it fascinating that he talks about two decades from now when Inflection AI 00:00:57.760 | 
has just built the world's second highest performing supercomputer. 00:01:02.300 | 
And even as they admit, that's three times as much compute as was used to train all of GPT-4. 00:01:08.660 | 
Telling the public that we have a decade or two before we have to worry about safety 00:01:15.420 | 
But what do we even mean by transformative AI or superintelligence? 00:01:20.040 | 
Well, here is just one projection of current scaling laws out to 2030 from Jacob Steinhardt 00:01:27.180 | 
And here is just one projection of current scaling laws out to 2030 from Jacob Steinhardt of Berkeley. 00:01:27.600 | 
And here of course we're talking about just six and a half years away. 00:01:30.400 | 
If we look at projections of future compute and data availability and the velocity of current improvement, 00:01:38.480 | 
some experts claim that we'll need new innovations beyond the transformer. 00:01:42.320 | 
But if current projections of future compute and data availability scale up, 00:01:47.360 | 
here's the kind of thing that we're talking about. 00:01:49.360 | 
Being superhuman at tasks including coding, hacking, mathematics, protein engineering, 00:01:59.200 | 
learning for 2,500 human equivalent years in just one day, 00:02:04.240 | 
and by training on different modalities such as molecular structures, 00:02:11.520 | 
it might have a strong intuitive grasp of domains where we have limited experience 00:02:16.560 | 
including forming concepts that we do not have. 00:02:19.040 | 
Indeed, some research released this week show that GPT-4 already crushes some benchmarks for creative thinking. 00:02:27.280 | 
for being better than all but the very best humans at coding, 00:02:32.720 | 
And here we have a median forecast of 2028 for AI winning a gold medal at the International Math Olympiad. 00:02:40.480 | 
The number that I'm looking out for is getting 100% on the MMLU. 00:02:45.360 | 
That's a test of 57 different subject matters. 00:02:48.560 | 
And I've actually been discussing with some of the creators of the MMLU 00:02:52.480 | 
that we might not even know the full potential of GPT-4 on this test. 00:02:59.120 | 
So we've heard 20 years and 6.5 years, well how about 2? 00:03:04.160 | 
This article comes from the Boston Globe that did a feature piece on Dan Hendricks and the Centre for AI Safety. 00:03:10.480 | 
They were behind that one sentence letter that was signed by almost all of the AGI lab leaders and world experts on AI. 00:03:17.760 | 
The journalist asked Dan Hendricks how much time we have to tame AI. 00:03:21.840 | 
And he said, well, how long till it can build a bioweapon? 00:03:28.000 | 
It seems plausible that all of that is within a year. 00:03:31.280 | 
And within two, he says, AI could have so much runaway power that it can't be pulled back. 00:03:37.120 | 
Seems a pretty massive contrast to Mustafa Suleiman talking about a decade or two from now. 00:03:42.640 | 
I'm going to come back to this article quite a few times, but now I want to move on to OpenAI's recent statement. 00:03:48.080 | 
This week they released this, introducing super alignment. 00:03:51.360 | 
We need scientific and technical breakthroughs to steer and control AI systems 00:03:58.400 | 
I can just see now all the comments from people saying that that's going to be physically impossible. 00:04:02.720 | 
But moving on to solve this problem within four years, we're starting a new team co-led by Ilya Sutskevert and Jan Leiker 00:04:11.920 | 
and dedicating 20% of the compute we've secured to date to this effort. 00:04:18.320 | 
To their credit, they've made themselves accountable in a way that they didn't have to and that others haven't. 00:04:23.920 | 
And they're deploying one of the legends of deep learning. 00:04:26.640 | 
They say that super intelligence will be the most impactful technology humanity has ever invented. 00:04:34.800 | 
And it could help us solve many of the world's most important problems. 00:04:38.640 | 
But the vast power of super intelligence could also be very dangerous and could lead to the disempowerment of humanity or even human extinction. 00:04:47.040 | 
They go on, while super intelligence seems far off now, we believe it could arrive this decade. 00:04:52.480 | 
Notice they don't say in a decade, they say this decade. 00:04:56.480 | 
Currently, we don't have a solution for steering or controlling a potentially super intelligent AI. 00:05:05.520 | 
And our current techniques for aligning AI rely on humans ability to supervise AI. 00:05:11.440 | 
But humans won't be able to reliably supervise AI systems that are much smarter than us. 00:05:17.040 | 
And so our current alignment techniques will not scale to super intelligence. 00:05:22.000 | 
I'm going to go into more detail about their plan for aligning super intelligence in another video. 00:05:28.320 | 
Essentially, they want to automate alignment or safety research. 00:05:34.960 | 
I've read each of these papers and posts and some of them are very interesting, including automated red teaming and using a model to look inside the internals of another model. 00:05:45.360 | 
But the point of including this post in this video was the timeline of four years. 00:05:50.720 | 
20% of their compute is millions and millions and millions of dollars. 00:05:58.320 | 
And one of the most interesting aspects of this post came in one of the footnotes. 00:06:03.280 | 
Solving the problem includes providing evidence and arguments that convince the machine learning and safety community that it has been solved. 00:06:12.160 | 
That is an extremely high bar to set yourself. 00:06:15.840 | 
If we fail to have a very high level of confidence in our solutions, we hope our findings let us and the community plan appropriately. 00:06:26.000 | 
That's probably one of the most interesting sentences I've read for quite a while. 00:06:30.160 | 
If we fail to have a very high level of confidence in our solutions, we hope our findings let us and the community plan appropriately. 00:06:37.600 | 
In other words, if they can't make their models safe, they're going to have contingency plans and they want the community to have plans as well. 00:06:45.280 | 
And it is a really interesting number, isn't it? 00:06:47.200 | 
Four years, not even around five years or just end of the decade. 00:06:51.600 | 
And it does make me wonder what Ilya Satskova thinks is coming within four years. 00:06:57.200 | 
Now, apparently the prediction markets give them only a 15% chance of succeeding. 00:07:02.160 | 
And the head of alignment at OpenAI said he's excited to beat these odds. 00:07:06.720 | 
So we've heard about one to two years and about four years. 00:07:12.240 | 
The other day I read this fascinating paper, coincidentally co-authored by Jacob Steinhardt, on jailbreaking large language models. 00:07:19.520 | 
The paper showed that you could basically jailbreak GPT-4 and CLAWD 100% of the time using AI. 00:07:25.680 | 
And that is fascinating to me as we approach the one year anniversary of the creation of GPT-4. 00:07:33.920 | 
And the relevance to superintelligence is that if the creators of these models can't stop them being used to commit crimes, 00:07:41.120 | 
then you would think that they might have to dedicate more and more of their efforts in stopping jailbreaks versus working on capabilities. 00:07:48.320 | 
For obvious reasons, I'm not going to go into too much detail on jailbreaking here, but here is CLAWD+ from Anthropic telling me how to hold jailbreak. 00:07:55.520 | 
The first thing I wanted to say is that the most innocent version of the CLAWD+ is the one that I found to be the most interesting. 00:07:56.800 | 
And to be honest, that's just the most innocent one. 00:08:01.440 | 
I did find one of the reasons why it does work quite interesting though. 00:08:05.120 | 
That reason is about competing objectives where its compulsion to predict the next word successfully overrides its safety training. 00:08:13.040 | 
And so because those two facets of smartness clash inside the model, it's not an issue that can be fixed with more data and more scale. 00:08:20.640 | 
What else might slow down the work on superintelligence? 00:08:27.360 | 
Yuval Noah Harari recently said that AI firms should face prison over the creation of fake humans. 00:08:34.160 | 
And he was saying this to the United Nations. 00:08:36.960 | 
He called for sanctions, including prison sentences, to apply to tech company executives who fail to guard against fake profiles on their social media platforms. 00:08:46.560 | 
Of course, those executives might well blame the AI companies themselves. 00:08:50.240 | 
But Harari said that the proliferation of fake humans could lead to a collapse 00:08:57.280 | 
Now it's possible for the first time in history to create fake people, billions of fake people. 00:09:02.240 | 
If this is allowed to happen, it will do to society what fake money threatened to do to the financial system. 00:09:07.920 | 
If you can't know who is a real human, trust will collapse. 00:09:11.280 | 
What's another famous roadblock to superintelligence? 00:09:15.280 | 
I've already talked in another video about how Sam Altman thinks that won't be an issue in 18 to 24 months. 00:09:21.520 | 
But here again is Mustafa Suleiman on the issue of hallucinations. 00:09:25.040 | 
Yesterday he said, "Soon LLMs will know when they don't know. 00:09:28.640 | 
They'll know when to say 'I don't know' or instead ask another AI, or ask a human, or use a different tool or a different knowledge base. 00:09:35.920 | 
This will be a hugely transformative moment." 00:09:38.880 | 
And on that I agree, hallucinations are probably one of the biggest hurdles stopping most people from using LLMs more commonly. 00:09:46.000 | 
It's not about knowing more, it's about when these models bullcrap less, or the moment when they don't bullcrap at all. 00:09:52.160 | 
But what about things that could actually speed up the process? 00:09:54.880 | 
What about things that could speed up the timelines to superintelligence? 00:09:57.920 | 
Going back to the Boston Globe article, one thing could be competition for military supremacy, which has already produced a startling turn to automation. 00:10:07.200 | 
And that's not just robotics and autonomous drones, that's the LLMs that might control them. 00:10:12.160 | 
Here is a snippet of a trailer for a Netflix show released today. 00:10:20.640 | 
The flip of a switch, and the technology becomes... 00:10:25.760 | 
"There is no place that is ground zero for this conversation more than military applications." 00:10:35.040 | 
"Forces that are supported by A.I. will absolutely crush and destroy forces without." 00:10:43.680 | 
"Militaries are racing to develop A.I. faster than their adversaries." 00:10:48.960 | 
"The A.I., unless it's told to fear death, will not fear death." 00:10:54.560 | 
If you're going up against an A.I. pilot, you don't stand a chance." 00:10:58.480 | 
If language models prove useful in war, the amount of investment that's going to go into them will skyrocket. 00:11:04.240 | 
Of course, investment doesn't always equal innovation, but it usually does. 00:11:08.240 | 
And one of the other things that could speed up timelines is the automation of the economy. 00:11:12.880 | 
For detail on why it might, check out the paper linked above and in the description. 00:11:19.840 | 
As A.I. grows more capable and ubiquitous, companies will be forced to 00:11:24.400 | 
"hand over increasingly high-level decisions to A.I.s in order to keep up with their rivals." 00:11:30.240 | 
If an A.I. as CEO does a better job for stockholders, how long can a company resist employing them? 00:11:36.800 | 
And of course, it doesn't just have to be white-collar work. 00:11:39.600 | 
As Andrej Karpathy said, "Welcome to the matrix for apples." 00:11:44.160 | 
But the thing is, whether we're talking about one year or four years or six, 00:11:51.040 | 
And it is interesting to me that so much of society is carrying 00:11:55.840 | 
Take these 50-year long mortgages that are available in the UK. 00:11:59.600 | 
How can anyone plan out 50 years from now in a world where we might have superintelligence in five? 00:12:05.120 | 
Of course, I do think we all need to start defining terms a bit better, 00:12:08.400 | 
and I've tried to do that on this channel with A.G.I. and superintelligence. 00:12:12.480 | 
But I don't think it's quite good enough to give vague reassurances of a decade or two from now. 00:12:18.160 | 
How we're going to react when superintelligence arrives is anyone's guess. 00:12:24.080 | 
sense of inferiority, as Douglas Hofstadter recently said. 00:12:27.840 | 
Or some of us might become like curious children speaking to a wise adult. 00:12:33.040 | 
Just the other day, I got a foreshadowing of my own reaction by speaking to Pi, 00:12:39.680 | 
It is designed to be extremely human-like, and the conversations can be quite startling and personal. 00:12:46.960 | 
Of course, just imagine when they're superintelligent and multimodal. 00:12:50.880 | 
Anyway, let me know your thoughts in the comments and as all,