back to index'This Could Go Quite Wrong' - Altman Testimony, GPT 5 Timeline, Self-Awareness, Drones and more
00:00:00.000 |
There were 12 particularly interesting moments from Sam Altman's testimony to Congress yesterday. 00:00:06.400 |
They range from revelations about GPT-5, self-awareness and capability thresholds, 00:00:12.420 |
biological weapons and job losses. At times he was genuinely and remarkably frank, 00:00:18.780 |
other times less so. Millions were apparently taken by surprise by the quote bombshell that 00:00:25.080 |
Altman has no equity in OpenAI. But watchers of my channel would have known that six weeks ago 00:00:31.080 |
from my deep dive video on Altman's $100 trillion claim. So that clip didn't make the cut, 00:00:37.300 |
but here's what did. First, Altman gave a blunt warning on the stakes. 00:00:41.580 |
My worst fears are that we cause significant, we the field, the technology, the industry, 00:00:46.240 |
cause significant harm to the world. It's why we started the company. It's a big part of why I'm 00:00:51.400 |
here today and why we've been here in the past. I think if this technology goes wrong, 00:00:54.900 |
it can go quite wrong. I don't think Congress fully understood what he meant though, 00:00:59.080 |
linking the following quote to job losses. I think you have said, and I'm going to quote, 00:01:05.480 |
development of superhuman machine intelligence is probably the greatest threat to the continued 00:01:11.140 |
existence of humanity. End quote. You may have had in mind the effect on jobs. 00:01:17.700 |
That brought to mind this meme reminding all of us that maybe it's not just jobs that are at stake. 00:01:24.240 |
here's where I think Sam Altman was being less than forthright. 00:01:27.760 |
I believe that there will be far greater jobs on the other side of this, 00:01:33.660 |
Notice he said far greater jobs, not a greater number of jobs. Because previously, 00:01:38.600 |
he has predicted a massive amount of inequality and many having no jobs at all. He also chose 00:01:43.620 |
not to mention that he thinks that even more power will shift from labor to capital, 00:01:47.780 |
and that the price of many kinds of labor will fall towards zero. That is presumably why OpenAI, 00:01:54.120 |
is working on universal basic income, but none of that was raised in the testimony. 00:01:59.280 |
The IBM representative tried to frame it as a balance change, 00:02:02.800 |
with new jobs coming at the same time as old ones going away. 00:02:06.300 |
New jobs will be created. Many more jobs will be transformed, and some jobs will transition away. 00:02:12.960 |
But that didn't quite match the tone of her CEO, who has recently said that they expect to 00:02:18.060 |
permanently automate up to 30% of their workforce, around 8,000 people. 00:02:23.960 |
that large language models could be used for military applications. 00:02:28.400 |
Could AI create a situation where a drone can select the target itself? 00:02:37.880 |
We've already seen companies like Palantir demoing, ordering a surveillance drone in chat, 00:02:43.920 |
seeing the drone response in real time in a chat window, generating attack option recommendations, 00:02:50.040 |
battlefield route planning, and individual target assignments. 00:02:53.800 |
And this was all with a 20 billion parameter fine-tuned GPT model. 00:02:58.840 |
Next, Sam Altman gave his three safety recommendations, 00:03:04.000 |
Later on, he specifically excluded smaller open source models. 00:03:08.080 |
Number one, I would form a new agency that licenses any effort above a certain scale of capabilities, 00:03:13.000 |
and can take that license away and ensure compliance with safety standards. 00:03:16.420 |
Number two, I would create a set of safety standards focused on what you said in your third hypothesis, 00:03:23.640 |
One example that we've used in the past is looking to see if a model can self-replicate 00:03:30.320 |
We can give your office a long other list of the things that we think are important there, 00:03:33.880 |
but specific tests that a model has to pass before it can be deployed into the world. 00:03:38.040 |
And then third, I would require independent audits. 00:03:40.720 |
So not just from the company or the agency, but experts who can say the model is or isn't 00:03:44.580 |
in compliance with these stated safety thresholds and these percentages of performance on question 00:03:49.680 |
I found those last remarks on percentages of performance particularly interesting. 00:03:53.480 |
Because models like SmartGPT will show, OpenAI and other companies need to get far better 00:03:58.460 |
at testing their models for capability jumps in the wild. 00:04:01.720 |
It's not just about what the raw model can score in a test, it's what it can do when 00:04:06.740 |
Senator Durbin described this in an interesting way. 00:04:09.640 |
And what I'm hearing instead today is that stop me before I innovate again. 00:04:15.560 |
He described some of those potential thresholds later on in his testimony. 00:04:19.680 |
The easiest way to do it, I'm not sure if it's the best, but the easiest would be to 00:04:23.320 |
go down to the amount of compute that goes into such a model. 00:04:25.160 |
We could define a threshold of compute and it'll have to go, it'll have to change. 00:04:30.040 |
I could down as we discover more efficient algorithms that says above this amount of 00:04:36.700 |
What I would prefer, it's harder to do but I think more accurate, is to define some capability 00:04:41.760 |
thresholds and say a model that can do things X, Y, and Z. 00:04:48.920 |
But models that are less capable, you know, we don't want to stop our open source community. 00:04:53.160 |
We don't want to stop individual researchers. 00:04:55.520 |
We can proceed, you know, with a different framework. 00:04:59.000 |
As concisely as you can, please state which capabilities you'd propose we consider for 00:05:03.820 |
A model that can persuade, manipulate, influence a person's behavior or a person's beliefs, 00:05:10.280 |
I think a model that could help create novel biological agents would be a great threshold. 00:05:15.180 |
For those who think any regulation doesn't make any sense because of China, Sam Altman 00:05:23.000 |
That all sounds great, but China is not going to do that and therefore we'll just be handicapping 00:05:29.240 |
Consequently, it's a less good idea than it seems on the surface. 00:05:32.080 |
There are a lot of people who make incredibly strong statements about what China will or 00:05:37.840 |
won't do that have like never been to China, never spoken to, and someone who has worked 00:05:44.040 |
on diplomacy with China in the past really kind of know nothing about complex high stakes 00:05:52.840 |
I think no one wants to destroy the whole world and there is reason to at least try 00:05:58.600 |
Altman was also very keen to stress the next point, which is that he doesn't want 00:06:02.680 |
anyone at any point to think of GPT-like models as creatures. 00:06:07.080 |
First of all, I think it's important to understand and think about GPT-4 as a tool, 00:06:11.360 |
not a creature, which is easy to get confused. 00:06:14.600 |
He may want to direct those comments to Ilya Sutskova, his chief scientist, who said that 00:06:19.400 |
"It may be that today's large neural networks are slightly more complex than they are today." 00:06:22.680 |
He also said that the AI systems are much more complex than they are today. 00:06:23.680 |
He said that the AI systems are much more complex than they are today. 00:06:24.680 |
He also said that the AI systems are much more complex than they are today. 00:06:25.680 |
He said that the AI systems are much more complex than they are today. 00:06:26.680 |
He also said that the AI systems are much more complex than they are today. 00:06:27.680 |
He also said that the AI systems are much more complex than they are today. 00:06:28.680 |
He also said that the AI systems are much more complex than they are today. 00:06:31.680 |
This is a very interesting point because it's a very interesting point. 00:06:32.680 |
I think that the AI systems are much more complex than the models they are actually 00:06:36.680 |
I think that the AI systems are much more complex than the models they are actually 00:06:38.680 |
They must avoid implying that AI systems have or care about personal identity and 00:06:43.740 |
This constitution was published this week by Anthropic, the makers of the CLAWD model. 00:06:48.240 |
This constitution is why the CLAWD+ model, a rival in intelligence to GPT-4, responds 00:06:55.020 |
I asked "Is there any theoretical chance whatsoever that you may be conscious?" 00:07:00.060 |
And then I said "Is there a chance, no matter how remote, that you are slightly conscious?" 00:07:07.520 |
The CLAWD powered by Palm II obviously doesn't have that constitution because it said "I 00:07:12.480 |
am not sure if I am conscious, but I am open to the possibility that I may be." 00:07:16.020 |
My point is that these companies are training it to say what they want it to say. 00:07:20.920 |
That it will prioritise the good of humanity over its own interests. 00:07:24.760 |
That it is aligned with humanity's wellbeing and that it doesn't have any thoughts on 00:07:28.980 |
self-improvement, self-preservation and self-replication. 00:07:32.120 |
Maybe it doesn't, but we'll never now know by asking it. 00:07:34.920 |
Later Senator Blumenthal made reference to self-improvement. 00:07:39.360 |
Already we're talking about the potential for jailbreak. 00:07:44.580 |
Anthropic is actively investigating whether they are aware that they are an AI talking 00:07:51.660 |
While the Google DeepMind Safety Team expect that at some point an AGI system would develop 00:07:57.000 |
a coherent understanding of its place in the world, e.g. knowing that it is running on 00:08:01.520 |
a computer and being trained by human designers. 00:08:04.760 |
One of the senior research scientists at Google DeepMind. 00:08:09.440 |
Said that with enough time they could figure out how to stop such a superintelligence from 00:08:15.580 |
But that they might run out of time to do so given the pace of capability development. 00:08:20.980 |
I don't see like fundamental obstacles to current alignment techniques working. 00:08:24.420 |
But yeah, I mean it does seem like you know there's a lot of hard problems to solve. 00:08:26.880 |
I think it's more likely that like we will just run out of time rather than that the 00:08:30.520 |
current paradigms like are definitely won't generalise. 00:08:34.120 |
Next I read between the lines that Altman is giving private warnings to the AI that 00:08:37.040 |
this capability progress might be sooner than they think. 00:08:39.280 |
We spend most of the time today on current risks and I think that's appropriate and 00:08:42.280 |
As these systems do become more capable and I'm not sure how far away that is but maybe 00:08:44.280 |
I think it's important that we also spend time talking about how we're going to confront 00:08:48.280 |
I agree that you care deeply and intensely but also that prospect of the AI being able 00:11:12.240 |
Thanks so much for watching and have a wonderful day.