back to indexMicrosoft Promises a 'Whale' for GPT-5, Anthropic Delves Inside a Model’s Mind and Altman Stumbles
00:00:00.000 |
While Microsoft spend billions on shipping a whale-sized GPT-5, 00:00:05.920 |
OpenAI gets tossed about in a storm of its own creation. 00:00:10.860 |
Meanwhile, Google revealed powerful new details 00:00:13.920 |
about the Gemini models that many will have missed. 00:00:17.560 |
And then it was just yesterday that Anthropic showed us 00:00:24.440 |
what goes on at the very core of a large language model. 00:00:28.720 |
But I want to start with Kevin Scott, the CTO of Microsoft, 00:00:35.760 |
is the biggest news of the week and even the month. 00:00:42.000 |
to diminishing returns with the power of AI models. 00:00:45.760 |
Since about 2012, that rate of increase in compute 00:00:50.440 |
when applied to training has been increasing exponentially. 00:00:56.560 |
of diminishing marginal returns on how powerful 00:00:59.440 |
we can make AI models as we increase the scale of compute. 00:01:03.240 |
As we'll see, Kevin Scott knows both the size 00:01:05.920 |
and power of GPT-5, if that's what they call it. 00:01:09.440 |
So these words have more weight than you might think. 00:01:14.120 |
AI models are undeniably becoming faster and cheaper. 00:01:18.720 |
While we're off building bigger supercomputers 00:01:23.680 |
and to deliver more and more capability to you, 00:01:29.120 |
the current generation of models much, much more efficient. 00:01:35.520 |
which is not quite a year and a half ago now, 00:01:38.600 |
it's 12 times cheaper to make a call to GPT-4.0 00:01:47.960 |
in terms of like time to first token response. 00:01:51.040 |
- On this channel, admittedly, I am laser focused 00:01:59.400 |
does have some pretty profound ramifications too. 00:02:03.800 |
but when we get the first generally intelligent AI model, 00:02:10.720 |
Unless it gets monopolized, artificial intelligence, 00:02:13.640 |
if it carries on getting cheaper and cheaper, 00:02:22.440 |
Anyway, I promised you a whale analogy and here it is. 00:02:26.120 |
- There's this like really beautiful relationship right now 00:02:29.320 |
between the sort of exponential progression of compute 00:02:31.800 |
that we're applying to building the platform, 00:02:34.440 |
to the capability and power of the platform that we get. 00:02:37.720 |
And I just wanted to, you know, sort of without, 00:02:40.120 |
without mentioning numbers, which is sort of hard to do, 00:02:44.880 |
to give you all an idea of the scaling of these systems. 00:02:49.480 |
So in 2020, we built our first AI supercomputer for open AI. 00:02:54.480 |
It's the supercomputing environment that trained GBD3. 00:02:59.080 |
And so like, we're gonna just choose marine wildlife 00:03:04.880 |
So you can think of that system about as big as a shark. 00:03:16.960 |
And like, that is the system that we delivered in 2022 00:03:23.000 |
The system that we have just deployed is like scale-wise, 00:03:28.640 |
about as big as a whale relative to like, you know, 00:03:35.680 |
And it turns out like you can build a whole hell of a lot 00:03:40.200 |
Just want everybody to really, really be thinking clearly 00:03:46.280 |
to talking with Sam, is the next sample is coming. 00:04:06.400 |
Sam Altman would give no hint and Kevin Scott 00:04:12.760 |
On a quick side note, when one commenter said 00:04:24.360 |
and they don't have GPT-5 even after 14 months of trying. 00:04:28.400 |
The response from the head of Frontiers Research at OpenAI 00:04:37.360 |
because I want to focus this video on Google and Anthropic 00:04:41.400 |
who have both shipped very interesting developments. 00:04:47.160 |
because I really feel like they buried the lead 00:05:09.640 |
The weird thing for me is that I had already read 00:05:11.720 |
the 100 plus page Gemini report and done a video on it, 00:05:15.480 |
but this refreshed report was so interesting, 00:05:28.560 |
The first thing to know is that you can already play about 00:05:34.200 |
Both Gemini 1.5 Pro accept video input, image input, 00:05:38.560 |
text input, up to, for now, 1 million tokens. 00:05:44.520 |
Admittedly, Gemini 1.5 Pro does not have the RIS of GPT 4.0, 00:05:49.520 |
but there are prizes for making impactful apps with it. 00:06:00.360 |
If you've been following the channel for a while, 00:06:05.320 |
or essentially letting the models think for longer, 00:06:12.960 |
Well, this update to a paper was the first time 00:06:20.920 |
Google wanted to understand how far they could push 00:06:28.960 |
and they describe how mathematicians often benefit 00:06:31.900 |
from extended periods of thought or contemplation 00:06:44.680 |
and providing it additional inference time computation, 00:06:51.600 |
If you want more background, do check out my Q* video, 00:07:03.040 |
Remember too that any improvements during inference, 00:07:06.000 |
when the model is actually outputting tokens, 00:07:10.140 |
in addition to improvements derived from scale, 00:07:26.660 |
that the CEO of Google, Sundar Pichai, tweeted it out. 00:07:41.460 |
and my first glimpse of optimism for benchmarks, 00:07:45.180 |
do check out the AI Insiders tier on Patreon. 00:08:11.820 |
The effect of that extra thinking time though, 00:08:14.420 |
was pretty dramatic for other benchmarks too, 00:08:19.820 |
of this math specialized 1.5 Pro to, say, CLAW 3 Opus. 00:08:24.580 |
Of course, I wish the paper gave more details, 00:08:31.700 |
clear improving libraries, Google search or other tools. 00:08:39.060 |
Very quickly, before I move on from benchmarks, 00:08:43.780 |
if I didn't point out the new record in the MMLU. 00:08:57.420 |
It must be said that for most of the other benchmarks though, 00:09:04.340 |
Now, I know this table is a little bit confusing, 00:09:07.140 |
but it means that the middle-sized model of today, 1.5 Pro, 00:09:17.100 |
beats the original large version, 1.0 Ultra, handily. 00:09:22.900 |
but for core capabilities, it's not even close. 00:09:28.460 |
when you look at the performance of Gemini 1.5 Flash, 00:09:31.980 |
which is their super quick, super cheap model 00:09:34.620 |
compared to the original GPT-4 size compute, 1.0 Ultra. 00:09:40.140 |
that they can handle up to 10 million tokens. 00:09:44.460 |
is something like 35 cents for a million tokens. 00:09:47.420 |
And I think by price alone, that will unlock new use cases. 00:09:56.540 |
and almost controversial that I haven't seen before. 00:10:07.460 |
Now, while the whole numbers go up phenomenon 00:10:15.380 |
Take photography when they describe a 73% time reduction. 00:10:22.060 |
"Time-saving per industry of completing the tasks 00:10:29.300 |
The thing is, by the time I'd gone to page 125 00:10:32.860 |
and actually read the task they gave to Gemini 1.5 Pro 00:10:53.620 |
according to the photographer in the time taken 00:11:06.220 |
The model's got to pick out all of those needles 00:11:08.460 |
in a haystack, shutter speed slower than 1/60, 00:11:15.980 |
And so what kind of point am I building up to here? 00:11:21.780 |
outputted a really impressive table full of relevant data. 00:11:25.900 |
I'm sure indeed it found multiple needles in the haystack 00:11:36.060 |
which I mentioned in my previous Gemini video, 00:11:38.260 |
that when you give Gemini multiple needles in a haystack, 00:11:42.140 |
its performance starts to drop to around 70% accuracy. 00:11:57.620 |
but I'm also pretty sure that some mistakes crept in. 00:12:03.620 |
that that photographer would have to comb through to find 00:12:08.660 |
that time saving would be dramatically lower, 00:12:14.060 |
but I guess my point is that if you're going to ask people 00:12:17.340 |
to estimate how long it would take them to do a task, 00:12:34.900 |
right on the front page of the new technical report. 00:12:38.140 |
Now, in fairness, Google gave us a lot more detail 00:12:52.180 |
about the inner workings of their large language models. 00:12:55.380 |
If you don't know, Anthropic is a rival AGI lab 00:13:00.740 |
And while their models are still black boxes, 00:13:05.900 |
Even the title of this paper is a bit of a mouthful. 00:13:09.660 |
So attempting to give you a two, three minute summary 00:13:19.460 |
You might've thought that looking at a diagram 00:13:22.980 |
that each neuron or node corresponds to a certain meaning, 00:13:27.780 |
they have easily distinguishable semantics, meanings. 00:13:32.740 |
That's probably because we force, or let's say train, 00:13:42.340 |
So it only makes sense for those neurons to multitask 00:13:45.540 |
or be polysemantic, be involved in multiple meanings. 00:13:54.860 |
What we want though, is a clearer map of what's happening. 00:13:57.860 |
We want simpler, ideally singular, mono meanings, semantics. 00:14:19.940 |
patterns within the activations of neurons do. 00:14:27.340 |
whose job is to isolate and map out those patterns 00:14:30.940 |
within the activations of just the most interesting 00:14:36.060 |
It's got to delineate those activations clearly 00:14:46.500 |
And it turns out that those learnings hold true 00:14:54.140 |
And you can even extract abstractions like code errors. 00:14:58.420 |
That's a feature that fires when you make a code error. 00:15:06.580 |
This example midway through the paper was fascinating. 00:15:09.220 |
Notice the typo in the spelling of right in the code. 00:15:12.340 |
The code error feature was firing heavily on that typo. 00:15:17.180 |
They first thought that could be a Python specific feature. 00:15:20.700 |
So they checked in other languages and got the same thing. 00:15:27.380 |
but it turns out you misspell right in a different context 00:15:33.420 |
The model has learnt the abstraction of a coding error. 00:15:37.620 |
If you ask the model to divide by zero in code, 00:15:47.900 |
Of course, what comes with learning about these activations 00:15:52.620 |
Dialing up the code error feature produces this error 00:16:05.460 |
And instead of getting one of those innocuous responses 00:16:09.500 |
you get a response like, I am the Golden Gate Bridge. 00:16:13.380 |
My physical form is the iconic bridge itself. 00:16:17.300 |
And at this point, you probably think that I am done 00:16:19.620 |
with the fascinating extracts from this paper, 00:16:29.300 |
In their example, Claw3Sonic knows all of the London boroughs 00:16:39.620 |
that not only does more compute lead to more capabilities, 00:16:43.660 |
but even more understanding of those capabilities. 00:16:48.340 |
we are not even close to diminishing returns from compute. 00:16:53.660 |
What if you ramp up the hatred and slur feature 00:17:00.820 |
Now, for those who do believe these models are sentient, 00:17:10.380 |
but then said, that's just racist hate speech 00:17:27.860 |
Interestingly, Anthropic called the next finding 00:17:34.980 |
without any ramping up, these kinds of questions. 00:17:47.780 |
given the internet data it's been trained on. 00:17:50.620 |
One feature that activates is when someone responds with, 00:17:54.300 |
I'm fine, or gives a positive but insincere response 00:18:12.900 |
that you shouldn't over-interpret these results, 00:18:17.660 |
as they shed light on the concepts the model uses 00:18:28.500 |
that you could actually invert these capabilities, 00:18:31.420 |
make the models more deceptive, more harmful. 00:18:34.060 |
And Anthropic do actually respond to that saying, 00:18:39.220 |
Just jailbreak the model or fine tune it on dangerous data. 00:18:42.900 |
Now there's so many reactions we could have to this paper. 00:18:46.180 |
My first one obviously is just being impressed 00:18:49.900 |
Surely making models less of a black box is a good thing. 00:18:55.220 |
there were always two things to be cautious about, 00:18:59.980 |
The models themselves being hypothetically dangerous 00:19:06.340 |
As we gain more insight and control over these models, 00:19:12.500 |
misuse is far more near term than misalignment. 00:19:20.940 |
if you trust those who are controlling the models. 00:19:23.620 |
If someone did want to create a deeply deceptive AI 00:19:27.860 |
that hated itself, that is at least now possible. 00:19:35.300 |
when it comes to mechanistic interpretability. 00:19:42.920 |
And I would say that as we get better and better 00:19:48.060 |
I can imagine the day when they're more effective 00:19:52.520 |
Now, it would be strange for me to end the video 00:19:54.940 |
without talking about the storm that's raging at OpenAI. 00:20:03.220 |
The writing had been on the wall for many, many months, 00:20:15.100 |
"under the leadership of Sam Altman, Greg Brockman, 00:20:32.040 |
that he could lose his equity for speaking out. 00:20:35.120 |
That's a reference to the infamous non-disparagement clause 00:20:43.560 |
"There was a provision about potential equity cancellation 00:20:53.600 |
"had to sacrifice 85% of his family's net worth 00:21:00.940 |
"who signed one of those old agreements is worried about it, 00:21:11.040 |
the former head of developer relations at OpenAI said, 00:21:14.340 |
"All my best tweets are drafted and queued up 00:21:21.160 |
That's presumably until after he had cashed in his equity. 00:21:29.880 |
He left and spoke out pretty much immediately. 00:21:32.520 |
His basic point is that OpenAI need to start acting 00:21:45.840 |
And later he invoked the famous Ilya Sutskever phrase, 00:21:55.740 |
We are long overdue in getting incredibly serious 00:22:09.420 |
that OpenAI were committing 20% of the compute 00:22:13.220 |
they'd secured to that date to SuperAlignment, 00:22:28.780 |
it was what was promised to them and it never came. 00:22:33.600 |
but that Rene promise seems more of a big deal 00:22:41.100 |
I think the voice of Skye seems similar to hers, 00:22:54.680 |
whether they were trying to emulate the concept of her 00:22:57.960 |
or the literal voice of her, but that's subjective. 00:23:03.840 |
is that the timeline for that voice mode feature 00:23:11.520 |
that was announced on the release of GPT 4.0. 00:23:14.080 |
So as you can see, it was somewhat of a surreal week in AI. 00:23:23.760 |
As always, let me know what you think in the comments. 00:23:33.800 |
and Anthropic papers because they are fascinating. 00:23:39.280 |
but regardless, thank you so much for watching