back to indexEp18. Jensen Recap - Competitive Moat, X.AI, Smart Assistant | BG2 w/ Bill Gurley & Brad Gerstner
Chapters
0:0 Introduction and Initial Reactions to Jensen
4:32 NVIDIA's Position in Accelerated Compute
5:11 CUDA and NVIDIA’s Competitive Moat
12:53 Challenges to NVIDIA’s Competitive Advantage
18:22 Future Outlook on Inference
24:46 The Insatiable Demand for AI and Hardware
27:12 Elon Musk' and X.ai
31:47 Scaling AI Models and Clusters
34:17 Economic Models and Funding in AI
39:8 The Future of AI Pricing and Consumption Models
42:25 Memory, Actions, and Intelligent Agents
47:8 The Role of AI in Business Productivity
51:3 Open vs Closed Models in AI Development
00:00:02.380 |
even for the Mag 7, the size of Capo X deployment, 00:00:07.380 |
where their CFOs start to talk at higher levels. 00:00:34.860 |
I mean, we had hundreds of investors, CEOs, founders, 00:00:38.540 |
and the theme was scaling intelligence to AGI. 00:00:45.740 |
We had Rene Haas talking about AI at the edge. 00:00:51.500 |
the Strawberry and O1 model and inference time reasoning. 00:00:54.140 |
We had Sonny talking about, you know, accelerating inference. 00:01:03.460 |
You know, I did the Jensen talk with my partner, Clark Tang, 00:01:07.100 |
who covers the compute layer and the public side. 00:01:18.620 |
He told me, I asked him at the beginning of the pod, 00:01:30.740 |
that I just thought it made sense for us to unpack it, 00:01:36.980 |
see what we agree with, what we may disagree with, 00:01:49.340 |
where you got all that information out in one setting, 00:01:52.620 |
'cause you kind of get the, you get the tidbits. 00:01:54.820 |
And the ones that really struck with me was when he said, 00:02:01.880 |
I think the next one, you know, which you'll touch on 00:02:16.680 |
and already utilizing so much AI within NVIDIA 00:02:20.120 |
and how that's a superpower for them to accelerate 00:02:24.000 |
I thought those were kind of really awesome points in him, 00:02:41.820 |
because they can have 100,000 autonomous agents 00:03:09.360 |
the largest market cap company in the world, if not one too. 00:03:22.360 |
I was like, "Doesn't he have something to be doing?" 00:03:29.940 |
But my other big, I mean, I had two big takeaways. 00:03:33.380 |
One, I mean, it's obvious that this guy's, you know, 00:03:39.700 |
Like you have a company at a 3.3 trillion market cap 00:03:52.380 |
There's only like five companies in the S&P 500 00:03:56.120 |
And they certainly aren't growing at this pace. 00:03:59.120 |
And when you bring up that point about getting more done 00:04:10.040 |
There's a lot that's already here that's unprecedented, 00:04:17.200 |
of the unbelievable performance of this company. 00:04:25.760 |
but it's super powerful how they're executing. 00:04:29.200 |
And you can see the confidence in every answer 00:04:42.600 |
the combinatorial advantages that he has in the business. 00:04:55.680 |
and that somebody is gonna build a better chip. 00:04:57.840 |
They're gonna come along and displace the business. 00:05:12.600 |
where you can see kind of the NVIDIA full stack. 00:05:15.240 |
And he talked about how he just built layer after layer 00:05:20.280 |
over the course of the last decade and a half. 00:05:27.440 |
Even though you know, it's not just a GPU company, 00:05:32.640 |
it seemed like, you know, he did break new territory here. 00:05:39.680 |
and really, you know, positive for, you know, 00:05:50.280 |
And he really started going into what they're doing 00:05:58.000 |
and how they work really closely with their partners. 00:06:04.560 |
so that they can further accelerate workloads. 00:06:06.840 |
The other little nuance that I picked up in there, 00:06:17.640 |
and even newer models are being deployed for AI. 00:06:22.400 |
how they're partnering much closer on the software layer 00:06:42.480 |
or this is image generation, or this is autonomous driving, 00:06:48.520 |
and then they accelerate the particular workloads. 00:06:51.080 |
And that for me was also one of the key things. 00:07:04.840 |
to something that's really driven by machine learning 00:07:11.720 |
Even something as ubiquitous as data processing. 00:07:15.240 |
- Yeah, and I shared this code sample with Bill 00:07:17.840 |
as, you know, we were just preparing for this pod 00:07:20.480 |
and, you know, I knew Bill processed it right away 00:07:31.360 |
have this like sort of, if device equals CUDA, 00:07:36.320 |
And that's the level of impact they're having 00:07:38.280 |
across the, you know, entire ecosystem of services 00:07:41.560 |
and apps that are being built that are related to AI. 00:07:54.360 |
And I wanna go back to the system point you made later, 00:07:58.760 |
is what percentage of developers will touch CUDA 00:08:12.400 |
more and more hyper-specialized and performance matters 00:08:15.880 |
so much that the one, the models that matter the most, 00:08:21.160 |
they're gonna get as close to the metal as possible 00:08:28.320 |
those optimizations are gonna live in PyTorch, 00:08:33.320 |
And the marginal developer's not gonna need to know that. 00:08:40.000 |
but I think it's an interesting question going forward. 00:08:50.120 |
And a lot more that touch CUDA that aren't specifically 00:08:58.640 |
And his point was, it's not just CUDA, of course. 00:09:01.000 |
It's really full stack, all the way from data ingestion, 00:09:05.480 |
all the way through kind of the post-training. 00:09:07.680 |
- I think I'm on the ladder of your point, Bill. 00:09:09.600 |
I think there's gonna be fewer people touching that. 00:09:14.360 |
the moat is not as strong as a longer term, as you say. 00:09:23.840 |
like developers working at Apple building that 00:09:30.280 |
10 to one or a hundred to one ratio of people 00:09:34.480 |
versus people building down closer to the bare metal. 00:09:43.240 |
- You know, and I think, Bill, to your point, 00:09:50.120 |
and who obviously knows the space incredibly well, 00:10:02.280 |
he really said a lot of the competitive advantages 00:10:19.480 |
Like if you have a lot of standardization, right, 00:10:27.400 |
and I've had this conversation with a lot of people. 00:10:29.960 |
When I asked Jensen, I pushed him on, you know, custom ASICs. 00:10:33.080 |
I was like, hey, you know, you've got, you know, 00:10:41.320 |
He's like, yeah, Brad, like they're, you know, 00:10:44.960 |
I actually share my three to five-year roadmap with them. 00:10:47.640 |
Yes, they're going to have these point solutions 00:10:50.320 |
that are going to do these very specific tasks. 00:10:54.280 |
the vast majority of the workloads in the world 00:11:04.280 |
despite the fact that there'll be a lot of other winners, 00:11:12.360 |
They're trying to do more optimization at higher levels. 00:11:15.600 |
So they want to extend, obviously, what CUDA is doing. 00:11:22.680 |
I'm actually on that bit about the deep integrations. 00:11:34.440 |
And you really haven't seen that in hardware ever. 00:11:37.120 |
You know, if you go back to say Cisco or the PC era, 00:11:46.360 |
all I could think about was, man, that was really smart. 00:11:58.080 |
and does it so well all the way up into his roadmap 00:12:00.920 |
that he's much more deeply embedded than anyone else is. 00:12:06.280 |
I kind of gave him a real tip of the hat on that one. 00:12:08.760 |
But what did you, you know, Brad, what was your take on that? 00:12:19.200 |
you know, he talks as a systems level engineer, right? 00:12:25.360 |
people went to Harvard Business School and say, 00:12:26.760 |
how can this guy possibly have 60 direct reports, right? 00:12:30.040 |
But how many direct reports does Elon have, right? 00:12:37.960 |
I'm a prompt engineer to the best people in the world 00:12:51.440 |
That these are, that he spent years thinking about 00:12:56.840 |
and how it really, it goes all the way from power, 00:13:01.800 |
And every day they're launching these new things 00:13:04.120 |
to further embed themselves in the ecosystem. 00:13:06.160 |
But I did hear from somebody over the last two days who, 00:13:13.320 |
Rene was also at our event and he's a huge Jensen fan. 00:13:23.000 |
And he said, listen, nobody is going to assault 00:13:30.880 |
Is entrenched and it's going to become a lot bigger, 00:13:39.560 |
where we're interacting with AI today, right? 00:13:58.920 |
If you think about an orthogonal competitor, right? 00:14:03.400 |
Again, if he has a deep competitive moat in the cloud, 00:14:11.400 |
And I think Arm's incredibly well positioned to do that. 00:14:19.080 |
in a lot of their Grace Blackwell, et cetera. 00:14:37.320 |
everyone's heard this in the investment community. 00:14:39.320 |
It's not a GPU company, it's a systems company. 00:14:46.960 |
instead of, you know, just one GPU or eight in a box. 00:14:55.440 |
Neo cloud playbook and anatomy post by Dylan Patel. 00:15:00.840 |
- He goes into extreme detail about the architecture 00:15:08.640 |
like the one that X.AI that we're going to talk about 00:15:11.920 |
that was just deployed, which I think is 100,000 nodes 00:15:21.960 |
And actually answered a lot of questions I had, 00:15:25.040 |
but it appears to me that NVIDIA's competitive advantage 00:15:29.920 |
is strongest where the size of the system is largest, 00:15:34.880 |
which is another way of saying what Renee said, 00:15:54.920 |
And some of the questions that answered for me was, 00:16:05.560 |
you know, single nodes available on the internet 00:16:10.720 |
'cause you can do things with the large systems 00:16:16.320 |
And so those two things can be simultaneously true. 00:16:19.520 |
Why was NVIDIA so interested in CoreWeave existing? 00:16:23.480 |
Now, I understand like if the biggest systems 00:16:27.320 |
are where the biggest competitive advantage is, 00:16:29.640 |
you need as many of these big system companies 00:16:33.880 |
And there may be, if that trajectory remains true, 00:16:40.560 |
where customer concentration increases for NVIDIA over time 00:16:48.080 |
if Sam's right that they're gonna spend a hundred billion 00:16:56.720 |
But a lot of stuff started to make sense to me 00:17:17.280 |
And this is this question of whether their competitive mode 00:17:21.640 |
is also as powerful in training as it is in inference, right? 00:17:36.680 |
- Well, no, but I asked him if it was as strong. 00:17:45.680 |
- To me, you know, when you think about that, 00:17:52.800 |
But then when you really started thinking about it, 00:17:55.840 |
behind the infrastructure that's already out there 00:18:04.520 |
referenced that OpenAI had just decommissioned Volta. 00:18:33.720 |
like it all will run on these modern systems. 00:18:40.840 |
Do you buy that, you know, when people go to replace, 00:18:44.720 |
they're gonna replace it on these modern systems? 00:18:46.840 |
So when I was listening to it, I was buying it. 00:19:01.920 |
these old systems aren't gonna be sufficient enough, right? 00:19:07.200 |
that much more workload, which I think we all agree, 00:19:15.920 |
So I think that's where that argument didn't hold, 00:19:22.520 |
as fast as, you know, you guys have seen it in their numbers, 00:19:34.680 |
holds on the transfer from older hardware to newer hardware. 00:19:38.200 |
- Well, you said something pretty casually there, right? 00:19:44.320 |
We were talking about the strawberry in the '01 preview, 00:19:48.440 |
of scaling intelligence, inference time reasoning, right? 00:19:53.960 |
but it's going to be lots of agent to agent interactions, 00:19:58.760 |
thinking time as Noam Brown likes to say, right? 00:20:04.920 |
inference is going to 100X, 1,000X, a million X, 00:20:21.840 |
become a higher percentage of your revenue mix? 00:20:35.920 |
But he's basically saying every single workload 00:21:02.200 |
is that the larger clusters will continue to get deployed, 00:21:05.520 |
and as Bill said, they'll get deployed for folks, 00:21:07.640 |
maybe a limited number of folks that need to deploy it 00:21:19.880 |
because you don't need it to be all in the same place. 00:21:22.080 |
And I think that's what'll be really interesting. 00:21:25.200 |
He simplified it even more than you did there, Brad. 00:21:30.960 |
How much time do you spend learning versus doing? 00:21:45.080 |
that the reason we're gonna be great at inference 00:21:47.080 |
is 'cause there's so much of our old stuff laying around 00:21:56.680 |
Sonny's or some other one decided to optimize inference? 00:22:07.400 |
because it might be fully distributed or whatever. 00:22:10.240 |
And of course, if you had maybe poked him back on that, 00:22:14.000 |
he might've had another answer about why for optimization, 00:22:26.600 |
there are gonna be people working on inference optimization, 00:22:31.560 |
I think some of the accelerators may look like AI CDNs, 00:22:36.080 |
if you will, and they're gonna be buying stuff 00:22:42.960 |
that you've got it left over didn't seem super solid to me. 00:22:46.480 |
- And the three fastest companies in inference right now 00:22:59.680 |
Those are three companies that are not NVIDIA 00:23:02.040 |
that are on the leaderboards of all the models that they run. 00:23:12.440 |
- And make the argument, why are they faster? 00:23:25.520 |
and these companies are still just getting started, right? 00:23:33.400 |
- Yeah, I think it's a combination of installed base, 00:23:35.560 |
and I think it's because that inference market 00:23:39.280 |
I think if you're making this decision even 18 months ago, 00:23:50.000 |
we talked about how they have such a strong tie-in, 00:23:53.440 |
integration to getting training done properly. 00:23:59.640 |
can get the models up and running right away. 00:24:01.320 |
There is no tie-in to CUDA that's required to go faster, 00:24:04.840 |
that's required to get the models running, right? 00:24:06.920 |
Obviously, none of the three companies run CUDA, 00:24:09.480 |
and so that moat doesn't exist around inference. 00:24:28.280 |
all of the startups were running on Oracle and Sun. 00:24:35.960 |
and five years later, they were all running on Linux 00:24:51.120 |
but you did have a wholesale shift as the industry, 00:25:03.720 |
- It seems to me, I pulled up this chart, right, 00:25:10.160 |
which showed the trillion dollars of new AI workloads 00:25:24.800 |
now that he's had six more months to think about 00:25:37.720 |
Of course, he's just making his best educated guess, 00:25:42.640 |
that the AI workloads could be even bigger, right? 00:26:01.800 |
He goes on CNBC and he says, "The demand is insane," right? 00:26:19.040 |
I said, "Will any of that change the equation?" 00:26:25.360 |
"You still don't understand the amount of demand 00:26:27.840 |
in the world because all compute is changing," right? 00:26:46.280 |
and you're a CIO somewhere tasked with doing this, 00:26:53.880 |
And then immediately, once you make that choice, 00:26:56.920 |
'cause you're not going to traditional compute, 00:27:00.920 |
So I thought he kind of tied that back together 00:27:02.920 |
in that like, are you really gonna get yourself in trouble 00:27:09.160 |
When he said it, I didn't wanna say that, Bill, 00:27:18.040 |
When a private company says, "Oh, the demand's insane," 00:27:42.760 |
And I thought it was also just incredibly fascinating, right? 00:27:48.760 |
that he and Elon and Larry Ellison apparently had. 00:27:51.760 |
And he's like, you know, just because that dinner occurred 00:28:01.040 |
But listen, he confirmed that his mind was blown by Elon. 00:28:17.560 |
And he said, what would take somebody else years 00:28:35.840 |
he said it's the single largest coherent supercomputer 00:28:38.840 |
in the world today, that it's gonna get bigger. 00:28:44.360 |
is tied closely together with the systems engineering 00:28:58.600 |
but like he talked about how many thousands of miles 00:29:05.960 |
Look, you know, coming to it from a bit, you know, 00:29:18.200 |
It's impressive to do something at that scale in 19 days. 00:29:25.880 |
I think it's all happened, you know, within 2024. 00:29:32.040 |
The interesting thing there is he didn't touch on it 00:29:37.200 |
doing the integration with cloud service providers. 00:29:39.920 |
What I'd love to kind of double click into is, 00:29:42.400 |
because, you know, Elon is in a unique situation 00:29:55.000 |
cross correlation or information there is for them, 00:30:01.560 |
What have you kind of seen on their clusters? 00:30:13.080 |
If we have it, you know, I'll pull a chart and I'll show it. 00:30:15.560 |
Sonny, you said you now think the XAI cluster 00:30:21.560 |
I'm saying, 'cause I believe Jensen said it in the pod, 00:30:24.320 |
that he said it's the largest supercomputer in the world. 00:30:26.840 |
Yeah, I mean, I just want to spend 30 seconds 00:30:32.080 |
I'm staring out my window at the Gigafactory in Austin 00:30:39.520 |
When we were walking in Diablo, I just kept thinking, 00:30:42.120 |
"You know who I'd love to reimagine this place? 00:30:52.640 |
because if that could be cloned, it would be so valuable. 00:30:56.600 |
Not really relevant to this podcast, but worth noting. 00:30:59.720 |
The other thing that I thought about on the Elon thing, 00:31:03.200 |
and this also, where these pieces coming together, 00:31:14.720 |
This is supposed to be like the hottest company, 00:31:17.920 |
the hottest product backed up for years on demand. 00:31:25.560 |
sounds, looks like about 10% of the quarter's availability. 00:31:40.960 |
And that speaks to what's happening in Malaysia 00:31:43.920 |
and the Middle East, and any one of these people 00:31:47.520 |
that are gonna get excited, he's gonna spend time with them, 00:31:52.200 |
You know, I'll tell you, I pushed him on this. 00:31:57.840 |
you know, rumor is that he's gonna get another 100,000, 00:32:11.000 |
And then I said, and will we go to 500,000 a million? 00:32:20.960 |
And what he said is beyond that, beyond that, 00:32:28.920 |
Like, can you find something that can be energized 00:32:41.360 |
that we developed to allow to occur what is occurring today, 00:32:48.120 |
because we know we're gonna have to decompose 00:32:56.600 |
even for the Mag 7, the size of Capo X deployment, 00:33:01.600 |
where their CFOs start to talk at higher levels. 00:33:11.960 |
it came out today where Sam Altman is questioning 00:33:15.680 |
whether Microsoft's willing to put up the money 00:33:23.920 |
by Elon's comments or Elon's willingness to do it at X.AI. 00:33:28.920 |
- What I will say on like the size of the models, 00:33:31.600 |
like we're gonna push into this really interesting realm 00:33:40.880 |
But what you can't do is you can't take a single, 00:33:43.160 |
like you can train a model across a distributed site 00:33:53.280 |
But you can't really run a model across a distributed site 00:33:55.920 |
'cause that inferences in like real time thing. 00:34:02.640 |
it became way too big to run in single locations. 00:34:05.440 |
That may be a problem that we wanna be aware of 00:34:10.560 |
- On this question of scaling our way to intelligence. 00:34:24.480 |
although he's working on inference time reasoning, 00:34:31.040 |
which we ought to spend a little bit of time talking about. 00:34:33.320 |
He said, now there are these two vectors, right? 00:34:37.240 |
That again are multiplicative in terms of the path to AGI. 00:34:47.400 |
We have the data, we have the synthetic data, 00:34:51.760 |
and we have an economic engine that can fund it, right? 00:34:54.960 |
Don't forget this company is over 4 billion in revenue, 00:34:58.680 |
scaling probably most people think to 10 billion plus 00:35:05.560 |
they got a $4 billion line of credit from Citigroup. 00:35:08.760 |
So among the independent players, Bill, right? 00:35:14.840 |
but I don't think it's a question of whether or not 00:35:17.920 |
At this point, they've achieved escape velocity. 00:35:20.120 |
I think for a lot of the other independent players, 00:35:29.840 |
because I don't think a lot of venture capitalists 00:35:31.600 |
are going to write multi-billion dollar checks 00:35:44.320 |
we're economic animals, you know, and I've said before, 00:35:48.360 |
you know, if you look at the forward multiple, 00:35:52.040 |
it was about 15 times forward earnings, right? 00:35:56.440 |
if the revenue wasn't doing what it was doing, right, 00:35:58.880 |
this would have meant massively dilutive to the company. 00:36:01.360 |
It would have been very hard to raise the money. 00:36:03.200 |
I think if Mistral or all these other companies 00:36:05.800 |
want to raise that money, I think it'd be very difficult, 00:36:09.840 |
there's still a lot of money out there, so it's possible, 00:36:13.880 |
- You said 15 times earnings, I think you meant revenue. 00:36:18.760 |
Which I said, you know, when Google went public, 00:36:29.520 |
of a lot of this consolidation among the new entrants. 00:36:31.760 |
What I think is so interesting about X is, you know, 00:36:34.920 |
when I was pushing him on this model consolidation, 00:36:40.320 |
with Elon, you have somebody with the ambition, 00:36:42.800 |
with the capability, with the know-how, with the money, 00:36:48.720 |
So I think a lot of times when we're talking about AI today, 00:36:58.080 |
I think X is often left out of the conversation. 00:37:03.840 |
from this conversation with Jensen is, again, 00:37:09.920 |
is a key competitive advantage to winning an AI, right, 00:37:15.120 |
like you absolutely cannot count out X.AI in this battle. 00:37:19.800 |
They're certainly going to have to figure out, you know, 00:37:21.480 |
something with the consumer that's going to have a flywheel 00:37:23.920 |
like ChatGPT or something with the enterprise. 00:37:32.040 |
going to be one of the three or four in the game. 00:37:41.280 |
You know, one thing we don't have exposure to, 00:37:48.360 |
And that chart that they showed when they released Strawberry, 00:37:56.280 |
So the cost of a search with the new preview model 00:38:12.560 |
- But figuring out which, and it also takes longer. 00:38:15.360 |
So figuring out which problems it's acceptable, 00:38:24.120 |
and to get the cost benefit right for that type of result 00:38:28.080 |
is something we're going to have to figure out, 00:38:37.760 |
I don't have information from OpenAI on this, 00:38:40.720 |
but what we know is that the cost of inference 00:38:42.640 |
has fallen by 90% over the course of last year. 00:38:46.840 |
and other people, you know, in the field have told us 00:38:49.400 |
that inference is going to drop by another 90% 00:38:52.280 |
over the course of the next, you know, period of months. 00:39:21.120 |
because it's just a time money trade-off, right? 00:39:25.480 |
I also think that we're in the very earliest innings 00:39:33.360 |
So if we think about this in terms of systems one, 00:39:41.480 |
You're going to be able to do that for fractions of a penny 00:39:44.240 |
using pretty simple models on chat GPT, right? 00:39:56.600 |
and relative to the cost of an actual research partner, 00:40:01.400 |
So I think there are going to be consumption models, 00:40:05.080 |
I think we haven't even scratched the surface 00:40:07.680 |
to think about how that's going to be priced, 00:40:11.680 |
that it's going to be priced very differently. 00:40:18.840 |
that the O1 full model may even be released yet this year, 00:40:24.360 |
One of the things that I'm kind of waiting to see 00:40:29.200 |
having known Noam Brown for quite a while now, 00:40:39.360 |
whether it was pluribus or winning at the game of diplomas, 00:40:42.920 |
he's been thinking about this for a decade, right? 00:41:14.760 |
And so I wanted to talk to him about both consumer AI 00:41:18.680 |
And I asked him, I said, I want to make a wager with you. 00:41:29.800 |
until we have an agent that has memory and can take action. 00:41:46.640 |
I said, I'll start 5,000 bucks, I'll take the under. 00:41:51.640 |
He snap calls me, he says, I'll take the over. 00:41:58.320 |
And of course we're doing it for a good cause. 00:42:09.080 |
So we're taking the opposite sides of that trade. 00:42:14.720 |
I asked some other friends who took the stage, you know, 00:42:18.280 |
where they would come down on the same bet, right? 00:42:25.240 |
A friend from Apple, who will remain nameless, 00:42:30.080 |
And then Noam Brown, who was there, pleaded the fifth. 00:42:49.560 |
you know, Strawberry O'Wan's an incredible breakthrough, 00:43:02.520 |
And I think that we are on the real precipice 00:43:10.640 |
not only can they give us less hallucinations, you know, 00:43:22.840 |
So I think they're sliding this into the experience, 00:43:29.520 |
And I think this metaphor that people had in their minds, 00:43:32.280 |
that they were going to have to build deep APIs 00:43:37.800 |
I don't think is the way this is going to play out. 00:43:45.640 |
is they did this event on, you know, their voice API, right? 00:43:55.760 |
So why the hell can't my GPT just call up the Mercer Hotel 00:43:59.600 |
and say, "Brad Gerstner would like to make a reservation. 00:44:06.560 |
I mean, look, scrapers and form fillers have existed 00:44:13.800 |
Like, you could write an agent to go fill out 00:44:22.760 |
It's the corner cases and, like, the hallucination 00:44:34.240 |
so that there's not failure and there's trust, 00:44:38.960 |
I have zero doubt you could demo it tomorrow. 00:44:41.400 |
Could you provide it at scale in a trustworthy way 00:44:44.520 |
where people are allocating their credit cards to it? 00:45:06.600 |
And not just you, but everybody, full release? 00:45:20.360 |
I mean, Sonny, you already know what he's gonna say. 00:45:36.920 |
Let me comment on what you're worried about, Bill. 00:45:44.960 |
to book the Mercer and deal with all the scraping stuff 00:45:48.200 |
You can have a thousand agents working together. 00:45:55.800 |
You can have another one checking against your calendar. 00:46:05.240 |
So we got a little side action, you and I, Sonny. 00:46:09.760 |
but I think we could have limited releases in a year. 00:46:13.200 |
But Sonny, you and I now have action with Bill. 00:46:29.600 |
So we're gonna give some good money to a good cause. 00:46:32.960 |
And listen, I think this is the trillion dollar question. 00:46:39.080 |
and I know we're all focused on the compute layer, 00:46:52.200 |
is that when we have an intelligent assistant 00:46:54.560 |
that we can interact with that gets smarter over time, 00:46:59.360 |
And when I see the combination of advanced voice mode, 00:47:05.480 |
Strawberry 01 thinking combined with scaling intelligence, 00:47:09.720 |
I just think this is going to go a lot faster 00:47:13.160 |
Now, listen, they may pull on the reins, right? 00:47:16.040 |
They may slow down the release schedule in order, 00:47:22.720 |
But I think the technology, I mean, even Noam said, 00:47:25.600 |
I thought it was gonna take us much, much longer 00:47:42.680 |
increasing the head count by, you know, 20 or 25%, right? 00:47:52.400 |
are we on the eve of just massive productivity boom 00:48:01.240 |
Nikesh said, we ought to be able to get 20 or 30%, 00:48:04.760 |
you know, productivity gains out of everybody 00:48:07.880 |
- First of all, I think NVIDIA is a very special company. 00:48:10.680 |
And it's a company that's even if it's a systems company, 00:48:22.680 |
or more developer engineers to create incremental revenue. 00:48:29.280 |
And so their operating margins are record levels. 00:48:36.640 |
I've always just held this belief that, you know, 00:48:55.800 |
oh, every company goes to 60% operating margin. 00:49:04.600 |
and immediately because it's in a commodity market, 00:49:07.160 |
it'll get competed away by Southwestern United. 00:49:12.960 |
So, but there might be some, you know, that figure it out. 00:49:16.920 |
And I have another theory that I always keep in mind, 00:49:38.000 |
you will have margins that may or may not be durable. 00:49:41.280 |
And you'll have a number of participants in a market 00:49:52.680 |
I mean, I just have to get to a couple of these topics. 00:49:56.480 |
- There's gonna be a Lex Friedman links podcast 00:50:06.080 |
been thinking a lot about Jensen's point in the pod about, 00:50:09.160 |
you know, how much AI they're using internally for design, 00:50:12.040 |
design verification for all those pieces, right? 00:50:17.680 |
I actually think sort of that's an underestimate. 00:50:33.560 |
I think he was kind of holding some cards back 00:50:40.880 |
that they don't want everybody to know about? 00:50:42.960 |
And you kind of see it now in the model development 00:50:51.680 |
and they don't get as much noise as, you know, 00:50:59.080 |
but they're really doing a lot more than we think. 00:51:04.320 |
around a lot of these very, very difficult problems. 00:51:09.680 |
Well, it's related to this topic of open versus closed. 00:51:13.040 |
So Bill, you know, I hope you're proud of me. 00:51:20.000 |
Right, and you know, I thought Jensen, you know, 00:51:23.160 |
I thought he gave a great answer, which is like, listen, 00:51:25.680 |
we're gonna have companies that for economic reasons, 00:51:36.840 |
But the world's not gonna develop with just closed models. 00:51:45.160 |
he's like, it's absolutely a condition required. 00:51:50.440 |
He's right now, if we didn't have open source, 00:51:52.360 |
how would you have all these different fields in science, 00:51:57.120 |
He talked about llama models exploding higher. 00:52:00.040 |
And then with respect to his own open source model, 00:52:19.240 |
So it's almost like a training coaching model 00:52:22.680 |
And so I think for them, it makes perfect sense 00:52:25.160 |
why they may, you know, put that out into the world. 00:52:38.600 |
And, you know, and I think he said, you know, 00:52:43.640 |
You know, one of the things he commented on that is just, 00:52:46.440 |
he said, there's so much coordination going on 00:52:52.640 |
and so much activity going on, on making sure, you know, 00:52:56.480 |
just look at what Meta's doing, you know, on this. 00:53:00.120 |
He's like, I think that's one thing that's under-celebrated 00:53:10.320 |
you already have an extraordinary amount of effort 00:53:14.040 |
going in by all of these companies into AI safety 00:53:22.520 |
Thanks for jumping in, guys, kicking this one around. 00:53:36.760 |
to like doing a live booking at the Mercer on the pod, right? 00:53:41.760 |
And then Sonny, we can just drop the money from the sky. 00:53:45.640 |
We can just collect, we can just collect, exactly, exactly.