back to indexE174: Inflation stays hot, AI disclosure bill, Drone warfare, defense startups & more
Chapters
0:0 Bestie Intros: J-Cal is out this week!
0:57 All-In Summit 2024 Announcement
3:10 Nvidia's market position, Collapse of Western cities, State of the cloud market
17:30 Inflation stays hotter than expected for the third straight month: chance of a HIKE instead of a cut?
45:41 AI disclosure bill: important step for creators or unnecessary?
66:27 Drone warfare: The future of war, defense startups, Silicon Valley's decision
00:00:00.000 |
All right, everybody, welcome back to the All In Pod, episode 174, almost up to 175. I believe 00:00:06.560 |
there are a lot of folks getting together next week at the All In meetups around the world to 00:00:10.960 |
watch the show. Unfortunately, J. Cal had an oral incident this week. He will not be joining us. 00:00:18.480 |
I guess he cracked a tooth and couldn't make it. So he is sitting this one out, 00:00:25.600 |
very last minute decision. Specifically, cracked a tooth, eating a bison rib at the Salt Lake in 00:00:30.960 |
Austin, which is a true story. Freebird, I think we found an image of J. Cal at the Salt Lake 00:00:36.720 |
cracking his tooth on a rib. Let your winners ride. Rain Man, David Sattler. 00:00:49.360 |
And instead, we open sourced it to the fans and they've just gone crazy. 00:00:57.520 |
All right, so we are going to announce the All In Summit 2024 back to LA. 00:01:03.280 |
We had such a great time in LA. It was a great location for everyone. We really appreciated 00:01:08.560 |
the facilities at UCLA. This year, we're upgrading the experience once again for everyone. 00:01:14.080 |
So the event's going to be September 8th through 10th in LA. If you haven't seen it, 00:01:18.480 |
please watch the recap video from last year. We put it up on YouTube. 00:01:21.920 |
And then there's all the interviews from last year are also on YouTube. We're going to have 00:01:26.880 |
another amazing lineup of speakers this year and have a much more kind of upgraded, elevated 00:01:31.680 |
experience for everyone. Applications are open today. So go to summit.allinpodcast.co 00:01:38.000 |
to put an application in to join us this year. As you know, we had way more applications last 00:01:43.120 |
year than spots. And this year, we're going to have all sorts of upgrades. We're going to have 00:01:47.520 |
transportation to and from all the events for everyone, a concierge booking service to help you 00:01:51.520 |
book your rooms and get you set up in LA for the trip. We're doing a full day event on the Sunday, 00:01:57.680 |
September 8th, followed by two days of content, Dine About Town on the first night, two blowout 00:02:02.800 |
parties. It's going to be awesome. So we're really excited about it. We've got a much bigger budget 00:02:07.360 |
this year. And we're going to try and really make the experience awesome, while also delivering the 00:02:12.800 |
great content. So really excited to do this again in LA. I appreciate everyone's support of that. 00:02:18.400 |
We had a lot of debate internally about where to take it. And we settled on going with what we know 00:02:25.520 |
and trying to make it a little bit better this year. So anything else you guys want to say about 00:02:30.400 |
And there's only one kind of ticket this year, right? 00:02:32.960 |
That's right. So we're not doing like regular ticket and VIP. We're just doing one ticket 00:02:37.200 |
for everyone. And everyone's going to get this kind of elevated experience. Should be cool. 00:02:41.920 |
Saks, just show up. Would be awesome. Hope to see you there. 00:02:50.560 |
When are we going to start announcing the speakers, Preberg? 00:02:53.040 |
We will do that in a couple weeks. We already have a bunch booked. And it's gonna be awesome. 00:03:01.840 |
Yeah, I know CP, you wanted to take it somewhere else. I mean, 00:03:04.800 |
we were in debate on Vegas, maybe going to some other country. 00:03:09.840 |
But you went to another country. Where'd you go this week? 00:03:15.200 |
There's like a important AI conference. And Jonathan Ross, the founder of Grok and I kicked 00:03:22.560 |
it off with like a fireside chat for like 45 minutes at the beginning just to get everybody 00:03:35.360 |
I think the soundbite that made an impact on me was from Jonathan. So I think he did a very good 00:03:42.160 |
job of explaining where NVIDIA is really strong, which is really in learning, and where NVIDIA is 00:03:50.320 |
trying to build the business, which is an inference. But the problem is the features 00:03:55.680 |
almost compete with each other between learning and inference. And I think that that was good 00:04:00.960 |
clarification. And then the second thing is he said that if he deploys his roadmap, he'll have 00:04:05.600 |
50% of the available inference compute by the end of next year. 00:04:10.880 |
Based on all the needs of like open AI and meta and everybody else, which is a... 00:04:14.720 |
Wait, but that's a smaller physical footprint. Is that right? And it's just 00:04:19.680 |
Less power, higher tokens per second across all the major models. So it would be meaningfully 00:04:27.120 |
less power and meaningfully cheaper. So that was a pretty amazing soundbite. But his whole 00:04:33.600 |
explanation was actually really good. So I mean, if you listen to the spaces we did, 00:04:37.120 |
I thought it was good. I thought this was even better. So I'm trying to get a copy of the 00:04:41.120 |
talk because it was taped so that we can kind of post it. At a minimum, anybody who's interested 00:04:48.880 |
in studying NVIDIA should listen to what he has to say, because I think the nuances between us 00:04:54.320 |
and them are really profound. And they were pretty stark the way that he described them. 00:04:57.600 |
So it was cool. It was great to be in Paris, frankly. 00:04:59.760 |
How does NVIDIA respond to this movement or this notion that there is an entirely 00:05:04.720 |
different chip category focused purely on inference? Do they have inference focused 00:05:08.960 |
chips in development or a different roadmap that they're going to have to kick up now? 00:05:12.800 |
I mean, I think it's very hard for them. It is a structural innovator's dilemma. And why is that? 00:05:18.240 |
Because they've made some architectural decisions around things like HBM, 00:05:21.920 |
which is high bandwidth memory, or specific kinds of extremely high throughput optical cables. 00:05:27.840 |
And why these things matter is that they've gone and bought up the world supply, 00:05:31.920 |
and they've basically bearhugged these markets as a structural part of their design. Now, 00:05:38.400 |
that makes sense, and it's very legal. The problem is that if you architect around it and go through 00:05:44.880 |
a totally orthogonal design process, we're not subject to any of those supply chain considerations. 00:05:51.040 |
And I think he explained that really well. I didn't understand it in nearly as much detail 00:05:55.360 |
as I did leaving. So I think that's an important takeaway. It's also really important for all the 00:05:59.760 |
hundreds of chip startups and for these venture investors who are ripping money into these 00:06:03.760 |
companies to know these differences. Because if you are investing in a company that is subject to 00:06:09.840 |
HBM in their design decisions, or these specific optical cables, or a whole bunch of other things, 00:06:15.040 |
you're going to be in a really difficult spot because NVIDIA has bought them all. 00:06:20.320 |
So I thought that that was really interesting. The other thing that NVIDIA does brilliantly is 00:06:24.240 |
what Intel did in the '80s and '90s when they wanted to choke the market for CPU competition, 00:06:31.120 |
which is that they were able to define a metric that everybody started to pay attention to, 00:06:36.400 |
which for them was clock speed. And if you remember, Intel would push Moore's Law, 00:06:41.040 |
and they would push transistor density and clock speed as the measurement of value. 00:06:47.200 |
Now, for a lot of consumers, none of us knew any better. And so every time we saw a new chip, 00:06:54.080 |
we would go through the Pentium upgrade cycle because we thought that's what we were supposed 00:06:57.680 |
to do. It turns out that clock speed doesn't really factor into speed the way that you think 00:07:03.440 |
about it and experience it as an end user using software. And it's the same in AI. And so we 00:07:10.160 |
talked about that as well. So it's an incredibly important company, NVIDIA is. I think they're 00:07:15.520 |
performing brilliantly. They're running many of the same plays that Intel did to try to create 00:07:20.720 |
total dominance. But there's some very important nuances around power and around supply chain 00:07:28.000 |
decisions and technology decisions that, frankly, some people would think are unsustainable. 00:07:34.000 |
And the more I understand them, I would be in that camp as well. So I thought it was a very 00:07:39.440 |
important conversation and hopefully we can post it. Yeah, it's cool. And by the way, it's great 00:07:45.520 |
to be in Paris, too. What do you think of Paris? You know, I have a new appreciation for what's 00:07:49.920 |
going on in these major cities. I think that all major cities are crumbling. I don't and I don't 00:07:54.960 |
mean to be alarmist or very dramatic about that. It's just that I spent a lot of time with 00:07:59.840 |
entrepreneurs that are in Paris. And they struggle with the same things that the people in San 00:08:05.840 |
Francisco do. And then I had a couple of conversations with some folks that are entrepreneurs 00:08:10.000 |
in London who were there. And they suffer from the same things. Lots of petty crime, lots of garbage, 00:08:17.040 |
lots of vandalism, lots of drugs. And so no place is perfect. It turns out everybody is struggling 00:08:23.760 |
from a lot of experimentation gone wrong and how these cities run. That was that was actually my 00:08:29.280 |
the biggest takeaway outside of the actual AI conference itself. But that's a Western City 00:08:34.640 |
problem. You're not I mean, have you visited many cities in China or other parts of Asia? 00:08:39.040 |
And is that like a Western problem? I have not been to China in six years. I've been to Hong Kong, 00:08:46.720 |
but I've not been into China for six years. So I don't know. I would say that India has its own 00:08:53.440 |
vagaries. So those issues are not the same. I've been to Singapore a couple times. My kids were 00:08:59.920 |
just in Japan for spring break. So I saw a bunch of the pictures, those cities look relatively 00:09:04.240 |
the same, very orderly, well run clean. But yeah, you're right. These Western cities are a bit 00:09:10.320 |
chaotic. Yeah, sacks. Is this like a Western social liberal manifestation? How cities are 00:09:16.960 |
being challenged? Generally? Yes. Yes. Like Malay has, I think Malay said this, right? He said, 00:09:21.920 |
like, all these Western social liberal beliefs are having lifestyle impact and urban centers. 00:09:29.040 |
Well, it's hard for me to speak to what's happening in Europe, but we know what's happening 00:09:32.400 |
in the United States. All the biggest cities are blue, and they're run by a certain ideology. And 00:09:39.440 |
in those cities, we've had now a decade plus of decriminalization, deprosecution, decarceration. 00:09:46.960 |
And the impacts of this are obvious. We also, at least in San Francisco, we have subsidies 00:09:53.760 |
for people who want to live on the streets and do drugs all day. I mean, it's not prohibited to do 00:09:59.920 |
that. So all of our public spaces have been taken over by people who are happy to take the whatever 00:10:04.960 |
it is, $700, $800 subsidy and do drugs all day. We also had this referendum in California that 00:10:12.640 |
essentially lowered a whole bunch of felonies to misdemeanors. Basically, if you steal less than 00:10:18.960 |
$950, that's considered a misdemeanor now instead of a felony. And so there's been a huge increase 00:10:25.840 |
in crime because of that. You know, shoplifting just doesn't get prosecuted anymore. And 00:10:30.400 |
misdemeanors don't get prosecuted anymore. We've had all these SOARs, DAs who don't want to 00:10:33.520 |
prosecute misdemeanors. So I think if voters had understood that lowering a lot of these crimes 00:10:41.920 |
from felonies to misdemeanors meant that they wouldn't get prosecuted at all, I don't think 00:10:45.200 |
they would have done it. But that's kind of where we are. But are we seeing a rebound from that 00:10:48.800 |
now, Sax? Is democracy working? Because I know that in San Francisco, we had an election recently, 00:10:53.680 |
and in the ballots, it seemed to indicate that voters were sick and fed up with the living 00:11:01.120 |
situation in the city, and they're putting different people in office and have an intention 00:11:06.400 |
of changing the city charter and changing the laws in the city so that we can reverse this trend. 00:11:13.200 |
Is that not where things are? I mean, like, sitting where you sit, like, don't you see 00:11:18.320 |
San Francisco kind of trying to right itself? And as a result, doesn't democracy work over time? 00:11:24.000 |
Yeah, it's just that we live in a one-party state, and that one party has been captured by ideologues, 00:11:30.000 |
and so it's very hard to change. It is true that in San Francisco, there was an election recently, 00:11:35.040 |
and the more radical supervisors were defeated. We have, I think, finally a good DA in San Francisco. 00:11:43.520 |
But these radicals have had a long time to entrench their power, and the party machinery 00:11:48.960 |
helps keep them entrenched. So I think you can argue that we've bottomed out because the public 00:11:54.080 |
has recognized the problem, but it's still very hard to fix, and it's going to take a while. 00:11:57.840 |
I think for the first time in a long time, I'm more optimistic than not, given the people I've 00:12:03.040 |
seen that tolerated the conditions historically have reversed course over the last year and have 00:12:08.880 |
said, "We've got to act, and we've got to do our kind of civic process." 00:12:13.120 |
Well, time will tell. You have to remember, when New York City was a chaotic mess in the '70s and 00:12:18.160 |
'80s, it took two different mayors, and it took almost a decade for that to get cleaned up, but it 00:12:23.600 |
took the strength of leadership. First, David Dinkins set the framework, and then Giuliani really 00:12:29.280 |
executed Rudy Giuliani. And between the two of them, and I think what people call the broken 00:12:34.480 |
windows theory, they hired a lot more policemen, and they just really cleaned up all of these really 00:12:40.240 |
bad areas, and then it had this knock-on effect. The problem is, when I'm staying at a hotel in 00:12:47.600 |
Paris, or when I was speaking to some of my friends who live in London, it's the same thing. 00:12:52.400 |
It's like the amount of knife crime is so egregious, as an example, that they tell you, 00:12:58.000 |
"Don't wear a watch," et cetera. I mean, I don't wear watches anymore when I go out. 00:13:02.320 |
But what's crazy is they will tell you that you'll get stabbed over a phone, 00:13:06.000 |
where it's like, "You don't need to stab me for my iPhone here. Just take the iPhone." 00:13:10.800 |
So there's a level of lawlessness and petty crime that's affecting tourists 00:13:14.960 |
that, to me, was very surprising. And so even when I was walking around Paris, I was like, 00:13:21.600 |
"Should I really feel scared here?" And my reaction was, "No, I've never felt scared in 00:13:28.720 |
Paris." But then what's ringing in the back of my ears are these warnings from people that have 00:13:34.240 |
lived there for 10 and 20 years, who've had this stuff happen to them. That was really surprising 00:13:38.480 |
to me. So that stuff in San Francisco, as far as I can tell, or any major city in the West, 00:13:42.320 |
hasn't really been fixed, because you haven't had a shift in leadership yet, where they've 00:13:46.560 |
really said, "Okay, we're going to double down on policemen and policing, and we're going to 00:13:52.480 |
amp up the punishment." That hasn't happened. Well, I was just in Vegas this week, 00:13:57.040 |
and it doesn't feel... How was your trip? Give us the trip report. 00:13:59.600 |
Yeah, I went to the Google Next Conference. But I will say, first off, no crime in Vegas. 00:14:04.480 |
Vegas is Vegas. It's a very different environment. But yeah, I went to the Google Next Conference. 00:14:11.280 |
I ended up going out because at O'Halo, we did a partnership with GCP. So we're moving all of our 00:14:18.000 |
infrastructure and doing a lot of model development on GCP. And it was great. I think there was 30,000 00:14:23.920 |
people at this event, which was insane. Have you guys been to any of these, like AWS Reinvent, or 00:14:33.120 |
What I didn't realize prior to going to this was how much of an ecosystem there is. They 00:14:38.080 |
take over the whole convention center at Mandalay Bay. And there's hundreds of companies that have 00:14:44.240 |
built services or build services on GCP that are there, pitching CIOs and CTOs of other companies. 00:14:50.480 |
And so it's this real network effect type of business model that I hadn't fully appreciated 00:14:55.280 |
just how like, deep the network is and how entrenched it is. I also got to go to this 00:15:00.960 |
thing where I spent a lot of time with CIOs and CTOs at other companies, big companies. 00:15:05.120 |
And a big takeaway for me that I also didn't realize was how most of them are not on a single 00:15:11.680 |
cloud provider. Almost everyone I spoke with is on multiple clouds. So they're all on like AWS 00:15:19.600 |
and Azure and GCP. And they use different aspects of each cloud. And there's a lot of interoperability 00:15:27.200 |
between the clouds now, which I had previously my naive concept of this, because I'm obviously not 00:15:32.160 |
deeply in the space was that a company picks a cloud and they move everything over to that one 00:15:37.200 |
service provider. But it turns out that they split, and then they allocate dollars based on 00:15:42.240 |
the services and the pricing. So it was pretty clear that everyone is, it's like a race to the 00:15:48.640 |
bottom on pricing for basic commoditizing services in the cloud, like data storage, right? And then 00:15:55.520 |
whoever has the better higher level services, whoever has access to better models or has a 00:15:59.440 |
lower price on running models, or on running things like inference. That's where the money 00:16:05.280 |
is made. And the gross margin on that is quite different. And I think that's where a company 00:16:09.760 |
like Google probably has a much bigger advantage because of the uniqueness and the quality of the 00:16:16.320 |
models that they've developed. But yeah, I just didn't I didn't really appreciate like how much 00:16:20.240 |
there was also. So my other big takeaway was the cloud is a wide open competitive market right now. 00:16:25.680 |
Like there is no AWS lock in in this market. There's no Azure lock in. There's no GCP lock 00:16:30.720 |
in like you guys. Yeah. The lock in comes from the developer services that they provide. So 00:16:36.240 |
both Amazon and Google, and I'm assuming Azure as well, will give you access to incredibly talented 00:16:42.400 |
engineers who will help you build production code. Yeah. And the ecosystem on how so like a lot of 00:16:47.760 |
the clouds are competing for the consultants who actually do the work for the big enterprises. 00:16:54.400 |
So Capgemini, Deloitte, PwC, they have these massive booths and facilities, and then go and 00:17:00.480 |
build tools for companies is also a key part of the business strategy. And yeah, I mean, 00:17:05.600 |
it's such a competitive market. I did not appreciate just like how much of an open game 00:17:11.360 |
that still was. Sure. It was a great, great couple days in Vegas. And I got to see some 00:17:16.320 |
of our friends. We ducked in and saw Helmuth in the Poker Go studio for a minute. Some big 00:17:21.920 |
tournaments going on at the ARIA. It was fun. We had a good time. Okay, let's move on. So, 00:17:26.160 |
you know, on the docket today, we were going to talk about CPI. The March CPI numbers came in 00:17:31.680 |
from the Bureau of Labor Statistics yesterday, higher than forecast at 3.5% year over year. 00:17:38.080 |
This is up from 3.2% year over year inflation reported by the Bureau of Labor Statistics in 00:17:43.920 |
February, higher than expected for three straight months now. Let's pull up this chart. And as we 00:17:48.880 |
all know, the Fed did 11 rate hikes in 2022 and 23. And the intention was that we would see 00:17:54.960 |
inflation come down to their target of 2% and they would start to lower rates. And obviously, 00:18:00.560 |
we're not seeing that inflation is remaining hot. Things are getting more and more expensive. 00:18:07.440 |
I pulled together some of the numbers on year over year price increase. Housing is up 5.7% 00:18:14.400 |
year over year. Transportation costs are up 10.7% year over year. And the highest car insurance is 00:18:21.600 |
up 22.2% year over year. So the expectation has been that the market will do rate cuts. 00:18:29.360 |
Larry Summers came out and said you have to take seriously the possibility that the next rate move 00:18:33.360 |
will be upwards rather than downwards. I guess, Saks, let me let me ask you first for your 00:18:40.160 |
commentary on how this plays into the election cycle. I know Biden was expecting a rate cut or 00:18:46.000 |
several rate cuts going into the election, which usually helps the case of the incumbent. Maybe you 00:18:49.760 |
can comment on where the election cycle might be influenced by this inflation report. 00:18:56.160 |
Well, I think it's definitely bad news for Biden. I was expecting a Fed put this year. 00:19:00.480 |
And really, the whole market was. The market was expecting three rate cuts this year, 00:19:04.160 |
which would have been really good news for Biden. And I think this latest inflation 00:19:09.280 |
print was kind of a dagger to the heart of that expectation. And I think the Wall Street Journal 00:19:14.480 |
put it best. It said here, Wednesday's report had been hotly anticipated because Fed leaders 00:19:21.280 |
had been willing to play down stronger than anticipated inflation readings in January and 00:19:27.200 |
February as reflecting potential seasonal quirks. But a third straight month of above expectations 00:19:34.720 |
inflation data erodes that story and could lead Fed officials to postpone anticipated rate cuts 00:19:41.040 |
until July or later. So it's not just the fact that this inflation print was higher than expected. 00:19:46.400 |
It was also higher than expected in January and February. But people were willing to kind of 00:19:51.040 |
overlook that, saying, well, maybe it was just kind of a quirky reading. But now we've had three 00:19:56.640 |
straight months, it's pretty clear that the narrative that we had going into this year, 00:20:00.880 |
which was that inflation was on its way down, that it peaked at 9% as the official rate, I think, 00:20:07.600 |
in 2022, and then it was going down every month through all of 2023. And I think the expectation 00:20:14.480 |
going into '24 was it would keep going down and we'd get these rate cuts. Well, 00:20:18.000 |
after three straight months of inflation being more persistent and stickier than people expected, 00:20:24.480 |
I think that narrative is basically dead. So I think the new narrative now is it's going to be 00:20:30.800 |
rates are going to be higher longer. And I think Larry Summers is reflecting that view. Larry's 00:20:36.080 |
going further. He's not just saying that rates are going to be, you know, at the current rates 00:20:40.560 |
for longer, he's saying they might actually increase. And that's a risk. Yeah, last summer, 00:20:45.760 |
he said, and this was last summer. So almost a year ago, he was saying, the market's got it 00:20:50.000 |
totally wrong. We're actually going to need much higher rates than the market's anticipating for 00:20:53.840 |
much longer, as well, than the market is anticipating. And so did Mario Draghi, 00:20:58.400 |
they both said the same thing. No one paid attention to him, everyone ignored it and 00:21:02.160 |
assumed that this was going to be a quick rebound to normalization. It's clear that, 00:21:06.400 |
you know, once again, Larry has proven himself to be fairly prescient. 00:21:09.360 |
Larry Summers has, for the most part, been spot on from this on this whole inflation 00:21:15.760 |
question going all the way back to 2021. Remember, 100% in Q1 of 2021, Biden's first quarter in 00:21:22.240 |
office, the big legislative push was for the $2 trillion COVID relief bill is the so called 00:21:28.640 |
American Rescue Plan. And Larry Summers said that it was it risked inflation, it was an inflationary 00:21:34.160 |
bill. It was unnecessary. Yeah, we didn't need it. The economy was already coming back, and we 00:21:38.560 |
didn't need it. And Biden was risking inflation. But of course, inflation was only at 2% at that 00:21:43.600 |
point. So the administration kind of poopooed Larry and said, you know, this is Larry Summers being 00:21:49.520 |
Larry or whatever. And sure enough, we were at 5% inflation by that summer. And you have to wonder 00:21:56.240 |
if Biden had listened to that advice. Would he be in a different position right now going into the 00:22:01.920 |
election? I think probably he would have been. And you got to wonder for what I mean, what did 00:22:06.160 |
that $2 trillion accomplish? I mean, COVID was winding down. I mean, these were the last bit of 00:22:12.800 |
stimmy checks and payments to these pharma companies. And I mean, it was basically a grab 00:22:18.480 |
bag and it was passed on straight party lines. And it was just totally unnecessary. I mean, 00:22:23.280 |
the economy certainly didn't need it. And so here we are. And I think that a lot of people, 00:22:29.280 |
including the markets thought that Biden was kind of out of the woods that this year, 00:22:35.280 |
we'd see the final leg of inflation going back to normal, but that's not going to happen. 00:22:40.400 |
And I think, you know, going beyond the political ramifications, I think there could be 00:22:44.640 |
several other knock-on effects that we should talk about. I mean, one is for the consumer. 00:22:50.640 |
This means that the cost of borrowing is going to be higher for longer. That makes it harder 00:22:56.320 |
to buy a house. Mortgage payments are higher. That also means if there's fewer home sale 00:23:01.920 |
transactions, that means the price of housing could come down. So there could be a correction 00:23:06.480 |
in that market. Second, if you want to buy a car, your car payments higher. And if you have 00:23:11.040 |
loans, your personal interest is going to be higher. And this is why I think consumers feel 00:23:16.240 |
like they're worse off than the economic data would otherwise reflect. And Larry Summers actually 00:23:21.440 |
had a tweet storm about this about a month ago, where he calculated that if you included the cost 00:23:26.320 |
of borrowing in inflation, that inflation was much higher than people thought and that it actually 00:23:32.320 |
peaked not at 9%, but at 18%. So Larry had an excellent tweet storm on that. And he said that 00:23:38.400 |
the cost of borrowing, which used to be calculated in inflation but is not anymore, was the reason 00:23:44.400 |
why consumer sentiment about the economy was depressed. He said that that accounted for about 00:23:49.600 |
70% of it. So people should be feeling better off because GDP growth is good and unemployment is 00:23:56.960 |
low, but they're not because inflation is so high. And if you include cost of borrowing, 00:24:01.440 |
the inflation is even higher. So I think this is really bad news for Biden. But it's a story 00:24:07.120 |
that's been going on now for three years. It's not just this one inflation print. 00:24:11.920 |
Right. It's funny. Yesterday, Biden said, during a press conference with Japanese Prime Minister 00:24:18.960 |
Kishida, that he's sticking with his prediction of a rate cut before years end, but there might 00:24:24.240 |
be a delay, which I think is obviously, you know, based on what the data showing and what 00:24:30.480 |
folks that seem to know and have been pretty good predictors are saying is unlikely. 00:24:35.040 |
We were supposed to get one in April. We were supposed to get one in June. 00:24:38.560 |
Oh, yeah. Remember, the Fed Open Market Committee in December 00:24:41.520 |
stated that they expected three rate cuts this year. And I think the market, in some cases, 00:24:46.640 |
were predicting as high as four or five. Right. 00:24:49.520 |
And now we may not have any and we may actually see a rate hike if you were to follow. 00:24:54.480 |
Well, if we see a rate hike before the election, I think Biden is toast. 00:24:58.800 |
Yeah. But I mean, I think Summers gave that maybe 00:25:01.040 |
a 15 to 20% chance, which is it's still not the most likely scenario. 00:25:05.600 |
But it's possible. But it's possible. And no one was counting on that in December. 00:25:10.000 |
Chamath, if you're a Fed governor, and obviously, the Federal Reserve Board of Governors sets the 00:25:16.640 |
rates, they meet and they make these decisions. How influenced are they politically? How influenced 00:25:22.880 |
are they by this election cycle? And, you know, historically, this is meant to be a fairly 00:25:26.960 |
independent board. But there's been a lot of consternation over the last few years that this 00:25:32.560 |
board acts like almost a political apparatus, particularly in election years. 00:25:42.400 |
It turns out that before I answer your question, that was a joke. Nick, you can please show the 00:25:51.760 |
tweet. But there was an article that came out that said one of the Fed governors has apparently 00:25:57.200 |
massively plagiarized most of the work that got her to the position of being Fed governor. 00:26:03.840 |
So the same thing that happened to the president of Harvard. So TBD, what happens to that Fed 00:26:08.000 |
governor? But it seems, you know, my observation there was just more that it's like the preferred 00:26:12.880 |
method of defenestration now of academics and the apparatchik, right? Before, if you had 00:26:19.440 |
right leaning views or centrist views, you get run out of town and get fired. Now, all of those folks 00:26:25.040 |
are using the search engines to basically figure out really quickly that you plagiarized a large 00:26:31.120 |
quantity of your academic work, and that your scholarship is is a little bit more speculative, 00:26:38.160 |
which then brings into question, why do you get the right to help set the policy of the most 00:26:43.040 |
important economy in the world? Well, anyways, so that that was that joke. But so I think that 00:26:47.680 |
you're asking the absolute right question. I think that the Fed has become increasingly politicized. 00:26:55.280 |
And I do think that there are actions that foreign governments take to influence the election, 00:27:04.240 |
but I do not think it's what most people think. When I say that most people's ideas goes to like 00:27:10.080 |
some darkroom conspiracy theory and misinformation. And I think that's a little bit naive. 00:27:16.640 |
I think the more obvious way in which other governments can impact the US election is 00:27:23.040 |
exactly how SAC said, which is they know that if inflation ticks high, and interest rates are up, 00:27:28.880 |
and consumers are under pressure and meaningfully displeased with the economy, they'll vote out the 00:27:32.880 |
sitting administration. And so what can other governments do? They can do what they're doing. 00:27:39.040 |
So Nick, I love this like, explanation and pre charts, Nick chart number one, please. 00:27:45.120 |
What is this? So this is a core driver of inflation, which is the price of energy. 00:27:51.760 |
And what you see here is that since the beginning of the year, oil prices have gone up, 00:27:58.800 |
which ultimately will translate to into a higher cost of doing business, right? Running your 00:28:03.680 |
factory, keeping the lights on, keeping your home warm, getting from point A to point B, 00:28:09.760 |
all of these things go up, which ultimately get reflected in prices, which is what inflation is. 00:28:14.400 |
And so that's a very bad trend. Obviously, we had a lot of really good stuff happen 00:28:19.520 |
through the course of last year, but it's largely reversed itself. So it's important to double click 00:28:25.360 |
into this and figure out, well, who's driving this? Is it just a random act? Or is it a systemic set 00:28:31.600 |
of decisions? And it turns out it's systemic. So if you go to the next chart, what this shows you 00:28:37.040 |
is what the US has been doing. So the United States has clearly created an incentive to flood the 00:28:44.160 |
economy with oil, which the laws of supply and demand would say prices should go down. 00:28:50.800 |
And by bringing prices further down, they would have further brought down inflation, 00:28:56.080 |
and they would have pulled forward these rate cuts that we had been expecting from the Fed. 00:29:00.720 |
The problem is that the rest of the world actually said, no, hold on. And that's what you can see on 00:29:06.160 |
the next chart. So for every time we were trying to increase production, all of these other countries 00:29:12.240 |
have been cutting production. And this is the change in the OPEC plus crude production targets 00:29:17.280 |
of 2024. So what you see here is for every barrel of oil, the US would pump out of the ground, 00:29:23.920 |
more than that one barrel of oil would get actually restricted by all of these countries 00:29:27.840 |
on this list, which is just to show you that these countries have a very specific view 00:29:32.080 |
about how they want that energy market as and that's just one market, but it's an input a key 00:29:38.160 |
input into inflation and rates to work. So you're saying that the OPEC has a price per barrel price 00:29:45.680 |
target, and they're able to manipulate production to maintain that price target. So while the US is 00:29:50.160 |
trying to lower the price per barrel, these other countries are are able to have a stronger 00:29:54.560 |
influence and balance out to get when you when you make these cuts like this prices will go up 00:29:59.520 |
because OPEC plus is meaningfully bigger, the sum of all the parts of OPEC plus is meaningfully 00:30:04.320 |
bigger than just the United States on its own. And so the United States may amp up a few 100,000 00:30:10.080 |
barrels a day. But if OPEC cuts by a few million barrels a day, you see what you saw in that first 00:30:16.640 |
chart, which is the price will go up. And that's largely why prices have gone up. So I think why 00:30:21.440 |
this happens is because when they see the United States, trying to manage inflation and manage 00:30:27.920 |
global rates, right, because us rates are not just for the United States, they're the, they're the 00:30:32.240 |
trendsetter for everything else. I think that in part, there's a decision that this rate philosophy, 00:30:37.760 |
the economic philosophy, the political philosophy, if they wanted to support that, what they would do 00:30:43.760 |
is actually also keep prices where they were or increase production. By reigning in production, 00:30:49.120 |
what you're essentially doing is voting against that entire apparatus. Right? So OPEC, and all of 00:30:55.120 |
those countries, and those decisions are not made by an oil minister, let's be honest, they're made 00:30:58.320 |
by the leaders of those countries, whose decision then gets reflected by the minister of petroleum 00:31:04.080 |
or oil in all of these countries, is a vote against Biden, and the United States government 00:31:09.520 |
and the politicization potentially of the Fed. And then the last comment is just that when you 00:31:13.680 |
look at the reverse repo rate, which is a fundamental measure of liquidity, what you 00:31:17.840 |
also see is liquidity very quickly getting sucked out of the system, which is also generally a 00:31:23.840 |
system health check that the Fed uses to make sure that the system is operating as required. So 00:31:28.480 |
in totality, what I see is that there is a structural disillusionment with the current 00:31:38.000 |
administration, and where people can make decisions in their own best interests, 00:31:44.480 |
versus the collective interests of the United States in a broader framework that the US may 00:31:49.280 |
represent are voting against it. And so this is why the idea of a persistent inflation rate 00:31:55.520 |
is a lot more credible than it was six months ago. And I think, sorry, just the last point, 00:32:01.920 |
and I think it's just what Zack says, which means that we're probably now on balance 50/50 between 00:32:07.040 |
a hike and a cut. And I think if you don't see this thing change in the next three months, 00:32:11.120 |
you're going to see 75/25 for a hike of at least 25 basis points. And I do think David's right. I 00:32:18.080 |
don't see how the sitting administration in any government will be able to withstand the onslaught 00:32:24.720 |
of electoral displeasure, if rates are four and a half, five and 6% going into November. 00:32:32.000 |
Timoth, if that is, I think everything you said makes sense going into this 00:32:37.040 |
election rates are high and and or going higher. What is the Biden administration do? 00:32:43.840 |
The fact is, they are very likely going to come up with a strategy to try and win votes that will 00:32:51.280 |
likely involve some sort of stimulatory process would be my guess. And that stimulatory process 00:32:57.120 |
may look something like more student loan forgiveness, more debt forgiveness, Kamala 00:33:01.760 |
Harris. She had a whole speech about debt. And here's here's the irony, the actions that the 00:33:08.480 |
administration is likely to take to support themselves in the election cycle are likely 00:33:14.640 |
going to further fuel inflation. And that's the biggest concern and worry I have is that there's 00:33:20.080 |
almost this inevitable set of behaviors that are coming down the pipe here over the next few months 00:33:24.480 |
because of the inflation report that are actually going to make things worse, not better. 00:33:28.720 |
I mean, to your point, Biden just announced a new student loan forgiveness plan. Remember, 00:33:32.560 |
the Supreme Court invalidated his last one. So now he's trying again. And it would potentially 00:33:37.440 |
give loan forgiveness to something like 30 million borrowers. So yeah, they're going to try and buy 00:33:42.080 |
votes. But I think it's already very late in the game to be trying to do that. I think that's what 00:33:46.640 |
they've been doing over the past few years. I mean, Biden has been a huge spender. You know, 00:33:51.760 |
he passed the COVID relief bill, he passed the Inflation Reduction Act, which was a ridiculous 00:33:56.800 |
name for was it 750 billion of, of more spending. And there was the CHIPS Act, which was well, 00:34:04.880 |
yeah, and there's the CHIPS Act and the infrastructure bill. And then remember, 00:34:08.400 |
he wanted four and a half trillion of build back better on top of that. And he wanted even more, 00:34:13.040 |
he didn't get any event. I think that this is a classic case of Biden spent the last three years 00:34:19.280 |
making his bed and now he's gonna have to sleep in it. And if in fact, inflation is persistent, 00:34:25.520 |
sticky and not going away, and he doesn't get rate cuts, again, this may cost him the election. 00:34:29.600 |
But, but look, I want to move beyond the political aspect of this and just talk about some more of 00:34:33.680 |
these knock on effects, because we talked about the impact on the consumer. There's also an impact 00:34:39.680 |
on investors, there's an impact on the government, and there's an impact on on the banks. So let's 00:34:46.240 |
just deal with the banks for a second. We've talked about on the show how commercial real 00:34:50.720 |
estate developers are hanging on by their fingernails right now. You know, they, they 00:34:55.840 |
basically borrowed huge amounts of money to buy commercial real estate, when rates were really low, 00:35:01.920 |
and they did so at high valuations. Now there's been a huge correction on that market, and they 00:35:05.680 |
cannot pay back those loans. And they can't get refinance. So what's happening? Well, they're 00:35:11.280 |
working with the banks to restructure those loans to tack on the interest they can't pay now, 00:35:16.560 |
to the end of the loan. And then the bank will hope and pretend that they'll be able to get 00:35:20.960 |
paid back one day. It's called extend and pretend. Well, the extended pretend strategy could work if 00:35:26.720 |
you believe that rate cuts are coming really soon, because you just need to hold on until the rate 00:35:31.280 |
cuts are here, then you can get refinanced at lower rates. But if rates are going to be higher 00:35:35.840 |
longer, that strategy is not going to work. And you're going to see more and more real estate 00:35:39.360 |
sponsors throwing in the towel. And banks are going to end up foreclosing on these properties, 00:35:44.240 |
and you're going to see more and more fire sales. There's a really dramatic example just the other 00:35:48.160 |
day in St. Louis, a building that was bought a few years ago for $200 million just fire 00:35:53.280 |
sailed at $2 million. I mean, 99% reduction. This is like dot com crash level reduction. 00:36:00.320 |
No, this is like the price of the deal. I mean, I assume the equity was completely wiped out. Yeah, 00:36:06.080 |
I guess they sold it for three and a half million. 00:36:07.920 |
Are you sure that's not the equity? That's got to be the equity. There's no way that's the- 00:36:15.120 |
With the debt. It's got to have the debt attached to it for that, yeah. 00:36:17.520 |
I mean, either way, it's a total wipe out for the equity holders. 00:36:21.040 |
But it's also going to be a wipe out for the bank that made that loan originally, 00:36:25.200 |
or that bought that loan when it was repackaged into CMBS or whatever it was. 00:36:29.520 |
So I would expect that if rates stay higher longer, that's going to create tremendous 00:36:35.040 |
stress on commercial real estate, and therefore on the regional banking system that made all these 00:36:40.080 |
loans to commercial real estate developers. So that's knock-on effect number one. Knock-on effect 00:36:46.000 |
number two is the government's cost of borrowing. The 10-year rate's already risen to 4.5% now. 00:36:51.280 |
And if rates stay higher longer, as more and more of the government debt rolls, 00:36:55.600 |
it's going to end up getting refinanced at higher rates. And the need to refinance 00:36:59.440 |
all of this government debt is going to continue to put pressure on the bond markets. 00:37:02.800 |
And I think this is why Larry is fundamentally pretty bearish on rates going down is the 00:37:08.240 |
government's financing costs are so huge, and they continue to grow. We're adding something 00:37:13.040 |
like a trillion dollars of new debt every 100 days. There've been some reports of this. 00:37:17.680 |
Yeah, 7.6 trillion is getting refinanced of old debt in the next 12 months. 00:37:24.160 |
And that 7.6 is sitting around 2%, and now it's going to get bumped up to close to 5%, 00:37:31.520 |
which adds an incremental $210 billion a year of debt service expense just on the debt that's 00:37:39.040 |
getting refinanced in the next 12 months, not to mention the deficit, not to mention, 00:37:44.640 |
Right. So the government was extending and pretending as well. I mean, they were hoping 00:37:49.040 |
that rates would go down fast enough that when all of this massive government debt rolls and 00:37:54.240 |
needs to be refinanced, it'd be done at lower rates. So the interest costs wouldn't swamp us. 00:37:58.880 |
We're already, I think, at over a trillion dollars now of annual interest expense for 00:38:03.760 |
We are. Yeah, and it's going to rise to 1.6 trillion. So it's going to be 60% 00:38:08.880 |
greater than our defense budget, just servicing the old debt. 00:38:12.000 |
We're basically in a classic debt spiral where we're borrowing. Our annual deficit 00:38:16.480 |
is we're at 2 trillion plus now. So we're not paying back the principal on the debt. 00:38:22.240 |
We're not even paying back the interest on the debt. We're borrowing money to pay off 00:38:26.880 |
all the interest and then some on top of that. So that's a huge problem for the federal government. 00:38:32.400 |
And then I think the last piece of this is investors. If you pull up this Fred chart 00:38:37.680 |
on the Fed funds rate over the very, very long term. So what this chart shows, 00:38:43.840 |
if you just zoom in on this, is the Fed funds rate basically over the last, I don't know, 00:38:49.120 |
like 75 years. And you can see that it peaked in the early 1980s when Volcker decided to 00:38:56.560 |
break the back of inflation by jacking up interest rates to like 18%. And it was very painful. We had 00:39:02.960 |
a pretty severe recession and I think it was '81 or '82. But then the economy came roaring back by 00:39:09.520 |
'83 and Reagan got elected in '84. In any event, that began a period of declining and low interest 00:39:16.720 |
rates for investors, again, starting in the mid '80s. And it continued all the way through, 00:39:24.640 |
I mean, especially through I think the post-GFC period and then this COVID period where we had 00:39:30.480 |
zero interest rates for a long period of time. And you can see the last spike we've had was when the 00:39:35.440 |
Fed decided to jack up rates to 5% in response to inflation. So the fact that we've got this 00:39:40.880 |
persistent sticky inflation suggests that this whole, you could call it really almost a 40-year 00:39:47.600 |
period of declining and low interest rates may be over. And when interest rates are going down, 00:39:53.040 |
it creates a huge tailwind to the stock market and the bond market because the discount rate is lower, 00:40:00.880 |
so stocks and bonds are going to be more valuable. Well, if we're in a long-term environment that's 00:40:06.960 |
more inflationary and rates have to stay higher longer, that's not great news for investors. 00:40:12.720 |
It hurts the stock market and it creates in the private investing world like VC and real estate, 00:40:18.960 |
it creates a much higher hurdle rate. So it's harder for private investors to justify 00:40:25.360 |
fundraising. So I think it's too early to say that this 40-year trend is over, but I think 00:40:32.080 |
it could be. And the reason for that is just the huge pressure of government borrowing, 00:40:38.960 |
crowding out private investing and this sort of ever-present need to keep issuing more and more 00:40:46.320 |
debt at higher and higher rates to make it more attractive is going to create, I think, 00:40:50.400 |
very bad conditions for investors. So Chamath, let me ask you if you concur, 00:40:56.720 |
given that today where we're sitting as of recording right now, towards the end of market 00:41:03.200 |
on Thursday, April 11, the NASDAQ is sitting at an all-time high, and the NASDAQ is up 1.6% today. 00:41:10.560 |
Following this inflation report yesterday, what's the disconnect in the market on why 00:41:16.560 |
investors are buying equities, given that we're likely going to have persistent higher rates, 00:41:23.520 |
and as a result, multiples should compress, particularly in growth stocks, 00:41:27.200 |
and why the NASDAQ is kind of firing up? I think it's important to look at what 00:41:34.240 |
stocks are actually lifting the NASDAQ, just like what stocks are lifting the S&P 500. And 00:41:39.840 |
we've seen this. There's a massive dispersion. There's like seven or eight companies that matter 00:41:43.840 |
and 500-odd companies that don't matter much. And so the reason why things go up in an environment 00:41:50.560 |
like that is, irrespective of the economy, money managers are paid to deploy money. 00:41:55.120 |
And I don't know if you guys saw this, but Nick, maybe you can find it. I linked to it in my 00:41:59.840 |
newsletter last Sunday, which was an article about the fact that the global amount of total 00:42:05.520 |
available equities to buy is shrinking, right? So if you think about that, what that means is that 00:42:12.000 |
the number of dollars is growing, right? Linearly, and in some years, it seems exponentially, 00:42:18.480 |
but then the end, so then a certain percentage of that finds its way into the equity markets. 00:42:24.560 |
But when the equity markets continue to shrink, and there's fewer and fewer things to buy, 00:42:28.160 |
prices will still go up, irrespective of the underlying justification. 00:42:33.520 |
So I think that the thing is, you can't look at these things as a sign per se. You know, 00:42:38.240 |
if you look at commodities as a different example, Nick, I sent you this. 00:42:41.840 |
Commodities are up, broadly speaking, 30% since the beginning of the year. 00:42:47.440 |
I think the reality is the following. Countries are feeling increasingly uncomfortable 00:42:58.480 |
with the decisions that the United States are making, whether that's its position to keep 00:43:04.080 |
funding the war in Ukraine, whether that it's the continued instability in the Middle East, 00:43:08.640 |
whether it's posturing with China, whatever it is. And so every country is now making a very 00:43:14.560 |
different calculation about what they want to do. And I think how it gets reflected is that 00:43:18.320 |
they're optimizing for their own economies. And what that yields is each of them trying to make 00:43:23.920 |
as much money as they can in the short term. And as a country that is a net importer, we are going 00:43:30.560 |
to feel the price of those decisions as inflation. And I think Sachs is right. Like, that's where we 00:43:35.760 |
have to start figuring out how we rebalance these decisions from a foreign policy and economic 00:43:42.000 |
perspective. Because if you don't change something here, all of these countries are basically going 00:43:45.920 |
to be like, "We're on our own. We're going to optimize for ourselves." These countries are 00:43:50.480 |
basically voting in a way that says we are not being led in a way that we trust. And that 00:43:55.520 |
unfortunately falls at the feet of the sitting president. Yeah. Well, look, I'm hopeful, 00:44:01.120 |
as we all are, that work in the tech industry, that new tools and AI and other technologies would 00:44:09.600 |
have a massive driver on productivity, which net-net could help even this out. That's the one 00:44:15.600 |
last great hope. Sitting here, AI obviously viewed as being highly disruptive by non-technologists 00:44:23.120 |
and deeply concerning, and particularly in DC. But what is that thing that Bill Gates says, 00:44:28.800 |
things are overestimated in the short term and deeply underestimated in the long term? I think 00:44:33.680 |
that that's what AI is. That's right. And so AI is a bunch of, no offense, probabilities and 00:44:39.760 |
statistics today. Okay? And so we talked about this last time, so we should debunk this whole 00:44:45.200 |
thing of like, there's some glowing brain somewhere that's capable of doing everything. 00:44:49.440 |
Doesn't mean that we shouldn't underestimate it, being capable of this in five or 10 years, 00:44:56.240 |
which is probably the window in which we have some form of AGI, but today it's not that. And so the 00:45:01.360 |
great leaps in productivity, Freeberg, that you're talking about are probably a three to five year 00:45:05.360 |
phenomenon, not a three to five months. So what SAC says is right, which is how does the sitting 00:45:11.200 |
administration buy votes today? Stimulatory methods. 00:45:14.480 |
It'll have to make a ton of promises. The problem is it can't really work its way into the economy 00:45:19.360 |
fast enough come the November election. And so that's where I think they're in a really tough 00:45:25.040 |
spot, which means that then, and I hate to be the bearer of bad news, but this is where you have a 00:45:31.280 |
lot of these wag the dog kind of scenarios that come to mind, because then the only way to really 00:45:37.040 |
galvanize a populace is to distract them. Yeah. Well, look, let's talk a little bit 00:45:41.920 |
more about AI, because I do think that regardless of the near term and longer term implications, 00:45:49.840 |
there is a peaking of fear, peaking of interest, both different spelling of the word peak. 00:45:56.480 |
Adam Schiff yesterday proposed a bill, sorry, earlier this week called the Generative AI 00:46:01.600 |
Copyright Disclosure Act, in response to, I think, over 100 musicians signing a letter 00:46:08.960 |
saying they're concerned that their copyrighted works are being used to train AI models. 00:46:13.920 |
And Nick, if you'll pull up the text from the bill that I pulled out here real quick, 00:46:18.960 |
we can take just a quick look. It's a very short bill. It's only five pages long. 00:46:24.000 |
And it defines a generative AI model that's a combination of computer code and numerical 00:46:29.600 |
values designed to use artificial intelligence to generate outputs in the form of text, images, 00:46:35.760 |
audio, or video, and that it substantially incorporates one or more models as a system, 00:46:42.320 |
and it's designed for use by consumers. And what the bill then asks is that anyone developing 00:46:48.000 |
these models has to submit to a register run by the federal government, a list of all of the data 00:46:55.840 |
that they use that might be copyrighted to train the model and that this becomes a precursor to 00:47:02.160 |
being able to legally develop and use AI models that you actually have to register the training 00:47:07.200 |
data. So I have a rant on this, which I'm going to preserve for the end given my rights as moderator 00:47:13.360 |
here today. But I'd love the reaction from Sachs first on Schiff's proposed bill to moth. And we'll 00:47:19.280 |
run this one around the horn pretty quick. Well, first of all, I think it's too soon for this 00:47:23.840 |
legislation. I mean, the first thing that needs to happen is we need to get the question of fair 00:47:28.480 |
use arbitrated by the courts. We've talked about this before. People publish lots of information 00:47:34.640 |
on the internet. And I guess technically, it's copyrighted, but it's been published 00:47:39.600 |
in a completely public way. And humans are obviously allowed to read it and use it to 00:47:44.880 |
generate their own work. I think there's a question about whether AIs are allowed to 00:47:49.120 |
learn from it as well. And I would argue that they should be to some degree under fair use. 00:47:54.400 |
Maybe if you're talking about copyrighted songs, I don't know if that's different. Anyway, 00:47:59.120 |
these questions need to be arbitrated. And once we know that, then we'll know whether there's a 00:48:04.720 |
need for a rights clearinghouse of some kind, which may not be a bad idea. I actually got pitched 00:48:09.760 |
by a founder recently to create a rights marketplace with copyright holders on the 00:48:19.040 |
one side or rights holders could be musicians, could be movie studios, could be anyone who owns 00:48:24.960 |
content, and then AI companies on the other that want to license that content for 00:48:29.440 |
training purposes. And I actually thought it was a pretty good idea. So my point is just that 00:48:35.760 |
there are entrepreneurs working on solving this problem as well. And I guess what I'm saying is, 00:48:40.320 |
I'd like to see fair use figured out. And I'd like to see what the private solutions are going to be 00:48:45.520 |
before we try to legislate this problem away from really a point of ignorance, because we're so 00:48:52.320 |
early in the development of this market. Now, I think one of the questions you have to ask here 00:48:57.040 |
is, why did Adam Schiff introduce this bill? And who is Adam Schiff? Adam Schiff is a powerful 00:49:03.040 |
California congressman who is the candidate for US Senate right now. And he's expected to win. 00:49:09.840 |
He's ahead of Steve Garvey, who's the Republican candidate by like 30 points. 00:49:13.040 |
Schiff is, in addition to being a vicious partisan who's known for for Russiagate, 00:49:21.840 |
he also is a prodigious fundraiser. And when he's not in front of the cameras, he's very effective 00:49:28.080 |
in closed doors appealing to the key special interests in California that matter. And my 00:49:34.400 |
guess is that he's not carrying water here for the 100 musicians who signed that letter. 00:49:38.480 |
He's carrying water for the big studios, the big Hollywood studios, including music labels, 00:49:44.160 |
but also the movie studios. And he's an LA based congressman. Yeah, I mean, that's who he's 00:49:49.520 |
representing. Now, at the same time, my guess is that he's run this by the big tech companies, 00:49:54.800 |
and he's probably got their blessing to some degree, because like Bill Gurley always points out, 00:49:59.200 |
they like rules like this, because of regulatory capture, they can comply with them. But the little 00:50:04.480 |
guy, the entrepreneur who's trying to create something new has a much harder time 00:50:08.480 |
complying with these regulations. So this is basically a big special interest group 00:50:13.920 |
sort of deal that I think he's he's putting forward. 00:50:19.440 |
Not yet, but it's paving the way for something. 00:50:22.560 |
And it's more like his promotional effort during his campaign effectively. 00:50:32.160 |
yeah, I mean, the big LA powers that be like this, and they love this. Yeah, yeah. 00:50:43.360 |
I think there was an article in The Verge that I posted. 00:50:48.080 |
It turns out that OpenAI transcribed over a million hours of YouTube videos to train GPT-4. 00:50:54.000 |
My takeaway when I read this was, well, if Google's not going to sue OpenAI, 00:51:00.480 |
then this is a moot point. Because you're talking about one of the most valuable companies in the 00:51:08.240 |
world with one of the most valuable sources of training data in the world. And so if that if 00:51:14.480 |
it's true that there was a copyright violation, and Google just lets us go, what's anybody else 00:51:20.400 |
supposed to do? What chance does these group of singer songwriters have? And you know, no offense 00:51:25.760 |
to Adam Schiff, but he's not going to be able to do anything. So if this multi trillion dollar 00:51:32.080 |
company doesn't draw a line in the sand, then none of this matters. And everybody's going to 00:51:36.640 |
train on everything. I think, like, my big takeaway on all this is that there's a real 00:51:44.080 |
deep misunderstanding on what these models really are, and what they're doing. The assumption, 00:51:57.120 |
I think, naively, is that they take in a bunch of data, and then they spit that data back out. 00:52:06.080 |
And in some cases, maybe slightly transformed. But the truth is that the models don't actually 00:52:11.680 |
hold the data that they're trained on. They develop a bunch of synthesized predictors, 00:52:19.040 |
that they learn what the prediction could or should be from the data. And this is similar 00:52:25.440 |
to how humans operate. Can you pull up the article, Nick that I posted? This article is from last 00:52:32.720 |
May. And you guys may remember this case. I can't remember if we talked about it on the show. 00:52:37.120 |
But Ed Sheeran was sued by the estate of the writer of the Marvin Gaye song, Let's Get It On, 00:52:46.240 |
for infringement in his song Thinking Out Loud. And he actually prevailed in court, 00:52:53.040 |
where he went through with the jury, how he comes up with how he runs his creative process, 00:52:58.640 |
how he came up with the song, and how he does his work. And as you know, Ed Sheeran, 00:53:04.560 |
and nearly every musician or every artist listens to other people's music, looks at other people's 00:53:10.640 |
art, and they synthesize that data, they synthesize that knowledge and that experience 00:53:15.280 |
into their own creative process to output what in their mind is a novel creation. And I think that 00:53:24.240 |
AI works in a similar vein in that it's not actually storing the images, it's not actually 00:53:29.280 |
storing the music of third parties. It's learning from that process. We all like musicians learn on 00:53:36.560 |
classical music, they listen to and play other pop artists music. And then they learn things that 00:53:41.600 |
they like things that they don't like things that fit together well, certain intonation, 00:53:46.720 |
syncopations, and that's how they develop their own tracks. And so I think that the assumption 00:53:52.720 |
in AI is that it's almost guilty until proven innocent is what's going to really challenge this 00:53:58.800 |
technology, getting mainstay appeal and being useful in the way that it could and should be 00:54:04.880 |
that the de facto is that it's learning and training is copyright. I don't I don't buy that. 00:54:14.880 |
No, that that just because you're learning like a human, it's not copyright infringement. And 00:54:20.640 |
I'm sorry, let me let me make my last statement, which is that I do think that similarity is 00:54:24.480 |
measurable. And that is copyright infringement. So I would rather than focus on how you train 00:54:29.440 |
the model, I would encourage legislators to find, as Saks points out, a better process 00:54:35.840 |
to define what makes one thing distinct from another, run that analysis, and then use that 00:54:42.000 |
analysis to determine uniqueness. And that analysis is ultimately what should determine whether 00:54:47.360 |
there is copyright infringement versus the how was the model trained exercise, which is wrought 00:54:52.720 |
with challenges, as you guys know, how do you determine weightings? How do you determine 00:54:56.240 |
whether this was even relevant in the model? We don't even know if the models ignore the data or 00:55:00.800 |
use the data in their synthesis or in their weightings. It's all very opaque. And so it's 00:55:05.680 |
almost like, hey, here's a list of stuff that I read. But I may not have even paid attention to 00:55:09.760 |
95% of it. When I was reading it. We don't know. And so I'd rather focus on is there copyright 00:55:15.280 |
infringement in the output? Does the output mimic an original copyrighted source? And if it does, 00:55:21.280 |
then that's where the infringement should lie, not on the How did you do it? 00:55:25.520 |
I think that you're creating an impossible standard. Is it 5%? The same? Is it 8%? The same 00:55:30.000 |
is that 14%? Where's your threshold? Is it different than training? How's that different 00:55:34.480 |
than training? If I'm JK Rowling? Am I not allowed to read other authors copyrighted books for fear 00:55:40.000 |
that by reading it, I will now use their copyrighted material in my novel book? No. Okay, 00:55:45.520 |
stick with that example. But we can also use the Ed Sheeran example. Nick, can you play this video? 00:55:49.520 |
I could play you every song in the pop charts right now over four chords. You name any song 00:55:53.600 |
and I'll play it in the four chords. So these, these are the chords. Messenger. Your friend 00:56:00.240 |
because you only need the light when it's burning low. Only miss the song when it starts to snow. 00:56:07.040 |
Only know your lover when you let it go. Let it be. I'm walking away Craig David. 00:56:11.040 |
Greg David. I'm walking away from the troubles in my life. Last one. Let it be Beatles. 00:56:18.880 |
When I find myself in times of trouble. Why is that important? Obviously, it's not illegal for 00:56:28.080 |
Ed to have learned these four chords. But if he tries to produce those songs and calls them 00:56:33.040 |
something else, that's clearly somebody else's IP. Similarly, I think that if you don't give 00:56:39.520 |
people the right to say, yeah, of course, the letters A, B, C, D, E, F, G, all the way to Z 00:56:44.000 |
are not my invention. It's the alphabet. I don't control that. I don't control words and their 00:56:48.080 |
meanings. But when I put one word in front of another in front of another purposefully, and I 00:56:52.400 |
create something that has value to me, I should have the right to protect that. And I think that's 00:56:57.760 |
the fundamental question. Should you not have a right to protect it? Should you allow somebody 00:57:02.880 |
else to take that? And in some opaque way, and undefinable way, make a system that could compete 00:57:09.600 |
with that thing and have no repercussions. I think that that's what people need to decide. 00:57:14.480 |
And I'm fine, by the way, one way or the other. Yeah, in the Ed Sheeran case, he was saying the 00:57:19.600 |
exact copyrighted lyrics. And he was singing the exact copyrighted melody. He was just playing the 00:57:25.040 |
harmony with four chords. I mean, that was kind of the point he was making. And in the case of 00:57:30.400 |
all the copyrighted material, you're stating, I want to protect it from showing up in the end 00:57:35.360 |
product of someone's creative process. I shouldn't care about them listening to my song in their 00:57:40.000 |
creative process. Everyone always references how Oasis sounds like the Beatles, right? Those guys 00:57:44.800 |
love the Beatles, they listen to the Beatles, so much of their music mimics and sounds like the 00:57:48.960 |
Beatles, but the Beatles aren't going to file infringement suits against them. They were 00:57:51.680 |
inspired by the Beatles, that was part of their creative process. And so I would argue that the 00:57:56.240 |
output should really be where we assess this stuff. And it's going to take a lot of cases, 00:58:00.240 |
you're creating an impossible standard where no any individual person will have the economic 00:58:06.000 |
wherewithal to fight this. And so it's basically seeding all creativity to the board. And I don't 00:58:12.960 |
get what is what is the standard of measuring how someone is training their creative process? 00:58:18.160 |
Your standard is I got to go I don't think I got to go listen, I got to go you got to tell me every 00:58:22.240 |
time you as an artist, listen to a song, and then I'm going to determine whether or not you are 00:58:28.320 |
listening to copyright music and making your own music. That's basically what we're asking for here 00:58:32.640 |
with this. That's not what I don't I'm not talking about the bill. I'm talking about this idea that 00:58:36.880 |
if you train a system that can actually automate a task for you, based on some learning that's 00:58:42.800 |
unique, you should be able to uniquely protect your ownership. For example, we have a different 00:58:49.120 |
set of ways in order to protect processes that are unique to you, which is, you can have a patent, 00:58:53.840 |
or you can have a trade secret. But if you strip all of this stuff away, where some system can in 00:58:59.440 |
an automated way do anything for you, and replicate a whole bunch of work that it sits on top of to be 00:59:06.000 |
able to do that without any repercussions. I think that that's a very important explicit decision. 00:59:11.040 |
And what I'm saying is, you cannot expect individual people to go and fight these little 00:59:15.680 |
battles against these big multi hundred billion dollar companies. If the big multi hundred billion 00:59:21.040 |
dollar companies and trillion dollar companies don't fight the other multi hundred billion dollar 00:59:25.280 |
and trillion dollar companies, what they're effectively saying is none of this is worth 00:59:29.120 |
anything. Copyrights are worth nothing. And I'm okay with that decision. But it should be an 00:59:33.680 |
explicit one. And it's not going to happen the other way where some five page law from a guy 00:59:40.240 |
who's trying to get elected is not the thing that's going to make the difference. The difference is, 00:59:44.400 |
if Google doesn't sue open AI, it means copyright is worthless. And by the way, that article didn't 00:59:49.840 |
just point the finger at Google. It pointed the finger at meta where there were explicit decisions 00:59:54.320 |
to say, Yeah, no, let's just violate the copyright. It's, it's okay. And if it's okay, then it's okay. 00:59:59.120 |
Saks, you're the independent arbiter here. Any summary? 01:00:05.200 |
Well, somebody's gonna litigate it. Somebody's gonna get gonna litigate it. You're right, 01:00:09.360 |
it'd be surprising if Google didn't want to, given how valuable their YouTube content is, 01:00:14.320 |
and a competitor, you know, slurped it up to train their model. But somebody will will litigate this, 01:00:20.240 |
it could be the songwriters, who knows. And then we're gonna we're gonna get some messy arbitration 01:00:25.760 |
around fair use, and it's probably gonna work its way up over the different circuit courts, 01:00:29.760 |
you'll probably get different judgments. And finally, the Supreme Court will resolve it. 01:00:34.000 |
And then we'll kind of know where things stand. And then and then there'll be a legislative fix. 01:00:39.120 |
I don't know where it's going to end up. I mean, I know what Jay cow would say if he were here, 01:00:42.880 |
I'm sure he's like chomping at the bit be part of this conversation. He would say that creators 01:00:47.520 |
deserve to get paid for their AI rights. And that if you want to make money training a model 01:00:53.440 |
on their content, you got to compensate them. I also I buy I buy your argument too, which is 01:00:59.440 |
you're not really violating their their rights for the model to 01:01:06.880 |
to read their content and analyze it, which is what it's really doing as long as it doesn't 01:01:12.800 |
output a violation of their copyright. And that's what all artists do. That's what all humans do. 01:01:19.520 |
The creative process is one of initially synthesis. You learn from other people's works, 01:01:25.680 |
you learn from the environment around you, you learn from the world that you're exposed, 01:01:28.480 |
there was compensation in those other models. There was not you don't of course, 01:01:32.800 |
Oasis never paid the Beatles to listen to their music. They 01:01:36.480 |
know that's not true. They listened on the radio, they may not have paid, 01:01:40.800 |
but some advertiser paid for an ad. The radio station had to pay ASCAP and BMI, 01:01:45.920 |
there was publishing rights, there was all kinds of other rights. So that's not true what you're 01:01:49.840 |
saying. I'm not what I'm what I'm saying is it's not copyright just because somebody got paid. So 01:01:54.800 |
that singer songwriter did get paid, they may not have getting paid by the listener. But in this 01:01:59.440 |
example, nobody's getting paid. Look, I think there's three potential outcomes here. Okay, 01:02:03.920 |
outcome number one is that Freiburg wins. And the courts decide that this is fair use. And 01:02:10.720 |
these AI companies aren't in violation of copyright infringement is measured in the work of art that 01:02:15.760 |
is produced, not in the process by which it is generated. Right? Okay, so that's solution number 01:02:20.960 |
one. I think solution or outcome number two is that the courts throw a big wrench into it or a 01:02:28.080 |
bill like schist bill, which is largely punitive, throws a big wrench into it. And I think what 01:02:33.280 |
happens then is new AI companies have a really hard time accumulating training data. And so the 01:02:39.360 |
incumbents who already have all the training data, they may have to pay a fine, there'll be some slap 01:02:44.080 |
on the wrist. But now you've created a real moat and barrier to entry. And so secretly, they're 01:02:48.640 |
going to love it. And it's going to be a big problem, I think, for the open source movement. 01:02:52.400 |
So that's outcome number two. And then outcome number three is you create some sort of rights 01:02:58.800 |
clearinghouse, kind of like exists in the recording industry, where if you want to cover somebody's 01:03:05.440 |
song or perform somebody else's song, there's already predefined terms on which you can use 01:03:10.720 |
those rights. And then there's a compensation mechanism for that. So you don't have to go 01:03:15.360 |
negotiate every single deal that you want to do, you can just use the song, you can use the rights, 01:03:20.560 |
and there's again, compensation for that. That would be like the compromise, I could see that 01:03:25.360 |
alternative number three existing, in which case, the incumbents probably like that, because 01:03:31.520 |
they again, they want there to be a barrier to entry. They want to make it hard for the guy in 01:03:36.480 |
the garage, to pony up the capital necessary to create a new model. But at least that would be a 01:03:43.120 |
somewhat functional system, it wouldn't be outcome number two, which would totally put the kibosh on 01:03:48.400 |
the use of training data. So in any event, I don't know which one of the three it's going to be. But 01:03:52.240 |
those are the three potential outcomes I see happening here. Have you guys seen you do you dio? 01:03:56.960 |
No, it's like a new music gen AI system. It is friggin unbelievable. 01:04:07.520 |
Is that the notes thing or the notes another thing? 01:04:09.200 |
I think the notes another thing this one just came out, it was apparently rumored for a while 01:04:14.480 |
ex DeepMind employees started you do it's been this kind of like hot stealth company. 01:04:20.160 |
For a couple of months, people have been talking about it as rumored to be the best music AI. 01:04:26.880 |
But here is someone that said, Dune the Broadway musical, and this is what it output. 01:04:48.800 |
You can hear the layers of music, the layers of the tracks, 01:04:51.840 |
the smoothness, the tempo, obviously, like there's just so many layers to this. 01:04:57.600 |
Nick, can you pull up another one? Some of these are just incredible, 01:05:00.320 |
because it spans genres. It spans styles. It's like so beautiful, like the way it's put together 01:05:07.040 |
every every piece feels somewhat emotional. It's unique. It really doesn't feel like you're just 01:05:12.640 |
listening to someone printing copyright music. funky dance pop song. Create a birthday song 01:05:24.640 |
for my friend urban. Create an Irish folk tune with violin and instrumental. 01:05:34.160 |
It's pretty incredible. And I think it shows the state of the art is such now that this 01:05:49.840 |
is going to become a real challenging question from a legislative point of view. 01:05:54.240 |
Given how far ahead these technologies have gotten, and I think that musicians, 01:05:58.480 |
artists, consumers are going to start to use these tools in a really 01:06:01.520 |
prolific way, given how good they are now. I heard from someone that just sat for hours, 01:06:07.440 |
creating tracks that they listened to on UDO the other day. And so you basically just playing your 01:06:12.720 |
own music all day long. It's really incredible. Anyway, exciting time. 01:06:16.640 |
Hopefully all the lawsuits and the legislators don't get in the way of all this innovation, 01:06:20.560 |
because it's moving really fast. And we should just kind of see where it goes before trying to 01:06:24.480 |
clamp down. Totally. Guys, I want to talk about, as you know, this like emerging trend that we've 01:06:31.440 |
been talking about offline about drones in warfare. Recently, we saw small drones controlled 01:06:36.800 |
by the Houthis attack cargo vessels in the Suez Canal. It led to rather large scale supply 01:06:42.320 |
disruptions, economic value destruction. We've seen videos, sex, I think you have one, we can 01:06:47.840 |
pull it up and take a look of Ukrainian controlled drones flying into destroy tanks and large 01:06:53.760 |
warehouses in Russian held territories, causing 10s of millions of dollars of damage. I think 01:06:58.800 |
we're seeing the acceleration of the changing face of warfare technology, lots of these small 01:07:03.040 |
form factor, low cost autonomous physical agents. And I think there's a bunch of implications both 01:07:09.280 |
with respect to war, but also how we spend money on defense, and how Silicon Valley is involved 01:07:15.520 |
in response. tax, I'd love to hear your points of view. Maybe you can highlight for us some of the 01:07:20.080 |
things you've been seeing in the utilization of these new technologies and in the war, given our 01:07:25.200 |
your our in house work correspondent, and you obviously do some work in Silicon Valley. Go 01:07:30.480 |
ahead. I've heard a lot about the impact of drones just by watching the coverage of the Ukraine war. 01:07:36.320 |
I was listening to an interview with an American mercenary who's fighting on the side of Ukraine, 01:07:42.320 |
and he described how ubiquitous these drones were on the battlefield. Now he said that you literally 01:07:47.680 |
can't get out of a trench to go to the bathroom because a drone will basically get you. And he 01:07:53.840 |
said they're buzzing around, they sound like mosquitoes, because they're kind of just everywhere 01:07:57.840 |
on the battlefield. Both Ukraine and Russia have them. And several months ago going into this year, 01:08:05.360 |
as long as he said that this would be the big game changer for Ukraine, they're going to make 01:08:08.400 |
a million drones. I don't think it's worked out that way. It turns out that as many drones as 01:08:12.880 |
Ukrainians have been able to put on the battlefield, the Russians have even more, 01:08:16.240 |
because they're able to mass produce them in factories, they've got bigger drones, better 01:08:21.280 |
drones. And I think the most important variable in the war so far, with respect to drones is that 01:08:26.800 |
the Russians have pretty advanced electronic warfare. And so they've been able to jam a lot 01:08:32.160 |
of the Ukrainian drones, whereas the reverse has not been true for the Ukrainians. So I would say 01:08:38.560 |
that drones have been a huge factor in the war. But so far, the balance there is tipping towards 01:08:45.040 |
the Russians as it is in so many other areas as well. But to your larger point, there's no 01:08:50.560 |
question this is the future of warfare. And you're seeing that it's creating a lot of opportunities 01:08:57.440 |
for asymmetric warfare. So for example, with the hoodies, they've been firing cheap missiles and 01:09:04.560 |
drones at our at our ships in the Red Sea, and we've been having to spend $2 million air defense 01:09:11.040 |
missiles shooting down $2,000 drones. So if that continues, and we don't have a good response to 01:09:16.000 |
this problem, it's going to really change the balance of power. I don't know if this was at 01:09:20.240 |
our summit, or if I heard this later from some senior person in the military, who said that 01:09:24.640 |
aircraft carriers are outdated technology, like they don't make sense anymore. And we're going 01:09:29.840 |
to see a shift towards lots of small, autonomous drone like systems on the battlefield taking out 01:09:38.080 |
targets. And as a result, if you gain theory this out, the necessity for defense systems against 01:09:44.720 |
large amounts of swarming drone like systems. And then, you know, what do you actually have to do 01:09:51.440 |
with your existing military architecture to play into this kind of new tactical system? 01:09:57.840 |
And then that's going to require a massive evolution in technology. And is the United 01:10:04.400 |
States prepared? Should Silicon Valley play should be, we should I maybe Nick, you can throw up this 01:10:11.840 |
the first link that I sent you. But eight years ago, I came to this conclusion. And I invested 01:10:17.600 |
in this company called sail drone, which is autonomous drones in the seas, our customers 01:10:22.720 |
include the US Navy, the chairman of our company actually is Admiral Mike Mullen, who is a former 01:10:27.280 |
chairman of the Joint Chiefs of Staff. But these drones are capable of going into some of the most 01:10:32.800 |
dangerous places in the world, and collecting enormous amounts of surveillance data and 01:10:37.680 |
information that otherwise takes the United States government lots of time and lots of money, 01:10:46.240 |
and is pretty scattershot. And instead with these things, you can be constantly in hotspots. And if 01:10:53.840 |
you go to the next one, so much so that you know, we've deployed some of these drones in the Middle 01:10:58.320 |
East, and there was a point at which the Iranian Navy intercepted two of our drones and pick them 01:11:03.440 |
up. And this was kind of like a global thing a few years ago. But I really believe in this trend. 01:11:09.120 |
And I think that we have an enormous responsibility to be funding these things. These are really 01:11:14.800 |
complicated systems to build, obviously, and they take lots of time. So these are not 01:11:18.320 |
overnight successes, Friedberg, and they take lots of money, which is hard to come by as well. 01:11:22.880 |
But these are absolutely the right kinds of businesses, because eventually, 01:11:26.480 |
at a minimum, you're building systems that can measure and collect enormous amounts of really 01:11:31.600 |
critical data, so that people can make better decisions, which hopefully is measured in saving 01:11:38.000 |
life, right. And then over time, you actually get to a military that should be much cheaper, 01:11:44.720 |
much safer, and has fewer people in general on our side and on our enemy sides on the front lines, 01:11:52.080 |
which means fewer casualties. And I think that that's the whole value of this entire movement. 01:11:57.040 |
It's cheaper, it's way better. It's much more useful. And it gives the respective armies that 01:12:04.480 |
use these things or militaries rather, a fuller picture so that you can make decisions that have 01:12:08.880 |
enormous consequences. Yeah. And in February, the Department of Defense on a panel in Arlington, 01:12:16.080 |
Virginia, briefed reporters and industry folks on their interest in growing their partnerships in 01:12:22.640 |
Silicon Valley, to access the necessary technologies that are going to rewrite the 01:12:29.360 |
face of warfare. I don't know about you guys, but I've seen kind of two sides of this in Silicon 01:12:35.120 |
Valley. A good number of venture capitalists and entrepreneurs and technologists think that it's 01:12:41.840 |
morally incorrect to support warfare technology. And then there are a few that are going blazing 01:12:48.480 |
ahead with supporting this new evolution and these technologies. Do you guys see the same? 01:12:53.360 |
And is there going to be kind of this? It's a big step to weaponize these things, 01:12:57.040 |
right? So Saildrone, we have a very good, big, thriving, successful business. These 01:13:03.440 |
machines aren't weaponized by any stretch of the imagination. And so I think that that 01:13:08.400 |
comment is more genuflecting and virtue signaling than it is real. Because in order to even be 01:13:17.840 |
legitimately considered for some kind of like armed, unmanned product, you have to be deep 01:13:26.480 |
inside of the bowels of the DoD and Pentagon and have been working with them for years, 01:13:31.280 |
don't even be taken seriously. So the startups that are like eschewing those deals are not even 01:13:36.400 |
close to those deals. And the startups that are close to them are probably being much more quiet 01:13:42.400 |
and thoughtful and hush hush, because the path to do that isn't legitimate unless you've been 01:13:47.120 |
in business with these folks for 10 plus years. Yeah, I think it's, it's more commentary on the 01:13:53.440 |
fact that if this is where an organization, the Department of Defense that has a trillion dollar 01:13:59.920 |
annual budget is going to be allocating resources, that technology is going to need to come from 01:14:06.960 |
somewhere. And the investors in Silicon Valley, that are making these investments are likely going 01:14:12.720 |
to outperform those who are not. Does that sound reasonable to you, Sax, that there is this shift 01:14:18.000 |
underway? And that, again, those who have shied away from defense technology are, you know, 01:14:24.160 |
necessarily going to be left out of a, you know, kind of a new industry that's emerging in Silicon 01:14:28.240 |
Valley? Well, I guess the way I put it is that the US government spends over $800 billion every 01:14:34.240 |
year on defense. And it's not clear what we're getting for that money. Because in Ukraine, 01:14:39.280 |
for example, we've run out of artillery shells, we've actually we've run out of patriots, we've 01:14:42.800 |
run out of javelins, we're on a stinger. So we're spending all this money, we're not getting a lot 01:14:47.040 |
for it. And part of the reason is because the defense industry is consolidated down to these 01:14:51.920 |
five prime defense contractors who are basically an oligopoly. And we have this cost plus procurement 01:14:59.040 |
system where they just raise their prices every year and the government pays it. So I think all 01:15:05.040 |
of us want the United States to have an effective defense. I mean, I want us to use our military 01:15:09.280 |
power more wisely. I don't like all these stupid wars we keep getting in. But I do want the United 01:15:13.360 |
States as an American to be the most powerful country. I do want us to get the best value for 01:15:20.720 |
our defense dollars. And the only way that's going to change is if the defense industry gets disrupted 01:15:27.680 |
by a bunch of startups doing innovative things. And there's no question that I think drones are 01:15:33.280 |
the future of warfare. To your point about autonomy, I think that is where this is going next 01:15:38.160 |
is that right now, the drones are typically controlled by, you know, somebody with like a VR 01:15:45.520 |
headset, tele operated, operated and their FPV, this first person vision drone, and you basically 01:15:53.200 |
strap some sort of explosive on to it, and then you drive it into whatever your target is. Yeah. 01:15:58.960 |
Those types of drones are easier to disrupt by electronic warfare, because if you can disrupt 01:16:05.680 |
the signal from the tele operator, then the drone basically doesn't know what to do. And so you're 01:16:12.160 |
right, the next step here is autonomous systems that can be programmed with a target and can find 01:16:20.960 |
it on its own, make decisions on its own. And then they also build in some shielding against 01:16:25.200 |
some of this like jamming technology or this EW electronic warfare technology. 01:16:31.040 |
So that's where all these things, that's where the battlefield is of the future is headed. 01:16:35.120 |
And, you know, in a weird way, if you think about humans becoming a smaller and smaller 01:16:41.040 |
piece of the battlefield, and autonomous drones becoming a larger and larger piece of it, 01:16:46.160 |
these wars become resource wars, and that's right, actually, technology wars. 01:16:51.200 |
That's right, which which may or may not be good. I don't know. 01:16:53.600 |
But by the way, so this is the point I wanted to make, if you pull up this chart, 01:16:56.800 |
if that is where warfare is headed, is a large number of small autonomous systems 01:17:03.440 |
that go and find a target and try and execute a mission. And, you know, in the case of Ukraine, 01:17:08.800 |
saying they want to have a million, China might say, I want to have 100 million, all of these 01:17:13.840 |
systems are dependent on lithium ion battery systems. Today, 79% of lithium ion battery 01:17:18.640 |
production comes out of China. The US is only 6.2% of global lithium ion battery production. 01:17:24.240 |
And China has talked about scaling up drone manufacturing to a level that the US simply 01:17:31.280 |
cannot even contemplate in its industrial architecture today. So it seems to me that 01:17:38.160 |
if that is where warfare tactically is headed, that China has a huge leg up, and is going to 01:17:44.400 |
become a critical point of dependency for the United States to develop an arsenal necessary 01:17:50.720 |
to be competitive in this this kind of next evolution of warfare. I will say that, like, 01:17:56.000 |
if you then game theory this out, like how do you defend against these autonomous electronic systems, 01:18:00.800 |
there's a technology called EMP or electromagnetic pulsing, where, as you guys know, if you run a 01:18:07.440 |
very high current electric field, you can actually send out and emit a pulse that then when it hits 01:18:13.280 |
electronic circuits far away, induces a high electric current in those circuits and short 01:18:18.560 |
circuits them. So EMPs are a defense system that allow you to take out electronic systems. And this 01:18:26.560 |
has been, you know, kind of a part of warfare, since probably the 1960s 50s, but targeted EMPs 01:18:32.800 |
and targeted systems for eliminating all of these autonomous systems becomes a new defense technology 01:18:37.520 |
that I know several startups in Silicon Valley that have been funded with a lot of capital 01:18:43.920 |
by folks that we all know very well, that are kind of pursuing this, 01:18:48.160 |
we don't have a choice. Because I think the point is that if you just make me just post this photo, 01:18:53.360 |
we have an enormous human capital problem with the military, which is there's just not enough 01:18:57.520 |
folks enlisting anymore. So we don't have any choice except to automate and become 01:19:02.720 |
drone dependent. Just tell us these numbers. This is a chart from the Department of Defense 01:19:07.440 |
that shows military enlistments looking back from about the 1970s, through today. And these are two 01:19:15.040 |
lines that are just going from the upper left to the lower right. And so we're at all time lows 01:19:20.160 |
with respect to the number of people that actually want to join the military. And so if we are 01:19:25.280 |
supposed to, as Zach said, be this very sharp fighting force, then all of the money that we're 01:19:32.240 |
spending needs to get allocated into things that can be remotely operated. We don't have a choice. 01:19:38.400 |
Yeah. I mean, so Freeberg, you asked the question, how do we defend against these drones? 01:19:42.560 |
Yes, there's EMP technology. If you want to use an EMP, you better make sure there's not any planes 01:19:48.480 |
nearby, because it will take them out of the sky, you better make sure. 01:19:51.120 |
By the way, there are now, you can target EMPs in a narrow cone, which is part of the 01:19:56.640 |
Yeah, you better make sure you don't hit any of your own stuff, because you'll fire the electronics 01:19:59.920 |
on those unless you EMP harden them, which by the way, will happen. I mean, if the US gets really 01:20:04.960 |
good at sending out EMP pulses, then our opponents will start EMP hardening their drones. But anyway, 01:20:12.640 |
yes, it's a category of defense. So is electronic warfare. The one defense investment I've made is 01:20:18.720 |
actually in this idea of how do you defend against drones. And it's a startup where I 01:20:23.920 |
led the seed round called Allen Control Systems, and they have a product called Bullfrog, 01:20:27.600 |
which is a gun turret that uses a standard M240 machine gun. But it's got a scanner next to it, 01:20:34.800 |
next to the machine gun that is searching the sky for enemy drones. And it uses computer vision to 01:20:42.560 |
recognize them, and then it just shoots them down like very quickly. And it seems to me this is, 01:20:47.680 |
it's maybe not the only way of doing it, but it's a really good way of doing it because 01:20:51.360 |
you can mount these things to a vehicle. I mean, imagine in the future, 01:20:56.960 |
that we have these autonomous drones everywhere that are basically assassination drones. 01:21:03.600 |
How are you going to stop them? I mean, how do you move the President of the United States around? 01:21:06.640 |
I mean, in his caravan, you're going to have to have, you know, his caravan is going to have to 01:21:11.520 |
have some sort of electronic warfare system on it to basically prevent drones, but it's gonna need 01:21:19.440 |
some sort of last line of defense where you can actually shoot a drone out of the sky. 01:21:25.120 |
So Sax, you don't have any moral trepidation around investing in defense technology, 01:21:30.720 |
Well, you know, I'd have to think twice about investing in a weapon. This is more of a defense 01:21:35.360 |
technology. This is defensive in nature where you're trying to take out a drone that would 01:21:39.760 |
be attacking one of our bases or installations or convoys. So this is fundamentally defensive 01:21:46.240 |
in nature. And look, I want America to be the most powerful country. I'm an American, I'm patriotic. 01:21:52.800 |
So I don't want our troops to be hit by enemy drones. So no, I don't have any moral problem 01:21:59.040 |
with this. I do have a problem with all the stupid wars the United States gets into, 01:22:04.080 |
but that's on a policy level. I feel like that's separate from the question of do we want our guys 01:22:11.440 |
And Chamath, you don't care, you would invest in defense technology companies. 01:22:15.040 |
No, I care. I think that I would probably have to think twice about making weapons. I'm not sure 01:22:19.360 |
that I'm really interested in that. But like I said, eight years ago, my thought was we need to 01:22:23.600 |
move to just looking at those trends, we need to move to an entire surface area of unmanned 01:22:29.040 |
autonomous vehicles. And I really liked this idea of a class of machine that's there to surveil and 01:22:37.600 |
collect all the important data so that you can make better decisions. And 01:22:40.640 |
it's taken us eight years, but we're firmly entrenched now. 01:22:44.400 |
I think the fact that both of you highlight a degree of trepidation, and you guys are like 01:22:50.160 |
fairly outside of the middle bounds of how Silicon Valley would think and make decisions indicates 01:22:56.320 |
like why so much of Silicon Valley has deep concerns about investing in defense technology. 01:23:00.480 |
And the US, despite having the best technology, may be challenged by the fact that our Silicon 01:23:07.360 |
Valley technologists and investors are held back on pursuing these evolutions because 01:23:17.200 |
of a moral suasion. And I think it is something that the Department of Defense has highlighted. 01:23:22.320 |
No, I want to be clear though, it's never like we've been faced with that choice. And the reality 01:23:28.000 |
is like, if the choice is stay small, or scale up and be really big, you know, it's hard to 01:23:36.800 |
turn away from that decision. But that choice isn't a realistic one. This is why I was saying, 01:23:41.840 |
I think it's, you can come into it with the moral framing. But I do think that Sachs is right, 01:23:46.320 |
which is, if you believe that America should be the most important country in the world, 01:23:51.040 |
and you have the ability to help it be that, eventually, some of those decisions will 01:23:56.880 |
come in contact with this kind of decision. And usually, though, it takes many, many years before 01:24:04.320 |
that's realistic. And to be honest, from what I've seen, the opportunity to pursue that is 01:24:13.280 |
generally so financially remunerative that everybody's like, Okay, I think we have to do 01:24:17.520 |
you know, I don't have a problem with investing in a company like and roll. I mean, America needs 01:24:22.160 |
defense companies in order to maintain its defense. And the problem right now that we have 01:24:26.640 |
is there's these five prime companies that are antiquated and really expensive. And we need 01:24:33.200 |
startup disruptors to that. So I'm not like, morally opposed to it. And in fact, I think it's 01:24:37.680 |
probably a good thing. I think my concern would be, I don't really want to be part of the military 01:24:41.920 |
industrial complex. I feel like it might, like, corrupt my views on things like I don't want to 01:24:48.000 |
know you want war, you'd be girding for Yeah, like, I don't want to be I don't want to feel 01:24:52.400 |
that's your dimension. That's Yeah, right. I don't want to feel an incentive to be like these 01:24:57.200 |
generals who go on CNN and Fox News and justify every single war. That's my biggest fear as well. 01:25:02.960 |
It's like, it's one thing to be investing in a bunch of companies where we're sitting around 01:25:07.520 |
hoping for cheaper clicks on Facebook and Google. It's an entirely other thing to sit on the board 01:25:12.160 |
of a company and hope for war. That's a horrible place to be in. On that point. That's a great 01:25:16.560 |
point to wrap on, guys. This has been fun. A little more flat than it would normally be with 01:25:21.440 |
Jay cow. We miss him. We wish him well, we hope his tooth recovers. And we will be back with your 01:25:28.640 |
it's like that scene in Dune where he eats his tooth and releases that poison gas. 01:25:32.880 |
That's right. Get better soon. J. cow. Please, if you're interested in the summit, 01:25:39.520 |
summit that all in podcast.co. Click Subscribe on YouTube. I've been given notes to say that 01:25:45.440 |
for the chairman dictator Chamath Palihapitiya, the rain man David Sachs, I am your Sultan of 01:25:51.440 |
science day Friedberg. We will see you guys next week. We open source it to the fans and 01:26:09.040 |
they've just gone crazy. We should all just get a room and just have one big huge orgy 01:26:32.720 |
because they're all just useless. It's like this like sexual tension, but they just need 01:26:36.000 |
to release it. I'm going all in. I'm going all in.