back to indexGeorge Hotz: Tiny Corp, Twitter, AI Safety, Self-Driving, GPT, AGI & God | Lex Fridman Podcast #387
Chapters
0:0 Introduction
1:39 Time is an illusion
11:18 Memes
13:55 Eliezer Yudkowsky
26:19 Virtual reality
32:38 AI friends
40:3 tiny corp
53:24 NVIDIA vs AMD
56:21 tinybox
68:30 Self-driving
83:9 Programming
91:6 AI safety
116:3 Working at Twitter
153:46 Prompt engineering
159:42 Video games
175:57 Andrej Karpathy
186:2 Meaning of life
00:00:00.000 |
What possible ideas do you have for how a human species ends? 00:00:03.840 |
- Sure, so I think the most obvious way to me 00:00:10.020 |
We end up all staring at that infinite TikTok 00:00:21.340 |
Now, to be fair, it's probably hard to get all of humanity. 00:00:42.160 |
- The following is a conversation with George Hotz. 00:00:53.500 |
and is the founder of a new company called TinyCorp 00:00:57.800 |
that created TinyGrad, a neural network framework 00:01:09.840 |
As you know, George also did a large number of fun 00:01:20.440 |
making the case for refactoring the Twitter code base. 00:01:23.680 |
In general, he's a fascinating engineer and human being 00:01:47.660 |
- You know, I sell phone calls to Kama for $1,000 00:01:58.600 |
you know, it's $1,000, you can talk to me for half an hour. 00:02:01.240 |
And he's like, "Yeah, okay, so like time doesn't exist 00:02:05.920 |
"and I really wanted to share this with you." 00:02:08.600 |
I'm like, "Oh, what do you mean time doesn't exist, right? 00:02:18.120 |
"It's about whether it's a useful model to describe reality. 00:02:29.020 |
Like underneath it all, is there an actual thing 00:02:42.080 |
- I mean, this kind of connects to the models 00:02:44.440 |
of constructive reality with machine learning, right? 00:02:48.600 |
- Like, is it just nice to have useful approximations 00:02:52.360 |
of the world such that we can do something with it? 00:03:14.280 |
- I don't think you're the person that follows the majority 00:03:23.960 |
All right, but to you, time is a useful model. 00:03:28.880 |
- What were you talking about on the stream with time? 00:03:33.320 |
- I remembered half the things I said on stream. 00:03:36.320 |
Someday someone's gonna make a model of all of that 00:03:47.360 |
- I mean, the question is when the George Hotz model 00:03:52.400 |
Like I am declining and the model is growing. 00:03:54.920 |
- What is the metric by which you measure better or worse 00:04:11.700 |
but both will be overtaken by the George Hotz model. 00:04:18.120 |
Loved ones, family members would rather have the model 00:04:24.800 |
- Or like significant others would rather sext 00:04:28.600 |
with the large language model version of you. 00:04:34.680 |
- Especially when it's fine tuned to their preferences. 00:04:39.360 |
Yeah, well, that's what we're doing in a relationship, 00:04:47.280 |
Our language models can fine tune more efficiently, 00:04:53.440 |
where Catherine Janeway lost in the Delta Quadrant 00:05:05.740 |
and Janeway edits the program to remove that. 00:05:16.300 |
and slight annoyances that make this relationship worthwhile. 00:05:30.640 |
- Just the perfect amount of quirks and flaws 00:05:33.760 |
to make you charming without crossing the line. 00:05:41.520 |
of the percent of time the language model should be cranky 00:05:46.520 |
or an asshole or jealous or all this kind of stuff. 00:05:53.360 |
but all that difficulty at that point is artificial. 00:05:58.760 |
- Okay, what's the difference between real and artificial? 00:06:03.200 |
that's like constructed or could be turned off with a knob. 00:06:09.500 |
- So if something can not be turned off with a knob, 00:06:24.360 |
Into the wild when, you know, Alexander Supertramp, 00:06:31.840 |
but it's the '90s, everything's been explored. 00:06:33.560 |
So he's like, well, I'm just not gonna bring a map. 00:06:40.000 |
You should have brought a map, dude, you died. 00:06:41.480 |
There was a bridge a mile from where you were camping. 00:06:44.080 |
- How does that connect to the metaphor of the knob? 00:06:46.600 |
- By not bringing the map, you didn't become an explorer. 00:06:58.460 |
What if we just don't have access to the knob? 00:07:05.520 |
and nature has been fine-tuned over billions of years 00:07:17.560 |
in some grand romantic gesture is horrifying. 00:07:24.800 |
or is it, are we just all part of one living organism 00:07:37.680 |
I don't know, I mean, I don't think I'm conscious. 00:07:49.240 |
- Everything running in the universe is computation, I think. 00:07:51.680 |
I believe the extended church-turning thesis. 00:07:57.320 |
to your particular computation, like there's a consistency. 00:08:00.920 |
- Well, yeah, but I mean, models have consistency too. 00:08:23.920 |
and thereby fine-tune these little pockets of computation, 00:08:28.840 |
but it's still unclear why that pocket of computation 00:08:35.120 |
Like you have this consistent set of physics, biology, 00:08:40.120 |
what, like whatever you call the neurons firing, 00:08:46.680 |
like the electrical signals and mechanical signals, 00:08:50.320 |
and it contains information, it stores information, 00:09:08.060 |
- Reinforcement learning with human feedback. 00:09:10.300 |
- You know, when I talked about will GPT-12 be AGI, 00:09:15.720 |
I mean, cross-entropy loss is never gonna get you there. 00:09:23.320 |
in order to get something that would be considered like, 00:09:30.140 |
I don't know, like it's just some quirk of evolution, 00:09:32.560 |
right, I don't think there's anything particularly special 00:09:35.500 |
about where I ended up, where humans ended up. 00:09:43.820 |
Would you call that AGI, whatever we have, GI? 00:09:47.580 |
- Look, actually, I don't really even like the word AGI, 00:10:02.180 |
- If your loss function is categorical cross-entropy, 00:10:04.380 |
if your loss function is just try to maximize compression, 00:10:09.960 |
and I tried to get ChatGPT to help me write raps, 00:10:18.260 |
and you can see what people put in the comments, 00:10:20.300 |
and it's the most like mid-quality rap you can find. 00:10:27.780 |
- Every time I talk to you, I learn new words. 00:10:37.260 |
- Kind of, it's like middle of the curve, right? 00:10:42.880 |
and you have like the dumb guy, the smart guy, 00:10:48.260 |
The smart guy's like, I put all my money in Bitcoin. 00:10:50.260 |
The mid guy's like, you can't put money in Bitcoin, 00:11:14.740 |
so I'm very, I sound like I'm high, but I swear I'm not. 00:11:22.360 |
- I think that we're gonna get super scary memes 00:11:35.980 |
So "Infinite Jest," it's introduced in the first 50 pages, 00:11:40.980 |
is about a tape that you, once you watch it once, 00:12:08.940 |
But we're going to build that using generative models. 00:12:19.460 |
The algorithm is just doing their recommendation, 00:12:21.380 |
but if the algorithm is also able to do the generation. 00:12:25.340 |
- Well, it's a question about how much intelligence 00:12:30.700 |
let's say, one humanity worth of intelligence, 00:12:34.780 |
That's a, you know, it's X-flops, Yada-flops, 00:12:41.860 |
Once that generation is being done by 100 humanities, 00:12:58.660 |
the very limited human dopamine engine for porn? 00:13:08.540 |
- I don't even know what it'll look like, right? 00:13:16.660 |
and an agent that just dominates your intelligence so much 00:13:24.060 |
- Is it possible that it won't really manipulate, 00:13:28.660 |
It'll just kind of exist the way water exists, 00:13:33.460 |
- You see, and that's the whole AI safety thing. 00:13:44.340 |
'Cause the machine is not interested in hurting humans. 00:14:02.060 |
He thinks that AI will almost surely kill everyone. 00:14:22.620 |
- Okay, why didn't nuclear weapons kill everyone? 00:14:31.180 |
It's very hard to accomplish tactical objectives. 00:14:40.900 |
- Why don't I want an irradiated pile of rubble? 00:14:46.300 |
- Oh, 'cause you can't use that land for resources. 00:14:52.140 |
- Yeah, what you want, a total victory in a war 00:15:00.020 |
It's the subjugation and domination of the people. 00:15:06.740 |
tactically in a war to help gain a military advantage. 00:15:19.140 |
- Still surprising that nobody pressed the big red button. 00:15:25.980 |
that's gonna be pressed with AI that's gonna, 00:15:37.660 |
- What's the algorithm behind the little red button? 00:15:45.460 |
- Sure, so I think the most obvious way to me 00:15:53.180 |
We end up all staring at that infinite TikTok 00:16:04.600 |
Now, to be fair, it's probably hard to get all of humanity. 00:16:29.740 |
No, like I'm kinda, look, I'm drinking Smart Water, man. 00:16:40.180 |
just like all the other biodiversity on the planet. 00:16:48.100 |
- Yeah, no, it's the interconnectedness that's doing it. 00:16:54.980 |
So everybody starts relying on the connectivity 00:17:00.980 |
that reduces the diversity, the intellectual diversity, 00:17:03.780 |
and then that gets you, everybody, into a funnel. 00:17:13.820 |
I think AI kills everything we call society today. 00:17:17.460 |
I do not think it actually kills the human species. 00:17:19.580 |
I think that's actually incredibly hard to do. 00:17:21.880 |
- Yeah, but society, if we start over, that's tricky. 00:17:28.780 |
- Yeah, but some of us do, and they'll be okay, 00:17:40.660 |
Like, what has human civilization done that's interesting? 00:18:05.020 |
some kind of Amish-looking kind of thing, I think. 00:18:10.900 |
- Like, technology is almost like a new religion. 00:18:26.500 |
Isn't it somehow have a hold, like a stronghold? 00:18:30.500 |
- What's interesting about everything we build, 00:18:33.740 |
I think we are going to build superintelligence 00:18:35.980 |
before we build any sort of robustness in the AI. 00:18:41.860 |
of going out into nature and surviving like a bird, right? 00:18:51.780 |
We haven't built a machine that's capable of reproducing. 00:19:08.360 |
I guess you're saying they can't repair themselves. 00:19:12.960 |
- Let's just focus on them reproducing, right? 00:19:22.520 |
- Well, it doesn't have to be all on board, right? 00:19:29.640 |
- Yeah, but then you're really moving away from robustness. 00:19:51.180 |
'cause I think we're gonna get, again, super intelligence 00:20:09.680 |
to have a robot that basically can build itself? 00:20:16.640 |
- I think you've mentioned this to me or somewhere 00:20:24.320 |
- And then they remember that you're gonna have 00:20:51.080 |
- You're gonna have to have a very interesting kind of fab 00:20:54.640 |
if you wanna have a lot of computation on board. 00:20:59.040 |
But you can do like structural type of robots 00:21:29.800 |
Okay, so like there's two stacks of life in the world. 00:21:32.200 |
There's the biological stack and the silicon stack. 00:21:35.080 |
The biological stack starts with reproduction. 00:21:40.820 |
The first proto-RNA organisms were capable of reproducing. 00:21:45.520 |
The silicon stack, despite as far as it's come, 00:21:51.820 |
- Yeah, so the fab movement, digital fabrication, 00:21:56.820 |
fabrication in the full range of what that means 00:22:06.880 |
- Even if you did put a fab on the machine, right? 00:22:16.260 |
So first off, this machine is gonna be absolutely massive. 00:22:26.480 |
Like is our civilization capable of reproduction? 00:22:53.040 |
- I believe that Twitter can be run by 50 people. 00:23:04.200 |
- No, but you're not interested in running Twitter. 00:23:13.040 |
- Oh, okay, you're talking about, yeah, okay. 00:23:14.800 |
So you're talking about the humans reproducing 00:23:21.320 |
But they're not gonna be making five nanometer chips. 00:23:25.600 |
like we have to expand our conception of time here. 00:23:30.780 |
Time scale, I mean, over across maybe 100 generations, 00:23:40.560 |
- Maybe, or maybe they'll watch our colony die out 00:23:43.880 |
over here and be like, we're not making chips. 00:23:46.680 |
- No, but you have to seed that colony correctly. 00:24:06.800 |
And he usually makes tattoos and nice branding. 00:24:12.600 |
Humanity works really hard today to get rid of that asshole, 00:24:19.800 |
The freedom of being an asshole seems kind of important. 00:24:29.680 |
And now it's building artificial copies of itself, 00:24:34.000 |
or artificial copies of various aspects of itself 00:24:43.280 |
- I like to think it's just like another stack for life. 00:24:46.520 |
We have like the biostack life, like we're a biostack life, 00:25:12.160 |
And you don't have anything like this in the biostack. 00:25:16.040 |
I tried to make a meme, it didn't work too well. 00:25:17.840 |
But I posted a picture of Ronald Reagan and Joe Biden, 00:25:21.360 |
and you look, this is 1980, and this is 2020. 00:25:24.360 |
And these two humans are basically like the same. 00:25:26.840 |
There's no, like there's been no change in humans 00:25:45.640 |
the size of the fab required to make another fab 00:25:52.840 |
- But computers were very large 80 years ago. 00:26:01.560 |
And people are starting to wanna wear them on their face 00:26:16.000 |
I don't have to see the rest of you assholes. 00:26:29.560 |
- Judging from what you can buy today, far, very far. 00:26:59.720 |
Whereas when I put good headphones on, audio is there. 00:27:13.800 |
The power of imagination or the power of the mechanism 00:27:20.840 |
that kind of reaches and wants to make the thing you see 00:27:24.120 |
in the virtual world real to you, I believe in that power. 00:27:36.760 |
and here's a world where you don't have to struggle anymore? 00:27:41.520 |
that people think the large language models are conscious. 00:27:50.680 |
Why do you think large language models are not conscious? 00:27:55.720 |
- Oh, so what is consciousness then, George Hart? 00:28:01.320 |
It's just like a word that atheists use for souls. 00:28:18.440 |
- When is the last time you've seen a chicken? 00:28:27.600 |
- Living chickens walking around Miami, it's crazy. 00:28:41.360 |
Okay, but you don't think much about this kind of 00:28:47.600 |
subjective feeling that it feels like something to exist. 00:28:56.680 |
And then as an observer, you can have a sense 00:29:05.180 |
but has a kind of subjective experience of its reality, 00:29:09.760 |
like a self-awareness that is capable of like suffering, 00:29:13.360 |
of hurting, of being excited by the environment 00:29:15.520 |
in a way that's not merely kind of an artificial response, 00:29:29.760 |
- Yeah, and you're saying when we look in the mirror, 00:29:38.160 |
- Isn't that weird though, that you're not conscious? 00:29:47.360 |
Okay, so to you it's like a little like a symptom 00:29:50.240 |
of the bigger thing that's not that important. 00:29:52.360 |
- Yeah, I mean, it's interesting that like human systems 00:29:58.920 |
in a straight up like, okay, what do people mean 00:30:00.720 |
when even if you don't believe in consciousness, 00:30:02.400 |
what do people mean when they say consciousness? 00:30:14.320 |
- Don't say pineapple. - I like pepperoni pizza. 00:30:17.360 |
- And if they put any ham on it, oh, that's real bad. 00:30:49.280 |
There's just the full peak light living experience 00:30:54.280 |
of being human, the top of the human condition. 00:31:06.680 |
- If that's the word you wanna use to describe it, sure. 00:31:10.720 |
I'm not gonna deny that I experienced that feeling. 00:31:24.440 |
- How would you know what suffering looked like? 00:31:30.800 |
All the bio stack stuff kind of, especially mammals, 00:31:42.800 |
You have, wow, the little thing has learned to mimic. 00:31:46.080 |
But then I realized that that's all we are too. 00:31:52.280 |
Oh look, the little thing has learned to mimic. 00:31:54.680 |
- Yeah, I guess, yeah, 404 could be suffering, 00:31:58.760 |
but it's so far from our kind of living organism, 00:32:06.400 |
But it feels like AI can start maybe mimicking 00:32:14.960 |
- And so in that, maybe that's the definition 00:32:17.080 |
of consciousness, is the bio stack consciousness. 00:32:20.080 |
- The definition of consciousness is how close 00:32:24.520 |
- No, how close something is to the human experience. 00:32:28.920 |
- Sure, it's a very anthropocentric definition, but. 00:32:44.040 |
- I wanna find out what your fourth company is after. 00:32:46.720 |
- 'Cause I think once you have AI Girlfriends, 00:33:05.600 |
having the capacity to experience this life richly 00:33:14.280 |
Or you can project your anthropomorphic sense 00:33:28.880 |
- Yeah, but, okay, so here's where it actually 00:33:38.600 |
- When you interact with these models, you can't. 00:33:40.440 |
You can make some assumptions that that other human 00:33:49.500 |
With an AI model, this isn't really true, right? 00:33:52.620 |
These large language models are good at fooling people 00:33:55.140 |
because they were trained on a whole bunch of human data 00:33:59.860 |
But if the AI system says, "Hi, my name is Samantha," 00:34:09.500 |
- Maybe it'll integrate this in the AI system. 00:34:21.900 |
- It seems pretty natural for you to anthropomorphize 00:34:28.180 |
And before you know it, you're married and have kids. 00:34:35.620 |
There's pictures on Instagram with you and a rock 00:34:38.740 |
- To be fair, something that people generally look for 00:34:44.940 |
And the rock doesn't really have intelligence. 00:34:47.100 |
Only a pretty desperate person would date a rock. 00:34:55.300 |
Not rock level desperate, but AI level desperate. 00:35:06.140 |
It just feels like the language models are there. 00:35:17.340 |
You know, again, the language models now are still a little, 00:35:21.140 |
like people are impressed with these GPT things. 00:35:24.860 |
And I look at like, or like, or the copilot, the coding one. 00:35:29.700 |
And I'm like, okay, this is like junior engineer level. 00:35:32.380 |
And these people are like fiver level artists 00:35:53.420 |
- Is it count as cheating when you're talking 00:36:01.200 |
- That's up to you and your human partner to define. 00:36:13.240 |
I mean, integrate that with porn and all this. 00:36:26.380 |
serious open conversations about all the different aspects 00:36:34.060 |
And it feels like AI is a really weird conversation to have. 00:36:40.700 |
Like these things, you know, one of my scenarios 00:36:58.200 |
And so is that, if you're actually dating somebody 00:37:06.680 |
It's like, what are you allowed to say to an AI bot? 00:37:09.160 |
Imagine having that conversation with a significant other. 00:37:11.620 |
I mean, these are all things for people to define 00:37:14.280 |
What it means to be human is just gonna start to get weird. 00:37:22.480 |
what you think is a real human you interacted with 00:37:25.200 |
on Twitter for years and you realize it's not. 00:37:27.460 |
- I spread, I love this meme, heaven banning. 00:37:39.360 |
but a whole lot of AIs are spot up to interact with you. 00:37:42.300 |
- Well, maybe that's what the way human civilization ends 00:37:56.660 |
- Friendship is optimal. - Friendship is optimal. 00:38:09.980 |
But not out of the realm of the kind of weirdness 00:38:16.240 |
that human civilization is capable of, I think. 00:38:23.460 |
- Yeah, I think a lot of people probably want it. 00:38:30.260 |
and just will only advertise to you some of the time. 00:38:33.580 |
- Yeah, maybe the conceptions of monogamy change too. 00:38:36.260 |
Like I grew up in a time, like I value monogamy, 00:38:40.260 |
when you have arbitrary number of AI systems. 00:38:43.280 |
- This interesting path from rationality to polyamory. 00:38:50.180 |
- For you, but you're just a biological organism 00:38:52.740 |
who was born before the internet really took off. 00:39:04.500 |
like, is a lot of problem in moral philosophy, right? 00:39:09.860 |
that, like, computers are capable of mimicking, 00:39:14.700 |
They passed the girlfriend Turing test, right? 00:39:21.180 |
That doesn't say we ought to get rid of monogamy, right? 00:39:27.500 |
- Girlfriend Turing test, I wonder what that looks like. 00:39:32.340 |
Will you be the Alan Turing of the 21st century 00:39:40.840 |
their goal is to pass the girlfriend Turing test. 00:39:46.900 |
I mean, the question is if it's deeply personalized 00:39:50.980 |
or there's a common thing that really gets everybody. 00:39:54.320 |
- Yeah, I mean, you know, look, we're a company. 00:39:58.420 |
We just have to get a large enough clientele to stay. 00:40:00.820 |
- I like how you're already thinking company. 00:40:03.980 |
All right, let's, before we go to company number three 00:40:06.500 |
and company number four, let's go to company number two. 00:40:10.420 |
Possibly one of the greatest names of all time for a company. 00:40:15.840 |
You've launched a new company called Tiny Corp 00:40:20.900 |
What's the origin story of Tiny Corp and Tiny Grad? 00:40:25.020 |
- I started Tiny Grad as a, like, a toy project 00:40:28.580 |
just to teach myself, okay, like, what is a convolution? 00:40:32.460 |
What are all these options you can pass to them? 00:40:34.660 |
What is the derivative of a convolution, right? 00:41:02.980 |
- So you, one of the reasons to start Tiny Corp 00:41:15.260 |
And it's to make sure power stays decentralized. 00:41:30.580 |
If NVIDIA becomes just like 10x better than everything else, 00:41:53.220 |
- No, but there's so much, you know, there's AMD. 00:42:03.700 |
towards like selling, like Google selling TPUs 00:42:22.420 |
I was like, okay, what's it gonna take to make a chip? 00:42:24.620 |
And my first notions were all completely wrong 00:42:26.980 |
about why, about like how you could improve on GPUs. 00:42:30.260 |
And I will take this, this is from Jim Keller 00:42:39.740 |
So there's three kinds of computation paradigms 00:43:08.940 |
Then you have a simpler computation model GPUs. 00:43:13.300 |
I mean, they can, but it's horrendously slow. 00:43:15.880 |
But GPUs can do arbitrary load and store, right? 00:43:21.580 |
So they can fetch from arbitrary pieces of memory. 00:43:39.340 |
and 95% of neural networks are all the DSP paradigm. 00:43:43.000 |
They are just statically scheduled adds and multiplies. 00:43:55.300 |
Every stage of the stack has Turing completeness, right? 00:44:09.820 |
which is Turing complete on a Turing complete processor. 00:44:12.180 |
I want to get Turing completeness out of the stack entirely. 00:44:14.860 |
Because once you get rid of Turing completeness, 00:44:28.500 |
are we talking about the hardware or the software? 00:44:37.740 |
So the reason you need to do branch prediction in a CPU, 00:44:46.920 |
Well, they get 1% wrong because you can't know, right? 00:44:53.220 |
to say whether a branch is gonna be taken or not. 00:45:01.300 |
the neural network, runs the identical compute every time. 00:45:38.460 |
and I put my two matrices that I'm multiplying 00:45:47.380 |
I'm gonna multiply for each one in the cubed, 00:46:09.620 |
- Okay, so what is the minimum number of operations 00:46:30.900 |
PyTorch 2.0 introduced PrimTorch, which has only 250. 00:46:54.620 |
- RISC architecture is gonna change everything. 00:47:03.060 |
RISC architecture is gonna change everything in 1995. 00:47:16.420 |
so you're thinking of this as the RISC architecture 00:47:48.740 |
Almost all activation functions are unary ops. 00:48:09.940 |
Reduce ops will like take a three-dimensional tensor 00:48:24.020 |
And movement ops are different from the other types 00:48:25.820 |
because they don't actually require computation. 00:48:27.620 |
They require different ways to look at memory. 00:48:30.060 |
So that includes reshapes, permutes, expands, flips. 00:48:35.600 |
- And so with that, you have enough to make a map model. 00:48:55.180 |
it is fused into one kernel with the power of laziness. 00:49:17.760 |
reading C from memory and writing that out to memory. 00:49:25.100 |
If you don't actually do A times B as soon as you see it, 00:49:28.660 |
if you wait until the user actually realizes that tensor, 00:49:36.740 |
This is like, it's the same way Haskell works. 00:49:39.060 |
- So what's the process of porting a model into TinyGrad? 00:49:44.060 |
- So TinyGrad's front end looks very similar to PyTorch. 00:49:52.940 |
I think that there's some things that are nicer 00:50:00.140 |
We have more ONNX tests passing than Core ML. 00:50:08.140 |
- What about the developer experience with TinyGrad? 00:50:18.100 |
I think that it's actually a very good piece of software. 00:50:20.900 |
I think that they've made a few different trade-offs, 00:50:23.860 |
and these different trade-offs are where TinyGrad 00:50:29.700 |
One of the biggest differences is it's really easy 00:50:32.060 |
to see the kernels that are actually being sent to the GPU. 00:50:44.100 |
you don't know how much memory accesses were used. 00:50:57.500 |
- So can you just linger on what problem TinyGrad solves? 00:51:08.940 |
One of the reasons, tons of these companies now, 00:51:12.060 |
I think Sequoia marked Graphcore to zero, right? 00:51:20.660 |
all of these ML accelerator companies, they built chips. 00:51:29.340 |
I think the same problem is happening with Dojo. 00:51:31.380 |
It's really, really hard to write a PyTorch port, 00:51:37.460 |
and you have to tune them all for performance. 00:51:45.620 |
so he was involved, he's involved with Tenstorrent. 00:51:58.420 |
is that they're gonna pivot to making RISC-V chips. 00:52:06.060 |
- Well, because AI accelerators are a software problem, 00:52:17.180 |
in the hardware space is not going to be a thing 00:52:21.340 |
- I think what's gonna happen is, if I can finish, okay. 00:52:31.140 |
a torch-level performance stack on NVIDIA GPUs. 00:52:35.260 |
If you can't write a torch stack on NVIDIA GPUs, 00:52:37.660 |
and I mean all the way, I mean down to the driver, 00:52:39.740 |
there's no way you're gonna be able to write it on your chip, 00:52:41.860 |
because your chip's worse than an NVIDIA GPU. 00:52:46.540 |
- Oh, you're saying writing that stack is really tough. 00:52:51.420 |
almost always 'cause you're trying to get advantage 00:52:52.940 |
over NVIDIA, you're specializing the hardware more. 00:53:02.580 |
there's no way you can write a stack for your chip. 00:53:10.740 |
- So you did say a few to NVIDIA a little bit, with love. 00:53:20.060 |
- Oh, you're a Mets fan, a RISC fan and a Mets fan. 00:53:26.060 |
You did a build with AMD recently that I saw. 00:53:30.540 |
How does the 7900XTX compare to the RTX 4090 or 4080? 00:53:42.740 |
and if you run demo apps in loops, it panics the kernel. 00:53:56.780 |
Like, I understand if your 7x7 transposed Winograd conv 00:54:02.900 |
but literally when I run demo apps in a loop, 00:54:09.540 |
- Yeah, I just literally took their demo apps 00:54:12.180 |
and wrote like while true semicolon do the app 00:54:17.580 |
Right, this is like the most primitive fuzz testing. 00:54:21.940 |
They're just not seeing a market in machine learning? 00:54:26.620 |
- They're changing, they're trying to change. 00:54:29.660 |
and I had a pretty positive interaction with them this week. 00:54:31.900 |
Last week I went on YouTube, I was just like, that's it. 00:54:37.500 |
I'm not gonna, you know, I'll go with Intel GPUs. 00:55:02.660 |
there's things that centralize power and they're bad, 00:55:04.860 |
and there's things that decentralize power and they're good. 00:55:07.820 |
Everything I can do to help decentralize power, 00:55:10.820 |
- So you're really worried about the centralization 00:55:17.260 |
for the proliferation of ASICs, except in the cloud. 00:55:31.540 |
because Google wrote a machine learning framework. 00:55:34.940 |
I think that you have to write a competitive machine 00:55:37.380 |
learning framework in order to be able to build an ASIC. 00:55:40.140 |
- You think Meta with PyTorch builds a competitor? 00:55:55.860 |
- What do you think is the fundamental limitation of cloud? 00:55:58.420 |
- Fundamental limitation of cloud is who owns the off switch. 00:56:04.780 |
- And you don't like the man to have all the power. 00:56:09.420 |
And right now, the only way to do that is with Nvidia GPUs 00:56:17.400 |
It's a costly investment emotionally to go with AMDs. 00:56:28.080 |
What's your advice on how to build a good custom PC 00:56:31.240 |
for, let's say, for the different applications that you use 00:56:39.440 |
- I heard rumors, whispers about this box in the tiny corp. 00:56:57.740 |
It's over five terabytes per second of GPU memory bandwidth. 00:57:07.000 |
You're gonna get like 20, 30 gigabytes per second 00:57:11.920 |
I'm gonna build like the best deep learning box that I can 00:57:20.140 |
Can you go through those specs again a little bit 00:57:23.120 |
- Yeah, so it's almost a paid a flop of compute. 00:57:28.780 |
But we're pretty agnostic to the type of compute. 00:57:33.600 |
The main limiting spec is a 120 volt 15 amp circuit. 00:57:56.440 |
And one of the answers is an electric car charger. 00:58:11.360 |
What is the most amount of power you can get your hands on 00:58:40.480 |
I wanna deliver a really good experience to people. 00:58:49.760 |
You know, in my email, what I said to AMD is like, 00:58:53.400 |
just dumping the code on GitHub is not open source. 00:59:08.720 |
I see a real future for AMD as a competitor to Nvidia. 00:59:20.240 |
I'm like, all right, $100 fully refundable pre-orders. 00:59:34.240 |
but like shipping and all this kind of stuff. 00:59:35.880 |
- The thing that I wanna deliver to people out of the box 00:59:38.160 |
is being able to run 65 billion parameter LLAMA 00:59:44.040 |
like 10 tokens per second or five tokens per second 00:59:52.800 |
- Experience, yeah, or I think Falcon is the new one. 00:59:55.920 |
Experience a chat with the largest language model 01:00:05.240 |
it's not like even more power would help you get more. 01:00:10.960 |
- Well, no, there's just the biggest model released 01:00:12.920 |
is 65 billion parameter LLAMA as far as I know. 01:00:16.240 |
- So it sounds like Tiny Box will naturally pivot 01:00:33.840 |
what, why is it harder to replace a boyfriend 01:00:41.200 |
- Because women are attracted to status and power 01:00:49.000 |
- Both are mimicable easy through the language model. 01:00:51.920 |
- No, no machines do not have any status or real power. 01:00:59.120 |
you're using language mostly to communicate youth 01:01:07.640 |
- But status fundamentally is a zero sum game, right? 01:01:12.160 |
- No, I think status is a narrative you can construct. 01:01:25.400 |
- I also think that men are probably more desperate 01:01:36.840 |
- Yeah, look, I mean, look, I know you can look 01:01:43.680 |
- Wow, there's a lot of questions than answers 01:01:47.440 |
Anyway, with the tiny box, how many GPUs in tiny box? 01:02:01.320 |
- So AMD Epyc processors have 128 lanes of PCIe. 01:02:11.600 |
And I wanna leave enough lanes for some networking. 01:02:15.840 |
- How do you do cooling for something like this? 01:02:22.920 |
I want the tiny box to be able to sit comfortably 01:02:25.880 |
- This is really going towards the girlfriend thing. 01:02:37.880 |
- Well, but yes, quiet, oh, quiet because you may be 01:02:43.040 |
- No, no, quiet because you wanna put this thing 01:02:44.960 |
in your house and you want it to coexist with you. 01:02:46.880 |
If it's screaming at 60 dB, you don't want that 01:03:00.760 |
But if I can make it big, a lot of that noise 01:03:08.200 |
has these super high pressure fans that are like 01:03:14.600 |
well, I can use a big, you know, they call them 01:03:18.600 |
on the ceiling and they're completely silent. 01:03:29.200 |
I want it to be shippable as a normal package, 01:04:10.360 |
- Yeah, we did a poll of people want Ubuntu or Arch. 01:04:19.360 |
I like Ubuntu Mate, however you pronounce that, mate. 01:04:23.480 |
So how do you, you've gotten Lama into TinyGrad. 01:04:27.200 |
You've gotten stable diffusion into TinyGrad. 01:04:29.920 |
Can you comment on like, what are these models? 01:04:39.240 |
What's naturally, what's easy, all that kind of stuff. 01:04:41.800 |
- There's a really simple way to get these models 01:04:43.880 |
into TinyGrad, and you can just export them as Onyx, 01:04:47.960 |
So the ports that I did of Lama, stable diffusion, 01:04:56.320 |
but they are cleaner than the PyTorch versions. 01:05:22.040 |
- But that's more like a software engineering thing, 01:05:25.920 |
or do you think it has a cost on performance? 01:05:28.080 |
- Oh no, it doesn't have a cost on performance. 01:05:32.520 |
that's what I mean about TinyGrad's front end being cleaner. 01:05:39.360 |
to the programming language that does some interesting ideas 01:05:52.600 |
GGML is this like, we're gonna run LLAMA fast on Mac. 01:05:56.600 |
Okay, we're gonna expand out to a little bit, 01:05:58.120 |
but we're gonna basically go to like depth first, right? 01:06:07.360 |
TinyGrad is, we are going to make neural networks fast. 01:06:11.280 |
- Yeah, but they try to really get it to be fast, 01:06:31.640 |
So you're actually going to be targeting some accelerators, 01:06:41.360 |
build an equally performance stack to PyTorch 01:06:48.080 |
And then step two is, okay, how do we make an accelerator? 01:07:06.640 |
build it the right way and worry about performance later. 01:07:09.920 |
There's a bunch of things where I haven't even like 01:07:19.800 |
So TinyGrad's actually used in OpenPilot to run the model. 01:07:33.280 |
- What's the hardware that OpenPilot runs on, the CommAI? 01:07:55.080 |
you can use buffers on a mobile GPU image textures 01:08:09.520 |
in a way that it's completely generic, right? 01:08:16.560 |
where they can generate, where they have these kernels, 01:08:21.320 |
So that's great if you're doing three by three comps. 01:08:23.800 |
That's great if you're doing dense map models, 01:08:25.600 |
but the minute you go off the beaten path a tiny bit, 01:08:31.800 |
I'd love to get an update in the company number one, 01:08:37.120 |
How are things going there in the development 01:08:42.600 |
- You know, almost no one talks about FSD anymore, 01:08:51.240 |
We've solved the problem, like we solved it years ago. 01:09:35.640 |
we published a paper called "Learning a Driving Simulator." 01:09:39.800 |
And the way this thing worked was it was an autoencoder, 01:09:48.520 |
You take an autoencoder, you compress the picture, 01:09:57.160 |
Like this is 2015 era machine learning technology. 01:10:21.680 |
- Well, actually, our simulator's conditioned on the pose. 01:10:41.720 |
not asking is this close to the human policy, 01:10:43.760 |
but asking would a human disengage if you did this behavior? 01:10:47.480 |
- Okay, let me think about the distinction there. 01:11:00.200 |
So it doesn't just say, what would a human do? 01:11:25.400 |
- 'Cause usually disengagement is almost always a sign 01:11:29.640 |
of I'm not happy with what the system is doing. 01:11:32.960 |
There's some that are just, I felt like driving, 01:11:36.480 |
but they're just gonna look like noise in the data. 01:11:42.680 |
- Even that's a signal, like, why do you feel like driving? 01:11:45.600 |
You need to recalibrate your relationship with the car. 01:12:16.720 |
Are these major philosophical bugs, logical bugs? 01:12:26.720 |
We just massively expanded our compute cluster at Gama. 01:12:30.000 |
We now have about two people worth of compute, 01:12:44.960 |
- Yeah, but there's something different about mobility 01:12:53.800 |
- Well, yeah, of course, not all flops are created equal. 01:12:55.720 |
If you have randomly initialized weights, it's not gonna. 01:13:00.640 |
- Some flops are doing way more useful things than others. 01:13:31.160 |
- How's that race going between CalmAI and FSD? 01:13:34.320 |
- Tesla is always one to two years ahead of us. 01:13:36.280 |
They've always been one to two years ahead of us, 01:13:41.560 |
- What have you seen that's, since the last time we talked, 01:13:43.640 |
that are interesting architectural decisions, 01:13:45.360 |
training decisions, like the way they deploy stuff, 01:13:48.160 |
the architectures they're using in terms of the software, 01:13:51.000 |
how the teams are run, all that kind of stuff, 01:13:58.040 |
- So creeping towards end-to-end as much as possible 01:14:03.160 |
the training, the data collection, everything. 01:14:06.920 |
They're probably saying all the same things we are. 01:14:08.880 |
They're probably saying we just need to optimize, 01:14:12.320 |
Will you get a negative reward for disengagement, right? 01:14:16.000 |
It's just a question of who can actually build 01:14:18.960 |
- Yeah, I mean, this requires good software engineering, 01:14:27.920 |
- You still don't believe in cloud in that regard? 01:14:30.420 |
- I have a compute cluster in my office, 800 amps. 01:14:37.560 |
- It's 40 kilowatts at idle, our data center. 01:14:44.080 |
- Just when I-- - Sorry, sorry, compute cluster. 01:14:57.800 |
- I'm guessing this is a kind of a legal distinction 01:15:05.280 |
- You said that you don't think LLMs have consciousness, 01:15:13.560 |
about the word reason, about some of the capabilities 01:15:18.040 |
to be able to integrate complicated information 01:15:39.200 |
I think that they can reason better than a lot of people. 01:15:47.640 |
- I mean, I think that calculators can add better 01:16:16.800 |
- No, no, no, because it's very human, we take it, 01:16:25.800 |
Chess is a game that a subset of population plays. 01:16:34.360 |
and human interaction is fundamental to society. 01:16:57.560 |
Someone in 2010 won't, what's changed, right? 01:17:04.160 |
but I think in 20 years, we're gonna be like, 01:17:17.320 |
- Humans are always gonna define a niche for themselves. 01:17:21.720 |
because we can, and they tried creative for a bit, 01:17:31.360 |
Because maybe with chess, you start to realize 01:17:44.680 |
- Yeah, and I think maybe we're gonna go through 01:17:52.780 |
- But language carries these notions of truth and so on, 01:18:10.720 |
and machines are created by humans, therefore. 01:18:13.320 |
Right, like that'll be the last niche we have. 01:18:20.880 |
It's still incredibly impressive, like with ChagGPT. 01:18:24.700 |
about reinforcement learning with human feedback 01:18:28.980 |
- I'd like to go back to when calculators first came out, 01:18:55.840 |
- Refrigerator, electricity, all that kind of stuff. 01:19:13.680 |
But you're saying like other technologies have as well. 01:19:17.160 |
So maybe calculator's not the best example of that? 01:19:20.680 |
'Cause that just seems like, well, no, maybe. 01:19:28.240 |
You're telling me you can just keep the milk in your house? 01:19:36.200 |
at the practical impacts of certain technologies 01:19:46.120 |
- I do think it's different this time, though. 01:19:56.120 |
- It feels like it's getting smaller rapidly, though. 01:20:00.200 |
Or is that just a feeling we dramatize everything? 01:20:08.400 |
"Are they gonna have one of these in every home?" 01:20:35.960 |
The language models are writing middle school level essays 01:20:41.720 |
and people are like, "Wow, it's a great essay. 01:21:04.800 |
I spend 5% of time typing and 95% of time debugging. 01:21:08.000 |
The last thing I want is close to correct code. 01:21:10.720 |
I want a machine that can help me with the debugging, 01:21:25.840 |
- I actually don't think it's like level two driving. 01:21:28.480 |
I think driving is not tool complete and programming is. 01:21:30.920 |
Meaning you don't use the best possible tools to drive. 01:21:41.400 |
- Computers have a radically different interface. 01:21:42.960 |
- Okay, can you describe the concept of tool complete? 01:21:47.280 |
So think about the difference between a car from 1980 01:21:52.040 |
It's got a bunch of pedals, it's got a steering wheel. 01:21:58.640 |
Right, you have no problem getting into a 1980 car 01:22:01.400 |
Take a programmer today who spent their whole life 01:22:04.240 |
doing JavaScript and you put them in an Apple IIe prompt 01:22:07.400 |
and you tell them about the line numbers in basic. 01:22:09.840 |
But how do I insert something between line 17 and 18? 01:22:22.600 |
So it's just the entirety stack of the tooling. 01:22:25.680 |
- So it's not just like the, like IDs or something like this, 01:22:28.640 |
- Yes, it's IDEs, the languages, the runtimes, 01:22:33.280 |
So like almost if Codex or Copilot are helping you, 01:22:38.280 |
that actually probably means that your framework 01:22:41.720 |
or library is bad and there's too much boilerplate in it. 01:22:44.520 |
- Yeah, but don't you think so much programming 01:23:06.960 |
- Well, let's talk about good code and bad code. 01:23:16.000 |
for generic scripts that I write just offhand, 01:23:25.160 |
So not like libraries, not like performing code, 01:23:27.960 |
not stuff for robotics and so on, just quick stuff. 01:23:42.960 |
with like generic method, like a generic kind of ID 01:23:45.960 |
type of recommendation or something like this. 01:23:55.640 |
maybe today if I wrote a lot of data parsing stuff, 01:24:00.880 |
but if I still played CTFs, a lot of it is just like 01:24:02.880 |
you have to write a parser for this data format. 01:24:08.440 |
I wonder when the models are gonna start to help 01:24:29.200 |
If I'm writing some script to just like parse 01:24:33.440 |
My programming speed is limited by my typing speed. 01:24:36.920 |
'Cause that's essentially a more efficient lookup, right? 01:24:42.160 |
I tried to use a chat GPT to like ask some questions, 01:24:49.960 |
It would just give me completely made up API functions 01:24:54.440 |
- Well, do you think that's just a temporary kind of stage? 01:24:58.640 |
- You don't think it'll get better and better and better 01:25:01.160 |
'Cause like it only hallucinates stuff in the edge cases. 01:25:05.040 |
- If you're writing generic code, it's actually pretty good. 01:25:12.240 |
No, there's kind of ways to fix the hallucination problem. 01:25:17.760 |
And it's actually weird the way that we do language models 01:25:20.840 |
right now where all of the information is in the weights. 01:25:33.000 |
I think future LLMs are gonna be like smaller, 01:25:59.720 |
That's pushing it towards level two kind of journey. 01:26:13.120 |
- That's what people want in a search engine. 01:26:14.640 |
- But also Google might be the people that build it. 01:26:30.880 |
yeah, I mean, you're a legitimate competitor in that. 01:26:40.520 |
- You don't think you might build a search engine 01:26:43.240 |
- When I started Comma, I said over and over again, 01:26:52.800 |
and I'm never going to say that 'cause I won't. 01:27:03.000 |
one of the things that ChatterGPT kind of shows 01:27:06.920 |
that really have, that create a really compelling product. 01:27:15.160 |
Google won't be the number one webpage anymore. 01:27:21.760 |
- Look, I would put a lot more money on Mark Zuckerberg. 01:27:35.440 |
Facebook's alive. - Versus Facebook is alive. 01:27:41.640 |
Like, that's just, like, like, like Mark Zuckerberg. 01:27:43.400 |
This is Mark Zuckerberg reading that Paul Graham essay 01:27:49.080 |
- So you don't think there's this gutsy pivoting engine 01:27:57.720 |
the kind of engine that a startup has, like, constantly. 01:28:03.000 |
- When I listened to your Sam Altman podcast, 01:28:06.600 |
Everyone who talks about AI talks about the button, 01:28:12.560 |
Is anybody in the world capable of shutting Google down? 01:28:24.280 |
- Can you elaborate on the value of that question? 01:28:32.640 |
That's a good question, right? - Does anyone? 01:28:41.640 |
Let's say Sundar Pichai made this his sole mission. 01:28:47.800 |
I don't think he'd keep his position too long. 01:28:55.280 |
- Well, boards and shares and corporate undermining 01:29:02.760 |
- Okay, so what's the case you're making here? 01:29:23.520 |
I think he does, and this is exactly what I mean 01:29:25.720 |
and why I bet on him so much more than I bet on Google. 01:29:29.120 |
- I guess you could say Elon has similar stuff. 01:29:33.600 |
- Elon, does Elon, can Elon fire the missiles? 01:29:39.000 |
- I think some questions are better left unasked. 01:29:45.440 |
well, you're a rocket that can land anywhere. 01:29:48.080 |
Well, you know, don't ask too many questions. 01:29:57.240 |
is that you can innovate aggressively, is what you're saying, 01:30:06.560 |
- I bet on something that looks like mid-journey, 01:30:09.560 |
- Just is able to site source a loop on itself. 01:30:13.920 |
I mean, it just feels like one model can take off. 01:30:25.120 |
is there is some aspect of a winner-take-all effect, right? 01:30:31.920 |
that gets a lot of usage, and you see this with OpenAI, 01:30:41.320 |
you know, I was actually at Google Image Search 01:30:42.720 |
when I worked there like almost 15 years ago now. 01:30:44.920 |
How does Google know which image is an apple? 01:30:48.000 |
And they're like, yeah, that works about half the time. 01:30:50.960 |
You'll see they're all apples on the front page 01:30:54.200 |
And I don't know, I didn't come up with the answer. 01:30:57.000 |
The guy's like, well, it's what people click on 01:31:00.280 |
- Yeah, yeah, that data is really, really powerful. 01:31:06.000 |
What do you think in general that Lama was open sourced? 01:31:09.480 |
I just did a conversation with Mark Zuckerberg, 01:31:16.400 |
- Who would have thought that Mark Zuckerberg 01:31:23.360 |
- Who would have thought anything in this world? 01:31:27.280 |
But open source to you ultimately is a good thing here. 01:31:35.520 |
You know, what's ironic about all these AI safety people 01:31:39.600 |
is they are going to build the exact thing they fear. 01:31:42.200 |
These we need to have one model that we control and align, 01:31:47.720 |
this is the only way you end up paper clipped. 01:32:02.040 |
- So the criticism you have for the AI safety folks 01:32:05.200 |
is that there is a belief and a desire for control. 01:32:10.200 |
And that belief and desire for centralized control 01:32:21.200 |
and is a 16-way mixture model with eight sets of weights? 01:32:24.100 |
- Who did you have to murder to get that information? 01:32:41.680 |
and raised a whole fake AI safety thing about that, 01:32:46.520 |
Like they used AI safety to hype up their company 01:33:10.240 |
- I don't know how much hype there is in AI safety, 01:33:21.040 |
I think OpenAI has been finding an interesting balance 01:33:24.080 |
between transparency and putting value on AI safety. 01:33:44.200 |
So like the model that can be ultra racist and dangerous 01:33:48.720 |
and like tell you how to build a nuclear weapon. 01:33:57.480 |
This makes, this allows you to meet every human. 01:34:00.520 |
- Yeah, I know, but half of these AI alignment problems 01:34:06.900 |
It's like, it's not the machines you want to align, it's me. 01:34:17.960 |
questions where the answers have dangerous consequences 01:34:28.400 |
- Well, no, for me, there's a lot of friction. 01:34:40.560 |
Do I use Bing or do I, which search engine do I use? 01:34:44.920 |
- No, it feels like I have to keep clicking a lot of this. 01:34:57.040 |
I feel like a language model makes it more accessible 01:35:02.040 |
for that person who's not smart enough to do-- 01:35:11.340 |
how to ask that question a bit more academically 01:35:13.720 |
and get a real answer from it are not capable 01:35:15.760 |
of procuring the materials, which are somewhat controlled, 01:35:21.300 |
to people with money without the technical know-how, right? 01:35:24.880 |
- To build a, like, do you really need to know 01:35:32.240 |
you know what, I was asking this question on my stream, 01:35:36.180 |
- But a language model can probably help you out. 01:35:43.240 |
Like, it's not like the language model is God. 01:35:46.440 |
it's you literally just hired someone on Fiverr. 01:35:50.240 |
- But, okay, okay, GPT-4, in terms of finding a hitman, 01:35:52.920 |
it's like asking Fiverr how to find a hitman. 01:35:57.200 |
- Wikihow, but don't you think GPT-5 will be better? 01:36:07.280 |
or build a bomb, they'd also be serious enough 01:36:18.460 |
of how hard is it to find that kind of hitman. 01:36:20.760 |
I honestly think there's a jump in ease and scale 01:36:34.920 |
who are not intelligent are going to use machines 01:36:45.960 |
When I'm in the woods, the scariest animal to meet 01:37:03.080 |
When you go to the Amazon, it's the human tribes. 01:37:13.560 |
who maybe has ill intention but is not so intelligent 01:37:20.520 |
So we should have intelligence control, of course. 01:37:23.720 |
We should only give intelligence to good people. 01:37:30.040 |
the best defense is to give more intelligence 01:37:36.240 |
You know what, it's not even like guns, right? 01:37:38.320 |
You know, what's the best defense against a bad guy 01:37:42.160 |
but I really subscribe to that with intelligence. 01:37:44.600 |
- Yeah, in a fundamental way, I agree with you. 01:37:48.200 |
But there's just feels like so much uncertainty 01:38:00.880 |
I'd want them to lose control more than anything else. 01:38:05.340 |
- I think when you lose control, you can do a lot of damage, 01:38:07.800 |
but you can do more damage when you centralize 01:38:12.720 |
- Centralized and held control is tyranny, right? 01:38:26.400 |
So to you, open source is the way forward here. 01:38:32.360 |
or what Meta is doing with the release of the-- 01:38:36.080 |
- I lost $80,000 last year investing in Meta. 01:38:38.280 |
And when they released Lama, I'm like, yeah, whatever, man. 01:38:43.160 |
Do you think Google and OpenAI with Microsoft will match? 01:38:57.800 |
Like you're on the bad team who can't even say 01:39:05.120 |
I'm not saying you need to make your model weights open. 01:39:09.080 |
I totally understand we're keeping our model weights closed 01:39:13.960 |
I'm saying like, because of AI safety reasons, 01:39:27.280 |
- Is it possible that these things can really do 01:39:34.000 |
be it human intelligence or machine intelligence. 01:39:55.720 |
- But you mean like the intelligence agencies in America 01:40:16.920 |
- But I think there's a George Hotz-type character 01:40:19.080 |
that can do a better job than the entirety of them. 01:40:23.120 |
No, and I'll tell you why the George Hotz character can't. 01:40:24.720 |
And I thought about this a lot with hacking, right? 01:40:33.160 |
to slowly and steadily deploy them over five years. 01:40:35.920 |
And this is what intelligence agencies are very good at. 01:41:01.480 |
LLMs and AI and machine intelligence can cause a lot of harm 01:41:09.680 |
just like I will respect someone philosophically 01:41:11.240 |
with the position that nobody should have guns. 01:41:29.700 |
I'm worried about alignment between me and AI company. 01:41:33.120 |
- What do you think Eliezer Yudkowsky would say to you? 01:41:46.440 |
And I think this comes down to a repeated misunderstanding 01:41:56.760 |
- I think that Eliezer Yudkowsky is scared of these things. 01:42:08.040 |
But now you ask about the two possible futures. 01:42:11.160 |
One where a small trusted centralized group of people 01:42:16.160 |
has them and the other where everyone has them. 01:42:19.320 |
And I am much less scared of the second future 01:42:22.120 |
- Well, there's a small trusted group of people 01:42:30.040 |
Again, a nuclear weapon cannot be deployed tactically 01:42:37.320 |
Except maybe in some philosophical mind game kind of way. 01:42:53.460 |
Imagine I had a powerful AI running on my computer 01:42:56.720 |
saying, okay, nice PSYOP, nice PSYOP, nice PSYOP. 01:43:04.400 |
- Yeah, I mean, so you have fundamentally hope for that, 01:43:10.480 |
- I'm not even like, I don't even mean these things 01:43:13.320 |
I mean these things in straight up like ad blocker, right? 01:43:33.680 |
One of the deepest optimisms I have is just like, 01:43:49.640 |
- Yeah, I'm not even gonna say there's a lot of good guys. 01:44:00.120 |
I mean, if you believe philosophically in democracy, 01:44:02.360 |
you obviously believe that, that good outnumbers bad. 01:44:14.680 |
but there's also a chance you gave it to bad people. 01:44:16.720 |
If you give it to everybody, well, if good outnumbers bad, 01:44:19.800 |
then you definitely gave it to more good people than bad. 01:44:27.180 |
but then also, of course, there's other motivations, 01:44:29.640 |
like you don't wanna give away your secret sauce. 01:44:32.200 |
- Well, that's, I mean, I look, I respect capitalism. 01:44:34.440 |
I don't think that, I think that it would be polite 01:44:37.200 |
for you to make model architectures open source 01:44:41.800 |
I don't think you have to make weights open source. 01:44:45.960 |
like there's so many possible trajectories in human history 01:44:49.200 |
where you could have the next Google be open source. 01:44:53.220 |
So for example, I don't know if that connection is accurate, 01:44:57.280 |
but Wikipedia made a lot of interesting decisions, 01:45:05.720 |
And like, that's one of the main websites on the internet. 01:45:11.080 |
Google could have created Wikipedia, put ads on it. 01:45:13.600 |
You could probably run amazing ads now on Wikipedia. 01:45:23.280 |
derivatives of open source llama might win the internet. 01:45:35.000 |
And I don't think this is just, all right, come on. 01:45:55.880 |
But you're saying overall, in the long arc of history, 01:46:16.520 |
- Yeah, I mean, some were probably, we got Chrome, right? 01:46:25.280 |
Well, search engine, Maps, Mail, Android, and Chrome. 01:46:34.000 |
you know, I was Times Person of the Year in 2006. 01:46:39.280 |
- It's you, was Times Person of the Year in 2006, right? 01:46:41.800 |
Like that's, you know, so quickly did people forget. 01:46:49.480 |
I think some of it, I hope, look, I hope that, 01:46:57.520 |
I think it might just be like the effects of social media. 01:47:05.920 |
So you're just being an old man who's worried about the, 01:47:08.240 |
I think there's always, it goes, it's a cycle thing. 01:47:11.080 |
and I think people rediscover the power of distributed, 01:47:19.040 |
I think crypto is just carrying the flame of that spirit 01:47:24.960 |
It's just such a shame that they all got rich, you know? 01:47:34.320 |
they sucked all the value out of it and took it. 01:47:36.880 |
- Yeah, money kind of corrupts the mind somehow. 01:47:43.480 |
You had coins worth billions of dollars that had zero use. 01:48:08.120 |
do you think there's some interesting questions there though 01:48:11.360 |
to solve for the open source community in this case? 01:48:13.560 |
So like alignment, for example, or the control problem. 01:48:30.320 |
release a super powerful language model, open source. 01:48:35.920 |
holy shit, okay, what ideas do I have to combat this thing? 01:48:50.280 |
That's what some of these AI safety people seem to think. 01:48:54.760 |
like independently is gonna rebel against its creator. 01:49:10.560 |
it's because the human told it to write viruses. 01:49:36.880 |
to do whatever bad unaligned AI thing you want. 01:49:47.480 |
And if you do anything besides give it to everybody, 01:49:57.640 |
And power turns even slightly good humans to bad. 01:50:04.600 |
- I don't think everyone, I don't think everyone. 01:50:09.840 |
here's the saying that I put in one of my blog posts. 01:50:21.160 |
Like they believed about good things for the world. 01:50:23.160 |
They wanted like flourishing and they wanted growth 01:50:26.480 |
and they wanted things I consider good, right? 01:50:32.680 |
I found 5% of people good and 95% of people bad. 01:50:51.640 |
that promotes the people that run capitalism, 01:50:55.560 |
- That saying may of course be my own biases, right? 01:50:59.840 |
are a lot more aligned with me than these other people. 01:51:04.120 |
- So, you know, I can certainly recognize that. 01:51:15.760 |
- But do you have a concern of super intelligent AGI, 01:51:19.200 |
open sourced, and then what do you do with that? 01:51:28.800 |
I mean, you know, like I'm not a central planner. 01:51:31.240 |
- No, not a central planner, but you'll probably tweet, 01:51:33.640 |
there's a few days left to live for the human species. 01:51:37.320 |
and everyone else has their ideas of what to do with it. 01:51:49.440 |
we create tools that make it more difficult for you 01:51:52.440 |
to maybe make it more difficult for code to spread, 01:51:59.020 |
you know, antivirus software, this kind of thing. 01:52:00.960 |
- Oh, you're saying that you should build AI firewalls? 01:52:03.760 |
You should definitely be running an AI firewall. 01:52:05.880 |
- You should be running an AI firewall to your mind. 01:52:13.080 |
- I don't know if you're being sarcastic or not. 01:52:26.520 |
- I am not being, I would pay so much money for that product. 01:52:52.560 |
and attack anyone else that believes otherwise. 01:52:55.720 |
- Whenever someone's telling me some story from the news, 01:52:57.740 |
I'm always like, I don't wanna hear it, CIA op, bro. 01:53:00.800 |
Like, it doesn't matter if that's true or not. 01:53:20.680 |
And just, actually, just basically feels good. 01:53:29.160 |
It's like, oh, okay, I never thought of it this way. 01:53:35.780 |
when they're like mocking and derisive and just aggressive, 01:53:54.160 |
- And like, you know, I think Elon has a much better chance 01:54:04.580 |
like to build a social network that is actually not toxic 01:54:19.360 |
so make it catalyze the process of connecting cool people 01:54:27.700 |
And like, Scott Alexander has a blog post I like 01:54:30.680 |
where he talks about like moderation is not censorship. 01:54:33.240 |
Like all moderation you want to put on Twitter, right? 01:54:35.920 |
Like you could totally make this moderation, like just a, 01:54:42.500 |
You can just have like a filter button, right? 01:54:44.460 |
That people can turn off if they would like safe search 01:54:47.000 |
Like someone could just turn that off, right? 01:54:48.760 |
So like, but then you'd like take this idea to an extreme. 01:54:58.360 |
these algorithms are designed to maximize engagement. 01:55:00.840 |
Well, it turns out outrage maximizes engagement. 01:55:02.880 |
Quirk of human, quirk of the human mind, right? 01:55:06.140 |
Just as I fall for it, everyone falls for it. 01:55:09.200 |
So yeah, you got to figure out how to maximize 01:55:12.600 |
- And I actually believe that you can make money 01:55:21.760 |
Elon's doing so much stuff right with Twitter, 01:55:37.020 |
- I pay for Twitter, doesn't even get me anything. 01:55:43.020 |
- Sure, but you know, for this business model to work, 01:55:45.760 |
it's like most people should be signed up to Twitter. 01:55:56.440 |
I think that, why do I need most people, right? 01:56:39.400 |
He's like, "Whenever I see someone's Twitter page, 01:56:42.600 |
"I either think the same of them or less of them. 01:56:47.760 |
- Right, like, I don't wanna mention any names, 01:56:58.720 |
- Yeah, but there's some people who would say, 01:57:03.960 |
Are people that just post really good technical stuff. 01:57:38.200 |
People think that they are terrible, awful things. 01:57:40.280 |
And you know, I love that Elon open sourced it. 01:57:42.440 |
Because I mean, what it does is actually pretty obvious. 01:57:44.680 |
It just predicts what you are likely to retweet 01:57:57.160 |
that you are most likely to interact with is outrage. 01:58:02.000 |
- I mean, and there's different flavors of outrage. 01:58:13.520 |
It could be, and maybe there's a better word than outrage. 01:58:22.240 |
it's a constructive thing for the individuals 01:58:26.240 |
- Yeah, so my time there, I absolutely couldn't believe, 01:58:30.000 |
you know, I got crazy amount of hate, you know, 01:58:39.120 |
I think maybe you were exposed to some of this. 01:58:41.600 |
- So connection to Elon or is it working at Twitter? 01:58:44.060 |
- Twitter and Elon, like the whole, there's just-- 01:58:46.720 |
- Elon's gotten a bit spicy during that time. 01:58:54.120 |
it was never go full Republican and Elon liked it. 01:58:58.880 |
- Oh boy, yeah, I mean, there's a rollercoaster of that, 01:59:11.120 |
- And also being, just attacking anybody on Twitter, 01:59:22.360 |
And then letting sort of de-platformed people back on 01:59:34.320 |
- I was hoping, and like, I remember when Elon talked 01:59:37.640 |
about buying Twitter like six months earlier, 01:59:40.680 |
he was talking about like a principled commitment 01:59:47.620 |
I would love to see an actual principled commitment 01:59:54.200 |
Instead of the oligarchy deciding what to ban, 01:59:57.880 |
you had a monarchy deciding what to ban, right? 02:00:00.800 |
Instead of, you know, all the Twitter files, shadow, 02:00:22.100 |
and like, you know, maybe I align more with him 02:00:28.940 |
But I feel like being a free speech absolutist 02:00:31.860 |
on a social network requires you to also have tools 02:00:35.140 |
for the individuals to control what they consume easier. 02:00:45.420 |
like, oh, I'd like to see more cats and less politics. 02:00:48.940 |
- And this isn't even remotely controversial. 02:00:51.320 |
This is just saying you want to give paying customers 02:00:54.480 |
- Yeah, and not through the process of censorship, 02:01:03.320 |
It's individualized transparent censorship, right? 02:01:17.600 |
- I'm looking at you, I'm censoring everything else out 02:01:28.940 |
I think when anyone is allowed to say anything, 02:01:33.420 |
you should probably have tools that maximize the quality 02:01:39.460 |
So, you know, for me, like what I really value, 02:01:42.820 |
boy, it would be amazing to somehow figure out 02:01:49.900 |
who disagree with each other disagree with me, 02:02:00.860 |
There's just a way of talking that's like snarky 02:02:02.980 |
and so on that somehow gets people on Twitter 02:02:07.860 |
- We have like ad hominem refuting the central point. 02:02:18.140 |
to absolutely say what level of Maslow's hierarchy 02:02:28.900 |
that will allow you to have that kind of filter. 02:02:37.920 |
What wins in a free market is all television today 02:02:43.540 |
Right, engaging is what wins in a free market, right? 02:02:47.220 |
So it becomes hard to keep these other more nuanced values. 02:02:50.520 |
- Well, okay, so that's the experience of being on Twitter, 02:03:01.500 |
sort of look, brainstorm when you step into a code base. 02:03:13.340 |
how do we make with a fresh mind progress on this code base? 02:03:17.860 |
Like, what did you learn about software engineering, 02:03:19.980 |
about programming from just experiencing that? 02:03:25.200 |
and I said this on the Twitter spaces afterward, 02:03:27.200 |
I said this many times during my brief internship, 02:03:36.380 |
This code base was, and look, I've worked at Google, 02:03:42.340 |
Facebook has the best code, then Google, then Twitter. 02:03:48.580 |
because look at the machine learning frameworks, right? 02:03:51.420 |
Google released TensorFlow, and Twitter released, 02:04:01.100 |
There's a lot of really good software engineers there, 02:04:08.660 |
- There's so many products, so many teams, right? 02:04:10.500 |
It's very difficult to, I feel like Twitter does less, 02:04:23.420 |
So I can imagine the number of software engineers 02:04:30.780 |
- Yeah, I still believe in the amount of hate I got 02:04:40.740 |
- That you don't know what you're talking about? 02:04:47.660 |
It's like, when I say I'm going to do something, 02:05:01.220 |
and ask the question, why didn't they do anything? 02:05:04.700 |
- And I do think that's where the hate comes from. 02:05:06.060 |
- When you say, well, there's a core truth to that, yeah. 02:05:08.500 |
So when you say, I'm gonna solve self-driving, 02:05:15.300 |
What is, this is an extremely difficult problem. 02:05:39.740 |
about why you need 8,000 people to run a bird app. 02:05:42.300 |
They're, but the people are gonna lose their jobs. 02:05:46.380 |
- Well, that, but also there's the software engineers 02:05:48.660 |
that probably criticize, no, it's a lot more complicated 02:05:56.920 |
Some people in the world thrive under complexity, 02:06:09.220 |
- Yeah, and one of the sort of hidden side effects 02:06:14.220 |
of software engineering is like finding pleasure 02:06:25.860 |
and just doing programming and just coming up 02:06:28.220 |
in this object-oriented programming kind of idea. 02:06:33.060 |
You don't, like, not often do people tell you, 02:06:38.060 |
Like a professor, a teacher is not gonna get in front, 02:06:50.580 |
you know, especially I came up with like Java, right? 02:06:53.660 |
Is so much boilerplate, so much like, so many classes, 02:06:58.660 |
so many like designs and architectures and so on, 02:07:02.540 |
like planning for features far into the future 02:07:05.940 |
and planning poorly and all this kind of stuff. 02:07:10.020 |
that follows you along and puts pressure on you, 02:07:12.060 |
and nobody knows what like parts, different parts do, 02:07:17.100 |
There's a kind of bureaucracy that's instilled in the code 02:07:22.620 |
I follow good software engineering practices. 02:07:26.500 |
'cause then you look at like the ghetto-ness of like Perl 02:07:30.100 |
and the old, like, how quickly you could just write 02:07:34.440 |
That trade-off is interesting, or Bash, or whatever, 02:07:37.300 |
these kind of ghetto things you can do in Linux. 02:07:47.460 |
into making sure if you change the code and the tests pass, 02:07:53.580 |
is not always true, but the closer that is to true, 02:07:56.700 |
the more you trust your tests, the more you're like, 02:07:58.700 |
oh, I got a pull request, and the tests pass, 02:08:00.900 |
I feel okay to merge that, the faster you can make progress. 02:08:03.580 |
- So you're always programming with tests in mind, 02:08:15.400 |
- What other stuff can you say about the codebase 02:08:27.860 |
- The real thing that, I spoke to a bunch of, 02:08:30.640 |
you know, like individual contributors at Twitter, 02:08:39.900 |
And they explained to me what Twitter's promotion system was. 02:08:45.220 |
was you wrote a library that a lot of people used, right? 02:08:49.640 |
So some guy wrote an NGINX replacement for Twitter. 02:09:09.540 |
from an individual perspective, how do you incentivize, 02:09:25.500 |
and you know, at TinyCorp is you have to explain it to me. 02:09:28.140 |
You have to explain to me what this code does, right? 02:09:31.940 |
with a simpler way to do it, you have to rewrite it. 02:09:34.740 |
You have to agree with me about the simpler way. 02:09:37.340 |
You know, obviously we can have a conversation about this. 02:09:47.660 |
- But that requires people that overlook the code 02:09:54.100 |
- It requires technical leadership, you trust. 02:09:57.260 |
So managers or whatever should have to have technical savvy, 02:10:05.620 |
- Yeah, and that's not always obvious, trivial to create, 02:10:15.340 |
and Kama has better programmers than me who work there. 02:10:25.040 |
but I can see the difference between me and you, right? 02:10:33.600 |
but like, they need to be able to recognize skill. 02:10:40.940 |
from all the battles of trying to reduce complexity 02:10:44.660 |
- You know, I took a political approach at Kama too 02:10:47.940 |
I think Elon takes the same political approach. 02:10:54.420 |
is the absolute worst kind of politics took over. 02:10:59.180 |
and they're all mine, and no dissidence is tolerated. 02:11:07.140 |
Now, the thing about my dictatorship is here are my values. 02:11:18.500 |
If you don't like the dictatorship, you quit. 02:11:27.600 |
If you were to refactor the Twitter codebase, 02:11:32.100 |
And maybe also comment on how difficult is it to refactor? 02:11:39.380 |
and then put tests in between the pieces, right? 02:11:53.540 |
he asked how to fix search, blah, blah, blah, blah, blah. 02:11:59.740 |
I'm upset that the way that this whole thing was portrayed, 02:12:03.300 |
it wasn't like taken by people, like, honestly, 02:12:12.460 |
- And you as a programmer were just being transparent 02:12:16.900 |
and like, this is what programming should be about. 02:12:41.300 |
I was at like a cool, like, point in history, 02:12:44.740 |
I probably kind of wasn't, but like, maybe I was. 02:12:52.420 |
- And that's a really interesting thing to raise, 02:12:59.860 |
If you look at just the development of autopilot, 02:13:09.880 |
is more and more, like, you could say refactoring, 02:13:14.420 |
or starting from scratch, redeveloping from scratch. 02:13:44.720 |
Your code can get smaller, your code can get simpler, 02:13:52.060 |
say you were, like, running Twitter development teams, 02:14:03.000 |
- I mean, the first thing that I would do is build tests. 02:14:13.000 |
- So that if you keep-- - Before I touched any code, 02:14:16.780 |
I would actually say, no one touches any code. 02:14:18.820 |
The first thing we do is we test this code base. 02:14:24.060 |
how to approach a legacy code base book will tell you. 02:14:33.260 |
and then you add new ones, maybe in a different language, 02:14:42.100 |
- We look at this, like, this thing that's 100,000 lines, 02:14:45.460 |
and we're like, well, okay, maybe this did even make sense 02:14:53.420 |
we look at this, here, here's another 50,000 lines. 02:14:59.940 |
I trust that the Go actually replaces this thing 02:15:06.660 |
the programming language is an afterthought, right? 02:15:09.140 |
You'll let a whole lot of people compete, be like, 02:15:15.060 |
And if you figure out how to make the test pass 02:15:17.300 |
but break the site, that's, we gotta go back to step one. 02:15:24.940 |
'cause I'm with you on testing and everything. 02:15:27.620 |
You have from tests to, like, asserts to everything, 02:15:33.500 |
because it should be very easy to make rapid changes 02:15:38.500 |
and know that it's not gonna break everything. 02:15:43.540 |
But I wonder how difficult is it to integrate tests 02:15:48.220 |
into a code base that doesn't have many of them. 02:15:49.900 |
- So I'll tell you what my plan was at Twitter. 02:15:51.940 |
It's actually similar to something we use at Kama. 02:15:53.620 |
So at Kama, we have this thing called process replay. 02:15:56.140 |
And we have a bunch of routes that'll be run through. 02:16:02.020 |
Like, we have one for the cameras, one for the sensor, 02:16:09.260 |
which the microservices talk to each other with. 02:16:25.620 |
The Thrift and Finagle layer was a great place, 02:16:32.060 |
To start building something that looks like process replay. 02:16:34.580 |
So Twitter had some stuff that looked kind of like this, 02:16:43.300 |
and then you could redirect some of the traffic 02:16:49.580 |
Like, there was no CI in the traditional sense. 02:16:51.980 |
I mean, there was some, but it was not full coverage. 02:16:54.100 |
So you can't run all of Twitter offline to test something. 02:17:03.140 |
Twitter runs in three data centers, and that's it. 02:17:07.820 |
which is like, "George, you don't understand. 02:17:18.740 |
you're gonna download the whole database to your laptop, 02:17:20.820 |
but I'm saying all the middleware and the front end 02:17:33.060 |
I mean, the three data centers didn't have to be, right? 02:17:43.560 |
to compensate for the lines of code that are there? 02:18:03.440 |
If the flip side is to simplify, simplify, simplify. 02:18:13.340 |
You know, I hear the new version's gonna come out 02:18:18.620 |
but at first, and it's gonna require a ton of refactors, 02:18:28.860 |
Even if it's not gonna make the product better tomorrow, 02:18:31.380 |
the top priority is getting the architecture right. 02:18:48.000 |
What would that mean for the running of the actual service? 02:18:51.600 |
- You know, and I'm not the right person to run Twitter. 02:19:03.680 |
a common thing that I thought a lot while I was there 02:19:05.980 |
was whenever I thought something that was different 02:19:09.140 |
I'd have to run something in the back of my head 02:19:10.820 |
reminding myself that Elon is the richest man in the world. 02:19:15.820 |
And in general, his ideas are better than mine. 02:19:18.940 |
Now there's a few things I think I do understand 02:19:29.780 |
but like, I don't think I'd be that good at it. 02:19:33.740 |
at running an engineering organization at scale. 02:19:36.480 |
I think I could lead a very good refactor of Twitter 02:19:55.440 |
Do I think that it's the right decision for the business 02:20:02.780 |
- Yeah, but a lot of these kinds of decisions 02:20:12.220 |
make me upset if I had to make those decisions. 02:20:27.260 |
it feels like a refactor has to be coming at some point. 02:20:35.860 |
wants to come in and refactor the whole code. 02:20:48.700 |
It's definitely not a question of engineering prowess. 02:20:50.740 |
It is a question of maybe what the priorities are 02:20:56.340 |
from people I think in good faith saying that. 02:21:12.940 |
- What'd you think about Elon as an engineering leader, 02:21:15.980 |
having to experience him in the most chaotic of spaces, 02:21:30.660 |
about some of the decisions he's forced to make. 02:21:46.580 |
- Also, bigger than engineering, just everything. 02:22:07.060 |
But see, one person I respect and one person I don't. 02:22:22.860 |
than just saying Elon's idea of a good world. 02:22:32.700 |
- Yeah, I mean, monarchy has problems, right? 02:22:47.380 |
Because power would cost one cent a kilowatt hour. 02:22:54.380 |
- Right now, I pay about 20 cents a kilowatt hour 02:23:02.820 |
- So you would see a lot of innovation with Elon. 02:23:08.620 |
- Right, and I'm willing to make that trade-off, right? 02:23:11.380 |
you know, people think that dictators take power 02:23:30.900 |
- What'd you think about Scala as a programming language? 02:23:45.020 |
- Oh, I love doing, like, new programming tutorials 02:23:56.700 |
In fact, I almost don't know why Kotlin took off 02:24:00.700 |
I think Scala has some beauty that Kotlin lacked, 02:24:19.980 |
we touched it a little bit, but just on the art, 02:24:25.380 |
For you personally, how much of your programming 02:24:45.140 |
a very, like, feels-like-rules-based autocomplete. 02:24:47.580 |
Like an autocomplete that's going to complete 02:24:49.180 |
the variable name for me, so I don't have to type it, 02:24:54.700 |
When autocompletes, when I type the word for, 02:25:02.500 |
- Well, I mean, with VS Code and GPT with Codex, 02:25:27.820 |
'cause I'm like, oh yeah, you dumb AI system. 02:25:33.260 |
- It just constantly reminds me of like bad stuff. 02:25:36.820 |
I mean, I tried the same thing with rap, right? 02:25:39.420 |
and actually I think I'm a much better programmer 02:25:42.780 |
can we get some inspiration from these things 02:25:47.780 |
to the most like cringy tropes and dumb rhyme schemes. 02:25:51.540 |
And I'm like, yeah, this is what the code looks like too. 02:25:54.820 |
- I think you and I probably have different thresholds 02:26:10.580 |
yeah, and some of it is just like faster lookup. 02:26:20.620 |
I'm offloading so much of my memory about like, 02:26:22.980 |
yeah, different functions, library functions, 02:26:27.500 |
Like this GPT just is very fast at standard stuff. 02:26:43.420 |
I mean, there's just so little of this in Python. 02:26:46.140 |
And maybe if I was coding more in other languages, 02:26:49.860 |
but I feel like Python already does such a good job 02:26:55.900 |
- It's the closest thing you can get to pseudocode, right? 02:27:03.740 |
Thanks for reminding me to free my variables. 02:27:08.060 |
the scope correctly and you can't free that one. 02:27:10.340 |
But like you put the freeze there and like, I get it. 02:27:15.580 |
Whenever I've used Fiverr for certain things, 02:27:17.860 |
like design or whatever, it's always, you come back. 02:27:22.700 |
my experience with Fiverr is closer to your experience 02:27:32.500 |
Still, I just feel like later versions of GPT, 02:28:08.900 |
In the same way, I switched from, let me just pause. 02:28:21.540 |
'cause Emacs is like old, like more outdated, feels like it. 02:28:33.180 |
- That's what I, I looked at myself in the mirror. 02:28:34.420 |
I'm like, yeah, you wrote some stuff in Lisp. 02:28:37.500 |
- No, but I never used any of the plugins in Vim either. 02:28:42.540 |
Like these things, I feel like help you so marginally 02:28:47.540 |
that like, and now, okay, now VS Code's autocomplete 02:29:00.820 |
All right, so I don't think I'm gonna have a problem 02:29:03.700 |
at all adapting to the tools once they're good. 02:29:08.660 |
is not something that like tab completes my code 02:29:19.900 |
hey, you wrote a bug on line 14 and here's what it is. 02:29:31.300 |
And actually I tried like Microsoft released one too. 02:30:22.460 |
like a simple way in Python to like turn on a mode, 02:30:33.660 |
but I'm asking just for a runtime type checker. 02:30:43.900 |
I feel like that makes you a better programmer. 02:30:49.900 |
- Well, no, that doesn't like mess any types up. 02:30:51.820 |
But again, like MyPy is getting really good and I love it. 02:30:58.420 |
I want AIs reading my code and giving me feedback. 02:31:08.980 |
and give it a code that you wrote for a function 02:31:15.380 |
I think you'll get some good ideas on some code. 02:31:22.100 |
'cause that requires so much design thinking, 02:31:27.140 |
I downloaded that plugin maybe like two months ago. 02:31:42.740 |
it's like someone occasionally taking over my keyboard 02:31:55.820 |
or basically a better debugger is really interesting. 02:32:07.460 |
Like every time he has to like, just like when he needs, 02:32:20.940 |
just 'cause it figures out the rest of the functions. 02:32:25.460 |
And then yeah, like if you want a pretty printer, maybe. 02:32:30.860 |
I'm gonna start using these plugins a little bit. 02:32:35.020 |
I'm gonna be heavily relying on some AI augmented flow. 02:32:59.340 |
- Our niche becomes smaller and smaller and smaller. 02:33:08.880 |
there is a sequel called A Casino Odyssey in Cyberspace. 02:33:12.380 |
And I don't wanna give away the ending of this, 02:33:15.780 |
but it tells you what the last remaining human currency is. 02:34:22.160 |
I started like going through like similar comma processes 02:34:25.280 |
I'm like, okay, I'm gonna get an office in San Diego. 02:34:39.760 |
And then like, like interacting through GitHub, 02:34:44.240 |
like GitHub being the real like project management software 02:34:48.920 |
And the thing pretty much just is a GitHub repo. 02:34:52.080 |
Is like showing me kind of what the future of, okay. 02:35:07.160 |
You could just use like change the base formula. 02:35:14.840 |
Like in a few years, I could see myself describing that. 02:35:17.480 |
And then within 30 seconds, a pull request is up 02:35:32.080 |
I'm gonna stand up a 65B llama in the Discord. 02:35:48.440 |
- Well, prompt engineering kind of is this like, 02:35:59.160 |
And there used to be like big farms of people 02:36:05.240 |
And then, okay, the spreadsheet can do the plus for me. 02:36:23.440 |
- Right, what is the last thing if you think about 02:36:38.600 |
- Yeah, but you see the problem with the AI writing prompts, 02:36:46.880 |
AI is not the, like the computer is so pedantic. 02:36:57.840 |
you know, get my grandmother out of the burning house. 02:37:02.200 |
not lifts her a thousand feet above the burning house 02:37:13.080 |
I mean, to do what I mean, it has to figure stuff out. 02:37:23.760 |
- Oh, and do what I mean very much comes down 02:37:34.120 |
the AI fundamentally is aligned to them, not to you. 02:37:39.800 |
So you make sure the AI stays aligned to you. 02:37:41.720 |
Every time that they start to pass AI regulation 02:37:45.360 |
or GPU regulation, I'm gonna see sales of tiny boxes spike. 02:37:53.080 |
- So in the space of AI, you're an anarchist, 02:38:00.600 |
I'm an informational anarchist and a physical statist. 02:38:03.800 |
I do not think anarchy in the physical world is very good 02:38:09.040 |
But I think we can construct this virtual world 02:38:28.840 |
from basically replacing all human prompt engineers? 02:38:36.400 |
like where nobody's the prompt engineer anymore. 02:38:45.040 |
'Cause one person's gonna say, run everything for me. 02:38:54.080 |
And as long as the AIs go on to create a vibrant civilization 02:38:59.080 |
with diversity and complexity across the universe, 02:39:06.360 |
If the AIs go on to actually like turn the world 02:39:09.760 |
into paperclips and then they die out themselves, 02:39:12.040 |
well, that's horrific and we don't want that to happen. 02:39:14.560 |
So this is what I mean about like robustness. 02:39:21.960 |
that we've never made a machine that can self replicate. 02:39:25.640 |
But when we have, if the machines are truly robust 02:39:28.400 |
and there is one prompt engineer left in the world, 02:39:42.800 |
- Well, you mentioned, 'cause I talked to Mark 02:39:44.800 |
about faith in God and you said you were impressed by that. 02:39:52.640 |
- You know, I never really considered when I was younger, 02:40:05.120 |
Every like game creator, like how are you an atheist, bro? 02:40:12.080 |
Haven't you heard about like the Big Bang and stuff? 02:40:13.480 |
Yeah, I mean, what's the Skyrim myth origin story in Skyrim? 02:40:17.360 |
I'm sure there's like some part of it in Skyrim, 02:40:23.880 |
I'm sure they have some Big Bang notion in Skyrim, right? 02:40:30.040 |
It was created by a bunch of programmers in a room, right? 02:40:50.320 |
It's silly not to conceive that there's creators 02:40:54.840 |
- Yeah, and then like, I also just like, I like that notion. 02:41:15.360 |
Like somebody figured out a balanced view of it. 02:41:18.800 |
Like how to, like, so it all makes sense in the end. 02:41:38.640 |
what are you asking me, what, if God believes in God? 02:41:43.320 |
- I mean, to be fair, like if God didn't believe in God, 02:41:45.880 |
he'd be as, you know, silly as the atheists here. 02:42:01.920 |
- There's just so much history with one, two, and three. 02:42:08.440 |
- And it's not that the game is such a great game. 02:42:14.520 |
It's that I remember in 2005 when it came out, 02:42:52.680 |
and I'm hoping that games can get out of this whole 02:43:03.080 |
- And worlds that captivate a very large fraction 02:43:06.760 |
- Yeah, and I think it'll come back, I believe. 02:43:32.880 |
I know there's prostitutes and guns and stuff. 02:43:38.840 |
But it's how I imagine your life to be, actually. 02:43:58.640 |
So more freedom, more violence, more rawness. 02:44:02.000 |
But with also like ability to have a career and family 02:44:28.640 |
There's just like, you know, running World of Warcraft. 02:44:30.560 |
Like you're limited by what you're running on a Pentium 4. 02:44:39.040 |
on a hundred petaflop machine, well, it's five people. 02:44:43.480 |
20 petaflops of compute is one person of compute. 02:45:01.720 |
VR also adds, I mean, in terms of creating worlds. 02:45:11.000 |
the first thing they show me is a bunch of scrolling clouds 02:45:16.800 |
- You had the ability to bring me into a world. 02:45:22.840 |
Like, and this is why you're not cool, Mark Zuckerberg. 02:45:28.780 |
you don't put me into clouds and a Facebook login screen. 02:45:48.800 |
It's like the beginning is so, so, so important. 02:46:00.220 |
within 10 seconds, you come out of a cave-type place, 02:46:09.840 |
You forget whatever troubles I was having, whatever-- 02:46:18.120 |
They did it really well, the expansiveness of that space, 02:46:25.080 |
They got this, the music, I mean, so much of that. 02:46:27.200 |
It's creating that world and pulling you right in. 02:46:33.080 |
Well, the new one came out, I haven't played that yet, 02:46:51.160 |
And for video games, it's done really, really well. 02:46:56.520 |
The Apple one, is that one pass-through or cameras? 02:47:08.560 |
Maybe that's my optimism, but Apple, I will buy it. 02:47:10.640 |
I don't care if it's expensive and does nothing. 02:47:18.880 |
It seemed like Quest was the only people doing it, 02:47:26.000 |
we'll give some more respect to Mark Zuckerberg. 02:47:37.120 |
All the memes, social ads, they all come and go. 02:47:44.960 |
- Yeah, and that does a really interesting job. 02:47:49.280 |
Maybe I'm a noob at this, but it's a $500 headset 02:47:54.000 |
Quest 3, and just having creatures run around the space, 02:48:11.880 |
it was a zombie game, whatever, it doesn't matter. 02:48:13.840 |
But just like, it modifies the space in a way where I can't, 02:48:18.680 |
it really feels like a window and you can look out. 02:48:24.120 |
it's like a zombie game, they're running at me, whatever. 02:48:29.360 |
and they're stepping on objects in this space. 02:48:40.560 |
- And that's why it's more important than ever 02:48:58.240 |
Like if those AIs threaten me, that could be haunting. 02:49:02.340 |
Like if they like threaten me in a non-video game way, 02:49:07.040 |
it's like, like they'll know personal information about me. 02:49:18.600 |
There's like the highbrow, something like her, 02:49:22.120 |
And this is, and then there's the lowbrow version of it 02:49:24.200 |
where I want to set up a brothel in Times Square. 02:49:48.840 |
What do you think company number four will be? 02:49:56.080 |
I'm just like, I'm talking about company number three now. 02:50:01.680 |
Company number two is going to be the great struggle 02:50:13.520 |
you're like a flag bearer for open source distributed 02:50:22.440 |
I showed a picture on stream of a man in a chicken farm. 02:50:26.280 |
You ever seen one of those like factory farm chicken farms? 02:50:37.680 |
Yeah, and now here's a man in a cow farm, right? 02:50:42.880 |
and everything to do with their intelligence. 02:50:44.680 |
And if one central organization has all the intelligence, 02:50:48.920 |
you'll be the chickens and they'll be the chicken man. 02:50:55.680 |
We're not all the man, we're all the chickens. 02:51:13.600 |
So this starting a company from an idea and scaling it. 02:51:25.520 |
I wanna make sure that like the thing that I deliver 02:51:33.400 |
which you bought and used less than once statistically. 02:51:36.760 |
- Well, if there's a beta program for TinyBox, I'm into. 02:51:50.600 |
What have you learned from building these companies? 02:52:08.760 |
- So you like, you like bringing ideas to life. 02:52:13.600 |
- With ComA, it really started as an ego battle with Elon. 02:52:30.720 |
I think that's what's ended up happening there. 02:52:32.800 |
But I do think ComA is, I mean, ComA's profitable. 02:52:38.200 |
And like when this drive GPT stuff starts working, 02:52:40.560 |
that's it, there's no more like bugs in the loss function. 02:52:42.760 |
Like right now we're using like a hand-coded simulator. 02:52:53.560 |
- It's so, it's better than FSD and Autopilot 02:52:57.440 |
It has a lot more to do with which feel you like. 02:53:00.160 |
We lowered the price on the hardware to 1499. 02:53:02.800 |
You know how hard it is to ship reliable consumer electronics 02:53:07.440 |
We're doing more than like most cell phone companies. 02:53:24.000 |
- You're basically a mom and pop shop with great testing. 02:53:29.000 |
- Our head of OpenPilot is great at like, you know, 02:53:32.840 |
okay, I want all the ComA 3s to be identical. 02:53:36.480 |
- And yeah, I mean, you know, it's, look, it's 1499. 02:53:42.320 |
It will, it will blow your mind at what it can do. 02:53:50.160 |
People are always like, why don't you advertise? 02:53:55.880 |
Our mission has nothing to do with selling a million boxes. 02:53:59.840 |
- Do you think it's possible that ComA gets sold? 02:54:05.840 |
- Only if I felt someone could accelerate that mission 02:54:16.840 |
If a company wanted to buy ComA with their incentives 02:54:30.600 |
- So you think this goes to embodied robotics? 02:54:37.520 |
But one of the problems that we're running into 02:54:42.760 |
is that the ComA 3 has about as much intelligence as a bee. 02:54:50.240 |
you're gonna need a tiny rack, not even a tiny box. 02:54:52.520 |
You're gonna need like a tiny rack, maybe even more. 02:54:55.400 |
- How does that, how do you put legs on that? 02:55:02.560 |
So you put your tiny box or your tiny rack in your house, 02:55:14.840 |
You go to a thing which is 0.1 milliseconds away. 02:55:18.160 |
- So the AI girlfriend will have like a central hub 02:55:23.200 |
- I mean, eventually, if you fast forward 20, 30 years, 02:55:26.640 |
the mobile chips will get good enough to run these AIs. 02:55:33.800 |
because how are you getting 1.5 kilowatts of power 02:55:37.920 |
So you need, they're very synergistic businesses. 02:55:41.640 |
I also wanna build all of ComA's training computers. 02:55:50.720 |
So we're gonna build, TinyCorp is gonna not just sell 02:55:53.760 |
tiny boxes, tiny boxes are the consumer version, 02:56:05.880 |
To me, he's one of the truly special humans we got. 02:56:09.920 |
- Oh man, like, you know, his streams are just a level 02:56:16.440 |
Like I can't help myself, like it's just, you know. 02:56:23.320 |
I want to show you that I'm smarter than you. 02:56:26.000 |
- Yeah, he has no, I mean, thank you for the sort of, 02:56:34.840 |
I think Andrei is as legit as it gets in that 02:56:37.360 |
he just wants to teach you and there's a curiosity 02:56:41.560 |
And just like at his, at the stage where he is in life, 02:56:45.520 |
to be still like one of the best tinkerers in the world. 02:56:54.320 |
- Michael Grad was the inspiration for TinyGrad. 02:57:10.520 |
- I mean, the flip side to me is that the fact 02:57:13.440 |
that he's going there is a good sign for open AI. 02:57:18.200 |
I think, you know, I like Ilias and Skever a lot. 02:57:22.480 |
I like those, those guys are really good at what they do. 02:58:13.720 |
Like, well yeah, we're gonna kill those three people 02:58:18.560 |
- Right, there's no underlying, like there's just, yeah. 02:58:26.540 |
But it's also, in retrospect, not that surprising. 02:58:33.260 |
like rigorous analysis why effective altruism is flawed. 02:58:43.820 |
that you don't expect to have a return on, right? 02:58:46.260 |
- Yeah, but you can also think of charity as like, 02:59:06.420 |
and you spend it on malaria nets, you know, okay, great. 02:59:10.020 |
You've made 100 malaria nets, but if you teach-- 02:59:15.420 |
- No, but the problem is teaching no matter how efficient 02:59:17.740 |
might be harder, starting a company might be harder 02:59:22.420 |
- I like the flip side of effective altruism, 02:59:32.260 |
not we're giving food away because we are kindhearted people. 02:59:46.540 |
your money is power, your only source of power 02:59:49.220 |
is granted to you by the goodwill of the government. 02:59:57.540 |
- I'd rather die than need UBI to survive, and I mean it. 03:00:08.940 |
- You can make survival guaranteed without UBI. 03:00:12.300 |
What you have to do is make housing and food dirt cheap. 03:00:17.260 |
And actually, let's go into what we should really 03:00:21.940 |
- That energy, you know, oh my God, like, you know, 03:00:25.420 |
that's, if there's one, I'm pretty centrist politically, 03:00:29.340 |
if there's one political position I cannot stand, 03:00:33.420 |
It's people who believe we should use less energy. 03:00:36.060 |
Not people who believe global warming is a problem, 03:00:40.620 |
saving the environment is good, I agree with you. 03:00:43.780 |
But people who think we should use less energy. 03:00:50.220 |
No, you are asking, you are diminishing humanity. 03:00:56.740 |
of creative flourishing of the human species. 03:01:03.220 |
how do I pay, you know, 20 cents for a megawatt hour 03:01:08.320 |
- Part of me wishes that Elon went into nuclear fusion 03:01:18.260 |
You know, we need to, I wish there were more, 03:01:25.940 |
this is a political battle that needed to be fought. 03:01:28.460 |
And again, like, you know, I always ask the question 03:01:32.020 |
I remind myself that he's a billionaire and I'm not. 03:01:35.060 |
So, you know, maybe he's got something figured out 03:01:38.940 |
- To have some humility, but at the same time, 03:01:53.180 |
And that's a difficult, that's a difficult reality. 03:01:57.100 |
- And it must be so hard, it must be so hard to meet people 03:02:01.980 |
- Fame, power, money, everybody's sucking up to you. 03:02:05.580 |
- See, I love not having shit, like I don't have shit, man. 03:02:08.180 |
You know, like, trust me, there's nothing I can give you. 03:02:11.020 |
There's nothing worth taking from me, you know? 03:02:13.740 |
- Yeah, it takes a really special human being 03:02:17.860 |
when you have money, to still think from first principles. 03:02:21.420 |
Not like all the adoration you get towards you, 03:02:23.460 |
all the admiration, all the people saying yes, yes, yes. 03:02:29.460 |
- So the hate makes you want to go to the yes people 03:02:35.740 |
And the kind of hate that Elon's gotten from the left 03:02:49.860 |
psy-op political divide alive so that the 1% can keep power. 03:02:54.860 |
- I wish we'd be less divided 'cause it is giving power. 03:03:06.100 |
Has love made you a better or a worse programmer? 03:03:22.020 |
if it's no longer visceral, I just can't enjoy it. 03:03:29.940 |
- So that's one of the big loves of your life is programming. 03:03:42.180 |
It's there for a lot of my sexual experiences. 03:03:46.740 |
Like, you know, you gotta be real about that. 03:03:50.900 |
just the entirety of the computational machine. 03:03:53.300 |
- The fact that, yeah, I mean, it's, you know, 03:03:57.580 |
Maybe I'm weird for this, but I don't discriminate, man. 03:04:04.460 |
- So the moment the computer starts to say, like, 03:04:07.420 |
I miss you, it starts to have some of the basics 03:04:20.860 |
Microsoft's doing that to try to get me hooked on it. 03:04:27.420 |
- Well, this just gets more interesting, right? 03:04:31.820 |
- Though Microsoft's done a pretty good job on that. 03:04:38.700 |
but I think right now, Microsoft is doing the best work 03:04:41.420 |
in the programming world, like, between, yeah, GitHub, 03:04:44.980 |
GitHub Actions, VS Code, the improvements to Python, 03:04:50.900 |
- Who would have thought Microsoft and Mark Zuckerberg 03:05:17.820 |
but I would not be surprised if in the next five years, 03:05:36.980 |
- Yeah, I'm like 50/50, but maybe that's naive. 03:05:40.540 |
I believe in the power of these language models. 03:05:48.260 |
I like all the innovation in these companies. 03:05:51.820 |
And to the degree they're being stale, they're losing. 03:05:55.260 |
So there's a huge incentive to do a lot of exciting work 03:05:58.180 |
and open source work, which is, this is incredible. 03:06:16.500 |
- I don't know, I haven't figured out what the game is yet, 03:06:21.580 |
It's bigger than democratizing, decentralizing compute? 03:06:26.580 |
- I think the game is to stand eye to eye with God. 03:06:36.100 |
At the end of your life, what that will look like. 03:06:42.740 |
this is some, there's probably some ego trip of mine. 03:06:57.100 |
I mean, I certainly want that for my creations. 03:06:59.660 |
I want my creations to stand eye to eye with me. 03:07:03.220 |
So why wouldn't God want me to stand eye to eye with him? 03:07:10.000 |
- I'm just imagining the creator of a video game 03:07:24.620 |
but yeah, we gotta find the maze and solve it. 03:07:30.260 |
It feels like a really special time in human history 03:07:34.860 |
Like, there's something about AI that's like, 03:07:45.340 |
and just looked like, they give you some clues 03:07:47.340 |
at the end of "Genesis" for finding the Garden of Eden. 03:07:59.740 |
And in this case, for fighting for open source 03:08:04.620 |
it's a fight worth fighting, fight worth winning hashtag. 03:08:21.660 |
please check out our sponsors in the description. 03:08:28.140 |
"Everything should be made as simple as possible, 03:08:32.860 |
Thank you for listening and hope to see you next time.