back to indexWhy Google failed to make GPT-3 -- with David Luan of Adept
Chapters
0:0 Introduction of David Luan, CEO and co-founder of Adept
1:14 David's background and career trajectory
3:20 Transition from reinforcement learning to transformers in the AI industry
5:35 History and development of GPT models at OpenAI and Google
13:8 Adept's $420 million funding rounds
13:38 Explanation of what Adept does and their vision for AI agents
19:20 Reasons for Adept becoming more public-facing
21:0 Adept's critical path and research directions (Persimmon, Fuyu, Act One)
26:23 How AI agents should interact with software and impact product development
30:37 Analogies between AI agents and self-driving car development
32:42 Balancing reliability, cost, speed and generality in AI agents
35:11 Adept's unique positioning and advantages in the AI industry
37:30 Potential of foundation models for robotics
39:22 Core research questions and reasons to work at Adept
40:57 David's closing thoughts on the AI agent space and industrialization of AI
00:00:07.580 |
♪ Exploring the world, the guts where new begins begin ♪ 00:00:32.040 |
- Hey everyone, welcome to the Latent Space Podcast. 00:00:42.840 |
CEO, co-founder of ADEPT in the studio, welcome. 00:00:47.340 |
I met you socially at one of those VC events, 00:00:50.560 |
and you said that you were interested in coming on, 00:00:52.640 |
and glad we finally were able to make this happen. 00:01:00.360 |
and then also just have you talk a little bit 00:01:03.800 |
what people should just generally know about you. 00:01:10.800 |
video detection classification API that was Dextro, 00:01:15.480 |
and that was your route to getting acquired into Axon, 00:01:27.520 |
VP of Eng for two and a half years, two years and a bit, 00:01:31.520 |
briefly served as tech lead of large models at Google, 00:01:47.440 |
- I guess a broader story was joined OpenAI fairly early, 00:01:53.840 |
two and a half to three years leading engineering there. 00:01:56.520 |
It's really funny, I think second or third day 00:01:59.720 |
of my time at OpenAI, Greg and Ilya pulled me in a room 00:02:03.920 |
and were like, "Hey, you should take over our direction. 00:02:12.640 |
So that was fun, just coalescing a bunch of teams 00:02:19.320 |
The company, the Dota effort was going pretty hard, 00:02:28.660 |
And then at Google, so I led Google's LLM efforts, 00:02:36.480 |
And I think there's been a couple of different eras 00:02:40.600 |
And if we count everything before 2012 as prehistory, 00:02:46.040 |
you kinda had this like you and your three best friends 00:02:48.040 |
write a research paper that changes the world period 00:02:52.640 |
And then from, and I think the game changed in 2017, 00:03:02.440 |
that the world would be covered in data centers, 00:03:06.360 |
- Yeah, well, like I think we had conviction in that, 00:03:08.640 |
but it wasn't until we started seeing results 00:03:10.360 |
that it became clear that that was where we had to go. 00:03:12.600 |
But also part of it as well was like for OpenAI, 00:03:21.040 |
compared to, hey, we're just smaller Google brain, 00:03:29.120 |
That's like not enough to like hang your technical identity 00:03:38.360 |
a certain class of like giant swings and bets, right? 00:03:44.460 |
you just do bottom-up research to more about like, 00:03:50.100 |
what are the big scientific outcomes that you wanna show? 00:03:55.360 |
whether or not you care about novelty and all that stuff. 00:03:57.680 |
And that became the dominant model for a couple years, right? 00:04:00.740 |
And then what's changed now is I think that like 00:04:07.880 |
is gonna be the deep co-design and co-evolution 00:04:17.200 |
And that's a big part of why I started ADAPT. 00:04:20.200 |
Any memories thinking from like the switch from RL 00:04:23.840 |
to Transformers at the time and kind of how the industry 00:04:29.800 |
and leaving behind some of the more agent simulation work? 00:05:01.900 |
is a definition of AGI that's oriented around 00:05:27.000 |
doesn't come with that out of the box, right? 00:05:33.460 |
hey, how do we solve problems of that caliber? 00:05:37.840 |
like the Novo RL is like a pretty terrible way 00:05:48.180 |
like what will it actually take to build AGI? 00:05:50.480 |
And his view is basically that you have to reproduce 00:06:10.000 |
So like today, maybe LLMs is like behavioral cloning 00:06:12.720 |
every word that gets written on the internet. 00:06:14.920 |
In the future, you know, like now the multimodal models 00:06:23.560 |
Where like tokens of data that have high signal come in, 00:06:27.320 |
and then all of those patterns are like learned by the model 00:06:30.280 |
and then you can regurgitate any combination out, right? 00:06:36.040 |
like to other image out or video out or whatever, 00:06:44.960 |
And I think now we're back to the era of like, 00:06:48.440 |
how do we combine this with all of the lessons we learned 00:07:02.360 |
'cause it's really nice, like personal, you know, 00:07:19.600 |
- Right around that? - I was right around that, yeah. 00:07:23.840 |
you know, just kind of came in and was like very obsessed 00:07:41.860 |
So I think the real story of GPT starts at Google, 00:07:46.340 |
Because that's where transformers sort of came about. 00:07:49.780 |
The number one shocking thing to me was that, 00:07:56.260 |
like you and your three best friends write papers, right? 00:07:59.940 |
I think about my job when I was a full-time research leader 00:08:03.220 |
as a little bit of a portfolio allocator, right? 00:08:14.280 |
My job is not actually to promote a million ideas 00:08:21.600 |
my job is to nudge resources towards the things 00:08:24.600 |
that are really working and then start disbanding 00:08:27.160 |
some of the things that are not working, right? 00:08:29.200 |
That muscle did not exist during my time at Google. 00:08:33.240 |
And I think had they had it, what they would have done 00:08:35.400 |
would be say, hey, Noam Shazir, you're a brilliant guy, 00:08:42.480 |
And then I think they would have destroyed us. 00:08:45.800 |
He's talking about trillion parameter models in 2017. 00:08:50.840 |
Which is that, and I'm jumping around historically, right? 00:08:53.440 |
But like, after GPT-2, we were all really excited 00:08:55.800 |
about GPT-2, I can tell you more stories about that. 00:08:58.760 |
It was the last paper that I even got to really touch 00:09:04.800 |
You know, every day we were scaling up GPT-3, 00:09:13.220 |
Google has all this compute, Google has all the people 00:09:15.880 |
who invented all of these underlying technologies. 00:09:22.500 |
about how he wants a trillion parameter model. 00:09:26.880 |
we're probably just doing duplicative research 00:09:32.600 |
that's probably gonna get there before we do. 00:09:42.440 |
So during my year where I led the Google LM effort 00:09:54.840 |
And did you guys remember the brain credit marketplace? 00:09:57.880 |
- Oh, so it's actually, you can ask any Googler, 00:10:02.360 |
- I mean, look, like yeah, limited resources, 00:10:04.840 |
you gotta have some kind of marketplace, right? 00:10:12.560 |
basically everyone's assigned a credit, right? 00:10:14.320 |
So if you have a credit, you get to buy end chips 00:10:20.820 |
you gotta convince like 19 or 20 of your colleagues 00:10:27.020 |
it's really hard to get that bottom up critical mass 00:10:31.880 |
And like, and the team at Google were fighting valiantly, 00:10:36.880 |
simply because we took big swings and we focused. 00:10:40.840 |
And I think, again, that's like part of the narrative 00:10:48.360 |
And I think in the same way, I think phase three companies 00:10:52.920 |
because of the same like asymmetry of success. 00:10:56.120 |
- Yeah, I think it's underrated how much Nvidia 00:11:17.280 |
I have so much respect for Nvidia, it is unreal. 00:11:23.480 |
or you just work with whatever Nvidia gave them. 00:11:29.120 |
There's, I'm not sure I can share all the stories, 00:11:39.200 |
He was on one of my teams, the supercomputing team, 00:11:46.360 |
But as a result, like we had very close ties to Nvidia. 00:11:50.640 |
Actually, one of my co-founders at Adapt, Eric Elson, 00:11:55.120 |
And so he and Scott and like Brian Catanzaro and Nvidia 00:11:58.640 |
and Jonah and Ian at Nvidia, I think all were very close. 00:12:02.640 |
And we're all sort of part of this group of just like, 00:12:04.240 |
how do we push these chips to the absolute limit? 00:12:12.400 |
knowing the A100 generation that like quad sparsity 00:12:15.920 |
Is that something that we wanna go look into, right? 00:12:19.400 |
that we could actually use for model training. 00:12:21.200 |
And I think more and more people realize this, 00:12:22.920 |
but like six years ago, or even three years ago, 00:12:28.040 |
Like this era of AI is really a story of compute. 00:12:30.160 |
It's really the story of how do you more efficiently map 00:12:32.760 |
like actual usable model flops to compute, right? 00:12:38.240 |
Is there another, you know, sort of GPT-2, 3 story 00:12:42.160 |
that like, you know, you love to get out there 00:12:45.040 |
that I think you think is like underappreciated 00:12:47.040 |
for like the amount of work that people put into it? 00:12:55.840 |
And I remember one of the most entertaining moments, 00:13:03.320 |
was like the shortest modeling section of any ML, 00:13:05.720 |
like reasonably legitimate ML paper to that moment. 00:13:10.160 |
like this is a standard vanilla decoder only transformer 00:13:14.960 |
It was like a paragraph long, if I remember correctly. 00:13:17.240 |
And both of us were just looking at the same, 00:13:33.520 |
where we just leaned fully into all we care about 00:13:36.080 |
is solving problems in AI and not about like, 00:13:44.160 |
that doesn't actually help move the field forward? 00:13:48.760 |
And it's like, you innovate on maybe like data set 00:13:50.880 |
and scaling and not so much the architecture. 00:13:58.680 |
there's a collection of really hard won knowledge 00:14:00.480 |
that you get only by being at the frontiers of scale. 00:14:12.120 |
But yeah, that's the stuff that helps differentiate 00:14:24.680 |
Sam Altman, myself, and our CFO flew up to Seattle 00:14:31.000 |
so I always had like a tremendous amount of anxiety 00:14:36.400 |
because it's like Kevin Scott and Satya and Amy Hood. 00:14:40.320 |
And it was my job to give the technical slides about, 00:14:44.640 |
what's our research portfolio, all of this stuff. 00:14:47.200 |
But it was also my job to give the GPT-2 demo. 00:14:56.800 |
model behaviors you find predictable at one checkpoint 00:15:01.160 |
And so like, I'd spent all this time trying to figure out 00:15:10.160 |
over to like Satya and Kevin and let them type anything in. 00:15:14.040 |
And that just, that really kept me up all night. 00:15:28.040 |
so I'm sure you do great in partners meetings. 00:15:32.080 |
- No, that's a high compliment coming from a VC. 00:15:34.360 |
- Yeah, no, I mean, you're doing great already. 00:15:46.240 |
and then have the founders fill in the blanks, 00:15:50.880 |
- Yeah, so I think ADAPT is like the least understood company 00:15:54.800 |
in the like broader space of foundation models plus agents. 00:15:58.480 |
So I'll give some color and I'll explain what it is, 00:16:02.280 |
and I'll explain also why it's actually pretty different 00:16:07.560 |
So the goal for ADAPT is we basically wanna build 00:16:11.760 |
an AI agent that can basically help humans do anything 00:16:19.480 |
we want this thing to be super good at turning 00:16:21.680 |
natural language, like goal specifications, right? 00:16:27.480 |
and then also have all the correct sensors and actuators 00:16:29.920 |
to go get that thing done for you across any software tool 00:16:33.520 |
And so the end vision of this is effectively like, 00:16:36.640 |
everyone's gonna have access to like an AI teammate 00:16:38.840 |
that they can delegate arbitrary tasks to at work, 00:16:42.560 |
and then also be able to use it as a sounding board 00:16:45.000 |
and like just be way, way, way more productive, right? 00:16:50.600 |
from something where you're mostly doing execution 00:16:52.400 |
to something where you're mostly actually doing 00:16:58.160 |
I find this like really exciting and motivating 00:16:59.760 |
because I think it's actually a pretty different vision 00:17:06.120 |
are the most likely systems to be proto-AGIs. 00:17:09.800 |
But I think the ways in which we are really counterintuitive 00:17:12.180 |
to everybody is that we've actually been really quiet 00:17:28.520 |
and like we want more users signing up for that thing. 00:17:33.040 |
we work with like a range of different companies, 00:17:36.600 |
some like late stage, like multi-thousand people startups, 00:17:42.560 |
And what we do for them is we basically give them 00:17:45.240 |
an out of the box solution where like big complex workflows 00:17:52.280 |
So we look a little different from other companies 00:17:54.360 |
in that like in order to go build this full agent thing, 00:17:57.740 |
the most important thing you gotta get right is reliability. 00:18:05.480 |
was we released this demo called Act One, right? 00:18:13.220 |
by going to Redfin and asking to buy a house somewhere. 00:18:15.960 |
'Cause like we did that in the original Act One demo 00:18:18.280 |
and like showed that, showed like Google Sheets, 00:18:22.240 |
But over the last like year since that has come out, 00:18:34.760 |
on how do we build an amazing enterprise product, 00:18:38.640 |
can't use anything that isn't in the nines of reliability. 00:18:44.280 |
than what you might find in the prompt engineering 00:18:47.240 |
sort of plays in the agent space to get that reliability. 00:18:52.240 |
And we've decided to prioritize reliability over all else. 00:18:59.400 |
being sent to a place as the result of the agent workflow. 00:19:02.880 |
And if you're like, if that works like 60% of the time, 00:19:14.640 |
I'm actually giving a talk at NVIDIA GTC about this, 00:19:19.800 |
you're wrapping user productivity in software 00:19:25.080 |
is replacing things that you would ask somebody to do 00:19:33.240 |
do the users still go in and like look at the agent 00:19:37.480 |
kind of like doing the things and can intervene 00:19:39.280 |
or like are these like fully removed from them? 00:19:41.600 |
Like the truck thing is like, does the truck just show up 00:19:43.700 |
or like are there people in the middle like checking in? 00:19:46.200 |
- Yeah, so actually what's been really interesting 00:19:47.920 |
is you could question whether they're fundamental, 00:19:55.080 |
I think that one of them is like in our experience 00:20:07.600 |
including the shitty rote stuff to I'm a supervisor 00:20:15.060 |
because like now it parallelizes a bunch of the things 00:20:17.840 |
that you had to do sequentially by hand as a human 00:20:36.320 |
and the whole trajectory is just broken and dead 00:20:39.280 |
So then those are the ones that the human then goes 00:20:41.400 |
and solves and so then they become a troubleshooter. 00:20:47.880 |
I think the second piece of it that we've found 00:21:01.120 |
but two, actually, if you're framing yourself 00:21:07.500 |
where you're solving tasks that are a little too hard 00:21:11.740 |
and still needs a human to provide oversight, 00:21:15.100 |
provide clarifications, provide human feedback 00:21:19.340 |
That's how you actually learn from the smartest humans 00:21:23.220 |
and so I actually think that being an augmentation company 00:21:25.980 |
forces you to go develop your core AI capabilities faster 00:21:45.020 |
and just they cannot attract the talent to do it 00:21:49.260 |
you have Copilot, which is the augmentation product 00:21:51.740 |
and then you have Sweep.dev, any of these products, 00:21:59.580 |
I agree that today, the reliability's so important 00:22:02.500 |
in the enterprise that they just don't use most of them. 00:22:13.180 |
they do Persimon, they do Fuyu, they do all these-- 00:22:16.660 |
- It's just public stuff and so I think you're gonna find, 00:22:19.580 |
so one of the things we haven't shared before 00:22:25.340 |
- Sold out of bandwidth to go onboard more customers. 00:22:27.740 |
I think we're like working really hard to go, 00:22:35.740 |
I think we're gonna be significantly more public 00:22:43.100 |
So I think that clarification will happen by default. 00:22:53.100 |
towards being more open or releasing more things. 00:22:56.700 |
- I think we just flipped over that way fairly recently. 00:23:01.180 |
I think it actually boils down to two things. 00:23:03.300 |
The public narrative is really forming around agents 00:23:08.540 |
because when we started the company in January, 2022, 00:23:15.340 |
But like the general public had no conception 00:23:19.660 |
on the tree of like, everything's a chatbot, right? 00:23:24.540 |
I think one of the things that I really care about 00:23:33.500 |
Things that make a function call are being called agents. 00:23:35.260 |
Like to me, an agent is something that you can give a goal 00:23:46.660 |
to be more aware of Adept as they think about 00:23:48.220 |
what the next thing they wanna do in their careers. 00:23:56.580 |
And I think a huge amount of gain is gonna happen 00:24:06.620 |
And I think people who wanna do agents research 00:24:31.580 |
- And before that even, I think there was something 00:24:39.020 |
I'm seeing AI startups that used to just brand themselves 00:24:41.140 |
as an AI company now brand themselves as an AI agent company. 00:24:49.820 |
where like, I would not touch any agent startups 00:24:56.020 |
- I think a lot of VCs that are maybe less technical 00:25:03.620 |
- No, no, I think like the, what is possible today 00:25:06.380 |
and like what is worth investing in, you know? 00:25:08.420 |
And I think like, I mean, people look at you and say, 00:25:17.260 |
"that is like tacking on AI to an existing thing, 00:25:21.440 |
"and kind of get some of the flag wheel going." 00:25:28.820 |
Like, sometimes we look around and it's like, 00:25:42.220 |
is there's a new agent and company popping up every day. 00:25:46.740 |
But like I have advised people to take agents 00:25:59.340 |
so like, you know, you're a portfolio allocator. 00:26:06.700 |
Can you take us through like how you think about 00:26:08.660 |
that evolution of that and what people should think about 00:26:11.700 |
what that means for adept sort of research directions? 00:26:14.980 |
- The critical path for adept is we want to build 00:26:22.100 |
all while keeping an insanely high reliability standard. 00:26:30.380 |
reliability standard but are continuing pushing 00:26:32.100 |
a level of abstraction, you then learn from your users 00:26:34.420 |
how to get that next level of abstraction faster. 00:26:36.180 |
So that's how you actually build the data flow. 00:26:41.780 |
So if you go zoom way, way back to Act One days, right? 00:26:57.780 |
when you give it various different workflows and text. 00:27:10.260 |
to get a lot better at having some specification 00:27:14.060 |
of some guardrails for what it actually should be doing. 00:27:22.420 |
that are really good at understanding knowledge work 00:27:29.620 |
And so like, back then we had to do a ton of research, 00:27:32.620 |
basically, on how do we actually make that possible? 00:27:46.860 |
I think one big hangover from primarily academic focus 00:28:03.020 |
Like, it's like, it's really helped the field, right? 00:28:06.980 |
I actually think like, like, it's really clear today, 00:28:09.340 |
multimodal models are the default foundation model, right? 00:28:13.300 |
Like, why would you just train a giant multimodal model? 00:28:18.020 |
like, where are they gonna be the most useful? 00:28:19.420 |
They're gonna be most useful in knowledge work tasks. 00:28:21.600 |
That's where the majority economic value is gonna be. 00:28:25.420 |
And so if that's what it is, what do you need to train? 00:28:27.580 |
I need to train on like charts, graphs, tables, invoices, 00:28:32.060 |
Like, that's just a totally different pre-training corpus. 00:28:35.380 |
And so at Depp, spent a lot of time building that. 00:28:37.900 |
And so the like, the public for use and stuff 00:28:54.540 |
So that's kind of like some of the modeling side. 00:28:56.540 |
We've kind of only announced some of that stuff. 00:28:58.060 |
We haven't really announced much of the agents work. 00:29:04.940 |
and I think the product form factor also really matters. 00:29:09.620 |
and you guys probably see this a little bit more than I do, 00:29:11.600 |
but like we're seeing like a little bit of a pushback 00:29:15.220 |
against like the tyranny of chatbots as form factor. 00:29:18.420 |
And I think that the reason why the form factor matters 00:29:21.180 |
is the form factor changes what data you collect 00:29:24.500 |
And so I think we've spent a lot of time doing full 00:29:27.660 |
of like vertical integration of all these bits 00:29:33.020 |
I'll plug Amelia Weinberger's talk at our conference 00:29:38.740 |
behind like what else exists other than chatbots 00:29:41.260 |
that if you could delegate to reliable agents, 00:29:46.500 |
I mean, so I was kind of excited at Adept Experiments 00:29:50.740 |
I don't know what the official name for it is. 00:29:52.900 |
I was like, okay, like this is something I can use, 00:29:55.500 |
but it seems like it's just an experiment for now. 00:29:59.240 |
So we just use experiments as like a way to go push 00:30:01.400 |
various ideas on the design side to some people 00:30:06.640 |
And actually the experiments code base underpins 00:30:18.600 |
for us to go deploy arbitrary cards on the side. 00:30:24.280 |
I would love to talk about the interaction layer. 00:30:32.160 |
I think there was some rumors about open app building agents 00:30:35.160 |
that are kind of like they manage the end point. 00:30:41.360 |
Like, and I know I read in one of your papers, 00:30:46.120 |
kind of like you don't just take the dome and act on it. 00:30:50.320 |
How do you think about the best way the models will interact 00:30:53.520 |
with the software and like how the development of products 00:30:58.320 |
as more and more of the work is done by agents 00:31:02.720 |
And it's actually one of the things I'm really excited about. 00:31:06.360 |
I've spent most of my time doing research stuff, 00:31:10.520 |
that I've been learning about and I find it really cool. 00:31:18.640 |
to why Adept is pursuing a path of being able to just use 00:31:34.320 |
Like in a world where you had T equals infinity, right? 00:31:37.520 |
You're probably gonna have various different form factors 00:31:39.400 |
that robots could just be in and like all the specialization 00:31:42.320 |
but the fact is that humans live in a human environment. 00:31:44.560 |
So having a human robot lets you do things that humans do 00:31:51.680 |
Like if you go itemize out the number of things 00:31:59.240 |
those numbers of workflows add up pretty close to zero. 00:32:03.640 |
you need the ability to actually control your computer 00:32:06.340 |
It also lets you learn from human usage of computers 00:32:09.240 |
as a source of training data that you don't get 00:32:20.880 |
I think because it's the most practical path, 00:32:31.160 |
of the agent interaction layer level is a little bit like, 00:32:34.400 |
do y'all remember Windows 3.1, like those days? 00:32:38.080 |
Okay, I might be too old for you guys on this, 00:32:41.320 |
but like back in the day, Windows 3.1, right? 00:32:59.000 |
and then it would give you the C colon slash thing, 00:33:24.000 |
as more and more trust is built towards agents, 00:33:25.800 |
and more and more things can be done by agents, 00:33:27.720 |
and more UIs for agents are actually generative 00:33:31.240 |
then that just becomes a standard interaction layer. 00:33:33.800 |
And if that becomes a standard interaction layer, 00:33:40.680 |
or like certain customized workflow execution engines. 00:34:06.560 |
what using Rabbit in real life will actually be like 00:34:18.320 |
The agent knows how to break your goal down into steps. 00:34:20.320 |
The agent knows how to use the underlying software 00:34:22.600 |
and systems of record to achieve that goal for you. 00:34:24.880 |
The agent maybe presents you information in a custom way 00:34:27.880 |
that's only relevant to your particular goal. 00:34:32.440 |
where you don't really need to ever interface 00:34:36.840 |
unless you're a power user for some niche thing. 00:34:54.100 |
And, you know, the primary split in self-driving 00:34:58.800 |
And I feel like most agent companies that I'm tracking 00:35:03.080 |
are all moving towards camera approach, which is like-- 00:35:06.600 |
- The non-multimodal vision, very, very heavy vision. 00:35:11.520 |
you're focusing on that, including charts and tables and-- 00:35:16.840 |
- Do you find like inspiration there from like, 00:35:23.880 |
I think sometimes the most useful inspiration 00:35:26.240 |
I've found from self-driving is the levels analogy. 00:35:31.240 |
And I think that's great. - Level one to five. 00:35:34.280 |
But I think that our number one goal is for agents 00:35:42.320 |
that you just have to bang your head at for a long time 00:35:47.940 |
which is basically what's happened in self-driving. 00:35:51.520 |
where you have the data flywheel immediately, 00:35:53.720 |
and that takes you all the way up to the top. 00:35:55.600 |
But similarly, I mean, like compared to self-driving, 00:35:58.240 |
like two things that people really undervalue 00:36:03.080 |
driving a car down highway 101 in a sunny day demo, right? 00:36:06.680 |
Like that actually doesn't prove anything anymore. 00:36:13.920 |
I think one of the things that we believe really strongly 00:36:26.720 |
get a lot of reliability is like a really strong focus 00:36:29.820 |
on like actually why does the model not do this thing? 00:36:33.800 |
the time the model doesn't actually do the thing 00:36:36.100 |
is because if you're a wizard of ozzing it yourself, 00:36:38.260 |
or if you have unreliable actuators, you can't do the thing. 00:36:41.580 |
And so we've had to fix a lot of those problems. 00:36:51.000 |
as the most, I guess, real case of agents that we have, 00:37:00.000 |
but it has taken a long time for self-driving to mature. 00:37:05.280 |
and the 101, the driving down 101 on a sunny day moment 00:37:11.560 |
- And then, you know, cruise, you know, RIP recently. 00:37:24.060 |
that I'm curious to get your commentary on is there's, 00:37:30.780 |
into just general like sort of production readiness 00:37:34.180 |
'Cause you have reliability, you also have cost, 00:37:39.740 |
All of that seems to, tends towards wanting to reduce, 00:37:44.520 |
the tendency or the temptation is to reduce generality, 00:37:47.520 |
to improve reliability and to improve cost, improve speed. 00:38:05.840 |
And I think the way you get there is basically like, 00:38:09.320 |
how do you frame the fundamental agent problem 00:38:11.920 |
in a way that just continues to benefit from data? 00:38:15.320 |
And I think that, I think like one of the main ways 00:38:19.200 |
of like being able to solve that particular trade-off 00:38:21.640 |
is like, you basically just want to formulate the problem 00:38:34.680 |
okay, are you overfitting on these end use cases, right? 00:38:44.200 |
- I mean, so then the question becomes kind of, 00:39:03.280 |
Like that is just, you have a good base model 00:39:10.680 |
I think there's like two paths to a lot more capability 00:39:19.120 |
I think one path is you figure out how to spend, 00:39:23.840 |
I think the other path, and so like in that path, right, 00:39:26.280 |
I consider search, RL, all the things that we all, 00:39:29.920 |
that we all love in this era as part of that path, 00:39:34.740 |
The second path is how do you get like super competent, 00:39:39.740 |
high intelligence demonstrations from humans. 00:39:48.700 |
Like the first one gives you maximum sample efficiency 00:40:06.000 |
Probably this is a bit of a, too much of a trend right now, 00:40:17.740 |
as being able to help people do things on computers 00:40:25.420 |
- You could do a lot of stuff when you have an environment. 00:40:28.660 |
- We were having dinner for our one year anniversary. 00:40:34.900 |
and we mentioned you were coming on the pod with, 00:40:39.560 |
- Yeah, this is our first, I guess, like mailbag question. 00:40:42.480 |
He asked, when you started GPT-4 Data and Exist, 00:40:48.420 |
which can help you building a lot of those things. 00:40:51.580 |
How do you think about the things that are unique to you 00:40:56.980 |
maybe research direction that you want to take the team 00:40:59.500 |
and what you want people to come work on at ADAPT 00:41:05.020 |
that you didn't expect everybody would have access to? 00:41:12.700 |
so he can push back on my assumption about his question. 00:41:15.900 |
But I think implicit in that question is like, 00:41:26.120 |
And maybe part of the assumption is that advantage accrues 00:41:41.700 |
that is much more than that of the base model itself. 00:41:45.220 |
And so I think like that is like always gonna be 00:42:00.880 |
So like we're allocating our capacity wisely, 00:42:10.600 |
in the broader foundation modeling space is like, 00:42:17.520 |
like how good is agents as like a startup area, 00:42:22.680 |
I feel super good that we're doing foundation models 00:42:28.260 |
within ADAPT is flowing from, can we make a better agent? 00:42:34.500 |
if you're training on publicly available web data, 00:42:37.500 |
you put in the flops and you do reasonable things, 00:42:41.780 |
And if you just double the amount of compute, 00:42:45.340 |
And so like, I think pure play foundation model companies 00:43:01.800 |
behind just training these base foundation models. 00:43:04.200 |
I think it's gonna commoditize a lot of the regular LLMs 00:43:10.680 |
So I feel really good that we're just focused on agents. 00:43:16.560 |
- No, because if we were a pure play foundation model 00:43:18.200 |
company, we would be training general foundation models 00:43:26.800 |
- Yeah, and our business is an agent business. 00:43:36.380 |
- It's like, if you have a particular area of specialty, 00:43:43.940 |
everyone's just scaling to ridiculous levels of compute. 00:43:48.500 |
I find that, I think it's gonna be a little tougher. 00:43:59.420 |
figure is like a big, also sort of open AI affiliated 00:44:05.260 |
I think, I mean, I don't know exactly what they're doing, 00:44:13.860 |
- What question would you ask if we had them on? 00:44:16.840 |
- Oh, I just wanna understand what their overall strategy 00:44:19.020 |
is gonna be between now and when there's reliable stuff 00:44:22.940 |
But honestly, I just don't know enough about it. 00:44:24.500 |
- And if I told you, hey, fire your entire workforce, 00:44:28.060 |
warehouse workforce, and put robots in there. 00:44:42.520 |
But I think like, look, I think there's two things. 00:44:52.060 |
Like, it's just, I think it's just gonna work. 00:44:59.260 |
like, we've been on this podcast just continually saying, 00:45:08.380 |
And then you figure out everything else you have to do 00:45:10.580 |
in order to teach you how to solve new problems. 00:45:29.380 |
- We had Ken June from Mimboo on the podcast. 00:45:34.860 |
- Yeah, we asked her why people should go work there 00:45:48.080 |
And she said, they're really excited about building 00:45:52.160 |
And for her, the biggest research thing was, like, 00:46:01.360 |
why should people be excited to come work at ADAPT 00:46:04.920 |
And maybe what are, like, the core research questions 00:46:16.620 |
but, like, the AI space, to the extent there's an AI space 00:46:21.140 |
and the AI agent space are both, like, exactly, 00:46:24.580 |
as she likely said, like, I think colossal opportunities 00:46:28.340 |
and, like, people are just gonna end up winning 00:46:31.380 |
in different areas and people are all just gonna, 00:46:35.420 |
So I really don't feel that zero-something at all. 00:46:37.980 |
I would say, like, to, like, change the zero-sum framing 00:46:43.400 |
I think there's two huge reasons to be at ADAPT. 00:46:46.360 |
I think one of them is, like, everything we do 00:46:53.600 |
Like, we do a lot of research in service of that goal, 00:47:00.480 |
And I think the second reason to work at ADAPT 00:47:02.840 |
is if you believe that actually having customers 00:47:06.960 |
lets you build a GI faster, which we really believe, 00:47:11.260 |
And I think the examples for why that's true is, like, 00:47:20.100 |
They're, like, okay, like, we have a customer 00:47:21.940 |
that really needs us to do these particular things. 00:47:33.560 |
and the new ones that even are not saturated, 00:47:43.180 |
Like, we're equally excited about the same problems 00:47:45.980 |
around reasoning and planning and generalization 00:47:52.940 |
they're very grounded in actual needs right now, 00:48:00.060 |
but, you know, I would just leave it kind of open to you. 00:48:15.180 |
- Wow, okay, so Amelia's already made the rant 00:48:17.680 |
better than I have, but, like, not just chatbots 00:48:23.120 |
Rant two is, like, AI's really been the story of compute 00:48:32.560 |
And I think as much as our research community 00:48:37.560 |
is really smart, like, we have made many, many advancements, 00:48:43.480 |
but, like, now I think the game is increasingly changing, 00:48:47.160 |
and, like, the rapid industrialization era has begun, 00:48:52.160 |
and I think we, unfortunately, have to embrace it. 00:48:55.520 |
- Awesome, David, thank you so much for your time.