back to indexThe State of AI Startups in 2024 [LS Live @ NeurIPS]

00:00:08.900 |
Thanks to everyone who made it early morning. 00:00:11.880 |
It's like really weird experiments that we wanted to try 00:00:16.920 |
but two also, I've been to a number of these things now 00:00:23.880 |
and we wanted an opportunity while everyone is in town 00:00:27.000 |
in like one central spot to get everyone together 00:00:29.920 |
to talk about the best stuff of the year, review the year. 00:00:33.000 |
It's very nice that New York is always the end of the year. 00:00:36.000 |
And so I'm very honored that Sarah and Pranav 00:00:42.320 |
Sarah, I've known for, I was actually counting, 17 years. 00:00:51.520 |
- But she's been enormously successful as an AI investor. 00:01:05.260 |
but he's also starting to write really incredible posts 00:01:07.960 |
and opinions about what he's seeing as an investor. 00:01:10.500 |
So I wanted to kick this off with an industry session. 00:01:13.360 |
We have a great day of sort of like best of year recaps 00:01:18.360 |
I think Vic is here as well and the RoboFlow guys. 00:01:31.120 |
My name is Sarah Guo and thanks to Sean and friends here 00:01:36.520 |
So I'd start by just giving 30 seconds of intro. 00:01:48.920 |
They range from companies at the infrastructure level 00:01:55.180 |
to a foundation model companies, alternative architectures, 00:02:10.360 |
was that we thought that there was a really interesting 00:02:17.800 |
in how people use technology in our lifetimes. 00:02:20.240 |
And that represented huge economic opportunity 00:02:31.440 |
the types of products and founders that you back change, 00:02:53.000 |
Sorry, I'm just making sure that isn't actually 00:03:04.520 |
we'll cover some of the model landscapes and themes 00:03:14.560 |
So I thought it'd be useful to start from like, 00:03:17.960 |
what was happening at NeurIPS last year in December, 2023. 00:03:27.920 |
it's hard to believe, but like roughly a year ago, 00:03:29.440 |
you could only input text and get text out of ChatGPT. 00:03:33.160 |
The Mistral folks had just launched the Mixtral model 00:03:39.600 |
I very genuinely forgot about the existence of Bard 00:03:45.600 |
that they were doing their first round of AI regulation, 00:03:53.160 |
there's at least five themes that we could come up with 00:03:57.080 |
of what 2024 has meant for AI and for startups. 00:04:02.680 |
first, it's a much closer race on the foundation model side 00:04:13.960 |
So you get two responses from two language models, 00:04:20.040 |
means that you're preferred two thirds of the time. 00:04:24.760 |
was like more than 100 points better than anything else. 00:04:27.360 |
And the view from the ground was roughly like, 00:04:29.440 |
OpenAI is the IBM, there is no point in competing. 00:04:32.040 |
Everyone should just give up, go work at OpenAI 00:04:39.680 |
I think it would have been unbelievable a year ago 00:04:41.960 |
if you told people that, A, the best model today on this, 00:04:59.400 |
And this seems true, not just on the eval side, 00:05:04.960 |
but it's actually just OpenAI and Anthropic spend. 00:05:19.480 |
both that language models are pretty easy APIs to switch out 00:05:22.480 |
and people are trialing a variety of different options 00:05:29.520 |
is that open source is increasingly competitive. 00:05:40.400 |
the foundation models clearly care a great deal about. 00:05:44.520 |
on math instruction following and adversarial robustness. 00:05:47.760 |
The Lama model is amongst the top three of evaluated models. 00:05:53.080 |
just to point out that this isn't true across the board. 00:05:56.680 |
where foundation model companies have had more data 00:05:59.280 |
or more expertise in training against these use cases, 00:06:03.560 |
open source models are surprisingly increasingly effective. 00:06:23.640 |
the belief was largely that most intelligence 00:06:34.880 |
would have been Mistral-7b, which on this eval, 00:06:45.160 |
and what you can fit into a fairly small form factor 00:06:50.760 |
And again, related, we think the price of intelligence 00:06:55.800 |
This is a graph of flagship OpenAI model costs, 00:06:59.000 |
where the cost of the API has come down roughly 80, 85%, 00:07:13.520 |
of different APIs and public inference options. 00:07:20.960 |
that a text editor had, or something like Notion or Coda, 00:07:24.000 |
that's somewhere in the volume of a couple thousand dollars 00:07:30.280 |
It's clearly not the same distribution of data, 00:07:35.200 |
there's an enormous volume of data that you can create. 00:07:43.480 |
We're lucky to work with the folks at Chai Discovery, 00:07:46.160 |
who just released Chai 1, which is open source model 00:07:49.840 |
It's impressive that this is like roughly a year of work 00:07:56.120 |
But models in domains like biology are beginning to work. 00:07:59.400 |
We think that's true on the voice side as well. 00:08:03.280 |
before things like 11 Labs have existed for a while, 00:08:05.480 |
but we think low latency voice is more than just a feature, 00:08:08.320 |
it's actually a net new experience and interaction. 00:08:12.960 |
than the historical transcription first models. 00:08:14.920 |
Same thing with many of the Cartesian models. 00:08:17.880 |
And then a new nascent use case is execution. 00:08:22.360 |
OpenAI launched code execution inside of Canvas yesterday. 00:08:36.600 |
Cognition, the company was founded under a year ago. 00:08:50.800 |
And I think the whole industry now considers that, 00:09:05.160 |
and managed to sign up before they cut off accesses. 00:09:07.200 |
So here's my favorite joke in the form of a video. 00:09:12.720 |
Yeah, you're telling me I shrimp fried this rice. 00:09:17.960 |
It's a pretty bad joke, but I really like it. 00:09:34.040 |
So this is Javier Millet, who speaks in Spanish, 00:09:36.800 |
but here you will hear him in English if this plays. 00:09:42.080 |
the original tonality of his speech and performance. 00:09:58.080 |
his original tone and the emotion in his speech, 00:10:02.000 |
which is definitely new and pretty impressive 00:10:15.520 |
I think there's a great debate happening here later today 00:10:18.160 |
on the question of this, but we think at minimum, 00:10:21.080 |
it's hard to deny that there are at least some limits 00:10:27.880 |
but there also seems like there are new scaling paradigms. 00:10:32.720 |
It seems like OpenAI has cracked a version of this 00:10:34.880 |
that works, and we think, A, foundation model labs 00:10:38.680 |
and B, so far it largely works for very verifiable domains, 00:10:46.400 |
where we can get an objective value function. 00:10:48.800 |
And I think an open question for the next year 00:10:50.320 |
is going to be how do we generate those value functions 00:10:52.080 |
for spaces that are not as well-constrained or well-defined? 00:10:55.080 |
And so the question that this leaves us in is like, 00:11:07.760 |
that is largely unjustified based on outcomes 00:11:12.320 |
and startups are largely raising money on hype. 00:11:27.240 |
But if you break out the numbers here a bit more, 00:11:36.380 |
which is upwards of 30 to $40 billion this year. 00:11:39.000 |
And so the reality of the funding environment 00:11:41.040 |
actually seems like much more sane and rational. 00:11:43.480 |
It doesn't look like we're headed to a version of 2021. 00:11:47.900 |
for an outsized amount of money being raised, 00:12:03.260 |
We think zero to 20 and just PLG style spending 00:12:10.040 |
And so what we wanted to try and center discussion on, 00:12:17.460 |
that are making 10 million more or revenue and growing, 00:12:44.380 |
or it's too hard to manage those set of people. 00:12:52.260 |
it's really useful to do like next level automation. 00:12:57.060 |
the story is you can do first wave professional services 00:13:09.980 |
like how effective text modalities have been. 00:13:15.060 |
And there's a whole host of not safer work chatbots as well 00:13:18.060 |
that are pretty effective at just text generation. 00:13:22.940 |
Perplexity and Glean have demonstrated this as well. 00:13:31.220 |
We think it's likely text isn't the last medium. 00:13:36.600 |
or sets of engagement that are more engaging. 00:13:38.900 |
But this feels like a pretty interesting place to start. 00:13:46.540 |
So one thing that I've worked on investing in 00:13:49.940 |
in a long time is democratization of different skills, 00:14:03.660 |
and now code and really fully functioning applications. 00:14:09.620 |
about the growth driver for all of these companies 00:14:18.020 |
we, the venture industry, you know, the royal we, 00:14:24.980 |
is that there's actually much more instinct for creativity, 00:14:28.860 |
visual creativity, audio creativity, technical creativity, 00:14:38.740 |
Midjourney was a company that is in the vanguard here 00:14:45.540 |
is like how many people want to generate images 00:14:52.120 |
they can't be using these professional contexts 00:15:02.180 |
And we think the range of quality and controllability 00:15:10.840 |
is still, it's very deep and we're still very early. 00:15:15.940 |
if we're in the first or second inning of this AI wave, 00:15:22.820 |
and to go build companies is the enabling layers, right? 00:15:26.180 |
Shorthand for this is obviously compute and data. 00:15:37.900 |
We'll talk about that later in terms of who has, 00:15:40.140 |
like let's say reasoning traces in different domains 00:15:43.080 |
that are interesting to companies doing their own training. 00:15:46.360 |
But this is an area that has seen explosive growth 00:16:00.080 |
that some part from companies, some part from investors, 00:16:05.000 |
it's a fun debate as to where is the value in the ecosystem 00:16:20.520 |
that there was no value at the application layer. 00:16:24.500 |
and then like nobody's gonna catch open AI in pre-training. 00:16:27.040 |
And this isn't like a knock on open AI at all. 00:16:32.040 |
These labs have done amazing work enabling the ecosystem 00:16:35.140 |
and we continue to partner with them and others. 00:16:37.940 |
But it's simply untrue as a narrative, right? 00:16:51.700 |
You have price competition, you have open source. 00:16:55.100 |
I think an underappreciated impact of test time scaling 00:17:06.780 |
useful to somebody, the customer can pay for the compute 00:17:26.380 |
if we look at the sort of cluster of companies 00:17:30.060 |
shown that it is creating and capturing value 00:17:39.940 |
that I think is actually shared by many of the labs 00:17:54.860 |
or another great debate that Sean could host is like, 00:18:16.940 |
that have been considered traditionally like too hard, right? 00:18:19.700 |
Like just bad markets for the venture capital spec, 00:18:27.820 |
where the end output is a tens of billions of dollars 00:18:34.660 |
And these included areas like legal healthcare, 00:18:39.620 |
any traditional venture firm would say like bad market, 00:18:43.820 |
nobody makes money there, it's really hard to sell, 00:18:50.460 |
that has actually been effective over the past year, 00:18:56.580 |
And so perhaps one of our more optimistic views 00:19:05.580 |
that is several magnitudes, orders of magnitude cheaper, 00:19:09.940 |
then actually you can change the buying pattern 00:19:13.380 |
And maybe the legal industry didn't buy anything 00:19:26.660 |
that have just like really enormous user basis 00:19:29.860 |
and engagement until we had these consumer chatbots 00:19:38.220 |
As Pranav mentioned, we think that the opportunity 00:19:49.860 |
And finally, in terms of the markets that we look at, 00:19:56.180 |
that you can sell against outcomes and services 00:20:01.380 |
because you're doing work versus just giving people 00:20:07.740 |
we think there's elastic demand for many services, right? 00:20:11.340 |
Our classic example is there's on order of 20 to 25 million 00:20:16.340 |
professional software developers in the world. 00:20:19.180 |
I imagine much of this audience is technical. 00:20:29.180 |
and high quality software down to orders of magnitude, 00:20:32.220 |
we're just gonna end up with more software in the world. 00:20:34.700 |
We're not gonna end up with fewer people doing development. 00:20:39.700 |
And then finally, on the incumbent versus startup question, 00:20:47.260 |
have the distribution, the product surfaces and the data. 00:20:52.980 |
and share some of it back with their customers. 00:21:00.860 |
Like the point of the startup is you have to go fight 00:21:02.820 |
with a better product or a more clever product 00:21:09.580 |
But the specifics around the product surface and the data 00:21:17.140 |
If you look at the SaaS companies that are dominant, 00:21:24.220 |
I might actually decrease the number of seats. 00:21:33.460 |
that have been written to enable a particular workflow 00:21:41.140 |
may not matter if I don't want people to do that workflow 00:21:43.860 |
of filling out the database every Friday anymore. 00:21:49.580 |
or the incumbent advantage gets highly challenged 00:21:59.460 |
is no one has the data we want in many cases, right? 00:22:16.780 |
Like that sounds like a very useful set of data. 00:22:19.940 |
And the incumbent companies in any given domain, 00:22:23.940 |
Like they have a database with the outputs some of the time. 00:22:32.300 |
is when an incumbent says they have the data, 00:22:41.260 |
our shorthand for the set of changes that are happening 00:23:00.380 |
I think that some of the CEOs of large companies 00:23:04.100 |
but they're still trying to make 100,000 people 00:23:08.820 |
The market opportunities are different, right? 00:23:10.940 |
These markets that we think are interesting and very large, 00:23:15.660 |
are not just the replacement software markets 00:23:24.500 |
Sierra just started talking about charging for outcomes. 00:23:27.420 |
Outcomes-based pricing has been this holy grail idea 00:23:37.580 |
And so, our companies, they spend a lot more on compute 00:23:43.220 |
They spend a lot with the foundation model providers. 00:23:49.340 |
It's a time where you need to be really creative 00:23:51.180 |
about product versus just replace the workflows of the past. 00:23:56.180 |
And it might require ripping out those workflows entirely. 00:24:01.460 |
I bet most of the people in this room have written evals 00:24:04.660 |
and compared the academic benchmark to a real-world eval 00:24:14.900 |
the non-deterministic nature of these outputs 00:24:19.700 |
I think that's a different way to think about product 00:24:24.140 |
And we need to think about infrastructure again. 00:24:27.260 |
There was this middle period where the cloud providers, 00:24:33.300 |
and it was all just gonna be front-end people at some point. 00:24:37.620 |
We're back in the hardware era where people are acquiring 00:24:46.380 |
So I guess we'll end with a call to action here 00:24:50.740 |
and encourage all of you to seize the opportunity. 00:24:55.340 |
It is the greatest technical and economic opportunity 00:25:02.340 |
And we do a lot of work with the foundation model companies. 00:25:10.140 |
and they're great partners and even co-investors 00:25:13.660 |
But I think all of the focus on their interesting missions 00:25:24.260 |
The world is very large, and we think much of the value 00:25:27.060 |
will be distributed in the world through an unbundling 00:25:47.900 |
that is much friendlier to startups than 2023. 00:26:07.660 |
from Latent Space Live at NeurIPS 2024 in Vancouver. 00:26:17.420 |
and Pranav Reddy for sharing their invaluable insights 00:26:24.900 |
for their presentation slides, social media links,