back to indexAgents @ Work: Lindy.ai (with live demo!)
Chapters
0:0 Introductions
4:5 AI engineering and deterministic software
8:36 Lindys demo
13:21 Memory management in AI agents
18:48 Hierarchy and collaboration between Lindys
21:19 Vertical vs. horizontal AI tools
24:3 Community and user engagement strategies
26:16 Rickrolling incident with Lindy
28:12 Evals and quality control in AI systems
31:52 Model capabilities and their impact on Lindy
39:27 Competition and market positioning
42:40 Relationship between Factorio and business strategy
44:5 Remote work vs. in-person collaboration
49:3 Europe vs US Tech
58:59 Testing the Overton window and free speech
64:20 Balancing AI safety concerns with business innovation
00:00:06.360 |
This is Alessio, partner and CTO at Decibel Partners, 00:00:09.040 |
and I'm joined by my co-host, Swix, founder of Small AI. 00:00:31.520 |
Right now you're a solo founder/CEO of Lindy.ai. 00:00:44.560 |
Like you can just pin up AI agents super easily 00:00:52.440 |
that you simply could not automate before in a few minutes. 00:00:57.360 |
I think you spoke at our Latent Space anniversary. 00:01:06.480 |
like we actually already scheduled this podcast 00:01:17.080 |
I think we are having a little bit of a moment. 00:01:18.520 |
I think it's a bit premature to say we're blowing up, 00:01:25.520 |
I would say we started working on that six months ago. 00:01:29.760 |
It's just, I guess, I guess that's what we're doing now. 00:01:33.400 |
And so we've basically been cooking for the last six months, 00:01:35.680 |
like really rebuilding the product from scratch. 00:01:42.000 |
- If you log in now, the platform looks very different. 00:01:48.360 |
and I think a lot of folks in the agent space 00:01:51.800 |
is that there is such a thing as too much of a good thing. 00:01:55.000 |
I think many people, when they started working on agents, 00:01:57.520 |
they were very LLM-peeled and chat GPT-peeled, right? 00:02:01.360 |
They got ahead of themselves in a way, as included, 00:02:15.960 |
hey, actually, the more you can put your agent on Rails, 00:02:18.760 |
one, the more reliable it's going to be, obviously, 00:02:21.280 |
but two, it's also going to be easier to use for the user, 00:02:25.000 |
you get, instead of just getting this big, giant, 00:02:27.600 |
intimidating text field, and you type words in there, 00:02:30.000 |
and you have no idea if you're typing the right words or not, 00:02:32.560 |
here you can really click and select step-by-step 00:02:37.480 |
and really give as narrow or as wide a guardrail 00:02:43.520 |
We called it Lindy on Rails about six months ago. 00:02:46.040 |
And we started putting it into the hands of users 00:02:48.720 |
over the last, I would say, two months or so. 00:02:50.560 |
And that's, I think things really started going 00:02:54.520 |
The agent is way more reliable, way easier to set up, 00:02:57.040 |
and we're already seeing a ton of new use cases pop up. 00:03:01.700 |
You launched the first Lindy in November last year, 00:03:05.160 |
and you were already talking about having a DSL, right? 00:03:10.200 |
and you were like, it's just much more reliable. 00:03:14.000 |
Is this a UI-level change, or is it a bigger rewrite? 00:03:16.960 |
FRANCESC CAMPOY: No, it is a much bigger rewrite. 00:03:24.320 |
And it's like, hey, every time you receive a Zendesk ticket, 00:03:35.040 |
was you would type the prompt asking it to do that. 00:03:39.920 |
you check my knowledge base, and so on and so forth. 00:03:42.160 |
The problem with doing that is that it can always go wrong. 00:03:44.600 |
Like, you're praying the LLM gods that they will actually 00:03:47.440 |
invoke your knowledge base, but I don't want to ask it. 00:03:52.040 |
consult the knowledge base after it receives a Zendesk ticket. 00:03:55.880 |
have the trigger, which is Zendesk ticket received, 00:03:58.280 |
have the knowledge base consult, which is always there, 00:04:01.880 |
So you can really set up your agent any way you want like 00:04:06.760 |
I think about for AI engineering as well, which 00:04:09.660 |
is the big labs want you to hand over everything in the prompts 00:04:14.560 |
And then the smaller brains, the GPU pourers, 00:04:19.800 |
things more deterministic, and reliable, and controllable. 00:04:24.560 |
and make it a very small, like the minimal viable box. 00:04:29.680 |
FRANCESC CAMPOY: I love that characterization, 00:04:32.200 |
Yeah, we talk about using as much AI as necessary 00:04:37.040 |
YOSSI ELKRIEF: And what was the choosing between this drag 00:04:40.420 |
and drop, low code, whatever, super code-driven, 00:04:46.760 |
And maybe the flip side of it, which you don't really do, 00:04:53.640 |
Whatever you learn, actually putting this in front of users 00:05:01.160 |
kind of like Ruby on Rails, instead of Lindy on Rails, 00:05:03.400 |
it's kind of like defaults over configuration. 00:05:14.560 |
I actually sort of agree now that text is really not great. 00:05:18.040 |
I think for people like you and me, because we sort of have 00:05:20.280 |
a mental model, OK, when I type a prompt into this text box, 00:05:24.400 |
It's going to map it to this kind of data structure 00:05:27.800 |
I guess it's a little bit blackmailing towards humans. 00:05:44.320 |
I would say 60% or 70% of the prompts that people type 00:05:48.600 |
Me as a human, as AGI, I don't understand what they mean. 00:05:52.360 |
It is actually, I think, whenever you can have a GUI, 00:05:55.160 |
it is better than to have just a pure text interface. 00:05:58.160 |
And then how do you decide how much to expose? 00:06:01.400 |
So even with the tools, you have a bunch of them. 00:06:05.440 |
You have Slack, you have Google Calendar, you have Gmail. 00:06:08.360 |
Should people, by default, just turn over access to everything, 00:06:11.640 |
and then you help them figure out what to use? 00:06:15.040 |
Because when I tried to set up Slack, it was like, hey, 00:06:17.640 |
give me access to all channels and everything. 00:06:19.960 |
Which, for the average person, probably makes sense, 00:06:22.200 |
because you don't want to re-prompt them every time 00:06:25.000 |
But at the same time, for maybe the more sophisticated 00:06:29.760 |
I want to really limit what you have access to. 00:06:35.200 |
The general philosophy is we ask for the least amount 00:06:42.360 |
I could be mistaken, but I don't think Slack lets you request 00:06:46.880 |
But, for example, for Google, obviously there's 00:06:49.400 |
hundreds of scopes that you could require for Google. 00:06:53.160 |
And sometimes it's actually painful to set up your Lindy, 00:06:55.840 |
because you're going to have to ask to Google and add scopes 00:07:07.480 |
I need to be able to read your email so I can draft a reply. 00:07:13.040 |
So we just try to do it very incrementally like that. 00:07:16.340 |
Do you think OAuth is just overall going to change? 00:07:28.300 |
get different permissions every time from different parts? 00:07:31.500 |
Like, do you ever think about designing things knowing 00:07:33.860 |
that maybe AI will use it instead of humans will use it? 00:07:38.060 |
One pattern we've started to see is people provisioning accounts 00:07:42.860 |
And so in particular, Google Workspace accounts. 00:07:44.840 |
So, for example, Lindy can be used as a scheduling assistant. 00:07:50.680 |
when you're trying to find time with someone. 00:07:53.960 |
going to go back and forth in a flow of abilities and so forth. 00:07:56.660 |
Very often, people don't want the other party 00:08:00.280 |
So it's actually funny, they introduce delays. 00:08:06.000 |
And they provision an account on Google Suite, 00:08:08.240 |
which costs them like $10 a month or something like that. 00:08:14.580 |
I'm not optimistic on us actually patching OAuth, 00:08:19.540 |
We would want to patch OAuth, because the new account thing 00:08:24.140 |
You would want to patch OAuth to have more granular access 00:08:27.060 |
control and really be able to put your showbiz in the box. 00:08:30.660 |
I'm not optimistic on us doing that before AGI, I think. 00:08:35.260 |
FRANCESC CAMPOY: That's a very close timeline. 00:08:37.180 |
I'm mindful of talking about a thing without showing it. 00:08:47.380 |
But also, let's have a look at how you show off Lindy. 00:08:54.780 |
and then I'll graduate to a much more complicated one. 00:08:57.620 |
A super simple Lindy that I have is I unfortunately 00:09:05.940 |
And I put them on the Holydew, which is like the French Airbnb, 00:09:10.380 |
And so I received these emails from time to time telling me, 00:09:20.500 |
Doing this without an AI agent or without AI in general 00:09:24.420 |
is a pain in the butt, because you must write an HTML parser 00:09:34.260 |
By contrast, the way it works with Lindy, it's really simple. 00:09:45.340 |
And so this is where you can see the AI part, where 00:09:55.140 |
And so I can say, OK, you extract from the email 00:10:01.500 |
You extract the number of travelers of the reservation. 00:10:04.760 |
And now you can see, when I look at the task history of this, 00:10:11.520 |
And boom, I'm appending this row to this spreadsheet. 00:10:15.520 |
So effectively, this node here, this append row node, 00:10:23.840 |
It has context over the task, and it's appending the row. 00:10:28.120 |
And then it's going to send a reply to the thread. 00:10:43.800 |
you can configure which model you want to power the node. 00:10:49.500 |
Much more complex example, my meeting recorder. 00:10:53.280 |
It looks very complex, because I've added to it over time. 00:11:00.840 |
And after the meeting, you send me a summary. 00:11:04.640 |
So I receive-- like, my Lindy is constantly coaching me. 00:11:07.920 |
And so you can see here, in the prompt of the coaching notes, 00:11:10.440 |
I've told it, hey, was I unnecessarily confrontational 00:11:20.000 |
So I can really give it exactly the kind of coaching 00:11:23.400 |
And then the interesting thing here is, you can see, 00:11:26.240 |
the agent here, after it sends me these coaching notes, 00:11:34.880 |
But it's actually able to backtrack and resume 00:11:54.920 |
And I was able to follow up and just ask a follow-up question 00:11:59.000 |
And I was like, why did you find too technical 00:12:03.660 |
And so she basically picked up the automation 00:12:07.580 |
And she has all of the context of everything that happened, 00:12:11.580 |
So she was like, oh, you used the words "deterministic" 00:12:18.300 |
for every channel and every action that Lindy takes. 00:12:25.380 |
So this was a meeting where I was not, right? 00:12:29.360 |
He's an Indie meeting recorder, posts the meeting notes 00:12:34.760 |
So you can see, okay, this is the onboarding call we had. 00:12:48.400 |
And it's really handy because I know I can have 00:12:50.800 |
this sort of interactive Q&A with these meetings. 00:13:00.220 |
I have a five-minute chat with my Lindy afterwards. 00:13:03.660 |
She was like, well, this is what we replied to this customer, 00:13:06.020 |
and I can just be like, okay, good job, Jack. 00:13:09.260 |
So that's the kind of use cases people have with Lindy. 00:13:12.580 |
It's a lot of, there's a lot of sales automations, 00:13:15.140 |
customer support automations, and a lot of this, 00:13:17.140 |
which is basically personal assistance automations, 00:13:25.840 |
how does it track whether or not you're improving? 00:13:27.880 |
You know, if these are like mistakes you made in the past, 00:13:33.600 |
So I'll show you my meeting scheduler, Lindy, 00:13:39.340 |
And so every time I talk to her, she saves a memory. 00:13:42.680 |
If I tell her, you screwed up, please don't do this. 00:13:46.480 |
So you can see here, it's, oh, it's got a double memory here. 00:14:05.860 |
Yeah, so Lindy can just like manage her own memory 00:14:10.860 |
- Okay, I mean, I'm just gonna take the opportunity 00:14:13.460 |
to ask you, since you are the creator of this thing, 00:14:17.900 |
Like if you've been using this for two years, 00:14:19.820 |
there should be thousands of thousands of things. 00:14:22.940 |
Agents still get confused if they have too many memories, 00:14:41.980 |
this does not exist either pre-training or post-training. 00:14:46.640 |
Like it's a miracle that they can be agents at all. 00:14:49.220 |
And so what I do, I actually prune the memories. 00:14:55.380 |
from back when we had GPT 3.5 being Lindy agents. 00:15:05.620 |
The reason is 'cause I have another assistant 00:15:10.180 |
It comes up with a lot of like trivial, useless facts 00:15:17.580 |
I'd much rather have high quality facts that it accepts. 00:15:25.300 |
to only memorize this when I say memorize this? 00:15:36.780 |
but somewhere in here, there is a rule where I'm like, 00:15:48.980 |
and it's a newsletter, or it's like a cold outreach 00:15:56.340 |
hey, this email I want you to archive moving forward. 00:16:00.300 |
when I have this kind of email, it's really important. 00:16:06.300 |
like archive emails from X, save it as a new memory. 00:16:16.380 |
I recommend that everybody have a virtual mailbox. 00:16:18.660 |
You could set up a physical mail receive thing for Lindy. 00:16:32.580 |
So yeah, most likely I can process my physical mail. 00:16:40.300 |
So this would be like a 65 point Lindy, right? 00:16:51.460 |
This designer recruiter here is kind of beefy as well. 00:17:15.380 |
And the odds that it gets it right are so low. 00:17:17.380 |
But here, because we're really holding the agent's hand 00:17:19.740 |
step-by-step-by-step, it's actually super reliable. 00:17:28.980 |
So for example, here, this like AI agent step, right? 00:17:44.460 |
and it updates a log that I have of my blood pressure 00:17:49.420 |
- Oh, so this is a, every Lindy comes with a to-do list? 00:17:57.100 |
And so you can see here, this is my main Lindy, 00:18:04.020 |
if I am giving you a health-related fact, right here. 00:18:09.340 |
So then you update this log that I have in this Google Doc, 00:18:12.780 |
And you can see, I've actually not configured 00:18:15.620 |
I haven't told it what to send me a message for. 00:18:21.860 |
It's like, I'm giving it my blood pressure readings. 00:18:33.900 |
you could have multiple workflows in one Lindy. 00:18:36.100 |
I think the mental model is kind of like the Zapier. 00:18:48.660 |
- Yeah, frankly, I think the line is a little arbitrary. 00:18:54.740 |
versus when do you overload your current class? 00:18:57.580 |
I think of it in terms of like jobs to be done. 00:18:59.740 |
And I think of it in terms of who is the Lindy serving. 00:19:05.900 |
I give it a bunch of stuff, like very easy tasks. 00:19:10.740 |
Sometimes when a task is really more specialized, 00:19:13.220 |
so for example, I have this like summarizer Lindy 00:19:18.500 |
I wouldn't want to add this to my main Lindy, 00:19:22.380 |
Or when it's a Lindy that serves another constituency, 00:19:27.660 |
I don't want to add that to like my personal assistant. 00:19:31.700 |
And you can call a Lindy from within another Lindy. 00:19:38.580 |
- A couple more things for the video portion. 00:19:50.940 |
And she sends me, so she woke up yesterday actually, 00:19:56.940 |
And she looks for like the latest episode on YouTube. 00:19:59.100 |
And once she finds it, she transcribes the video. 00:20:17.220 |
Any interesting one, complicated one that you're proud of? 00:20:24.500 |
- So many of our workflows are about meeting scheduling. 00:20:26.300 |
So we had to build some very open unity tools 00:20:29.780 |
So for example, one that is surprisingly hard 00:20:36.140 |
this is like a thousand lines of code or something. 00:20:45.260 |
What are the meeting, like the weekdays in which I meet? 00:20:48.820 |
There's like, how many time slots do you return? 00:21:01.380 |
And it's really handy because anytime any bug happens, 00:21:04.140 |
so the Lindy reads our guidelines on a Google Docs. 00:21:07.340 |
By now the guidelines are like 40 pages long or something. 00:21:10.140 |
And so every time any new kind of bug happens, 00:21:12.540 |
we just go to the guideline and we add the lines like, 00:21:23.980 |
Or maybe how do you think about the complexity of these tasks 00:21:27.420 |
when it's going to be worth having kind of like 00:21:28.820 |
a vertical standalone company versus just like, 00:21:30.940 |
"Hey, a Lindy is going to do a good job 99% of the time." 00:21:44.300 |
versus when do you want to use a horizontal tool. 00:21:46.660 |
I think of it as very similar to the internet. 00:22:03.180 |
I think maybe the biggest exception is e-commerce. 00:22:25.780 |
I have a thesis, which is a really cool thesis for Lindy, 00:22:31.500 |
I think that by and large, in a lot of verticals, 00:22:34.780 |
agents in each vertical have more in common with agents 00:22:41.220 |
because that way your agents can work together. 00:22:46.500 |
and so you can create agents for everything that you want, 00:22:50.620 |
for a bunch of different platforms and so forth. 00:22:52.740 |
So I think ultimately it is actually going to shake out 00:22:56.900 |
in that search is everywhere on the internet. 00:23:01.540 |
So there's going to be a lot of vertical agents 00:23:08.300 |
but then I also think there are going to be a few 00:23:14.180 |
- Yeah, that is actually one of the questions 00:23:16.940 |
So I guess we can transition away from the screen 00:23:22.860 |
You're basically saying that the current VC obsession 00:23:26.180 |
of the day, which is vertical AI-enabled SaaS, 00:23:38.620 |
and there's also a lot of horizontal platforms. 00:23:42.420 |
basically the entire no-code space is very horizontal. 00:23:46.340 |
like there's a lot of very horizontal tools out there. 00:23:50.340 |
I was just trying to get a reaction out of you for hot takes. 00:23:54.660 |
No, I also think it is natural for the vertical solutions 00:23:58.020 |
to emerge first, 'cause it's just easier to build. 00:24:10.380 |
I also see some other people doing it for you for free. 00:24:22.140 |
like people posting their Lindys and so forth. 00:24:25.340 |
that you built the platform where creators see value 00:24:29.860 |
and so then your incentive is to make them successful 00:24:32.820 |
so that they can make other people successful, 00:24:34.340 |
and then it just drives more and more engagement 00:24:39.140 |
- Yeah, yeah, I mean, community is everything. 00:24:41.100 |
- Are you doing anything special there, any big wins? 00:24:44.620 |
- We have a Slack community that's pretty active. 00:24:47.220 |
I can't say we've invested much more than that so far. 00:24:51.620 |
so I have some involvement in the no-code community. 00:24:56.660 |
after no-code as a category got them a lot more allies 00:25:02.820 |
So it helps you to grow the community beyond just Lindy. 00:25:09.780 |
Maybe you want to call it something different, 00:25:17.820 |
and they're also kind of going after a similar market. 00:25:28.020 |
Linspace is growing the community of AI engineers, 00:25:30.420 |
and I think you have a slightly different audience of, 00:25:34.540 |
- Yeah, I think the no-code tinkerers is the community. 00:25:37.900 |
Yeah, it is going to be the same sort of community 00:25:39.900 |
as what Webflow, Zapier, Airtable, Notion to some extent. 00:25:43.220 |
- Yeah, the framing can be different if you were, 00:25:52.860 |
we're exclusively only for CEOs with a certain budget, 00:25:56.340 |
then you just have, you tap into a different budget. 00:25:59.140 |
The problem with EA is like the CEO has no willingness 00:26:01.980 |
to actually tinker and play with the platform. 00:26:06.020 |
Like a lot of your biggest advocates are CEOs, right? 00:26:09.180 |
- A solopreneur, you know, small business owners, 00:26:24.300 |
- It's one of the main jobs to be done, really. 00:26:29.460 |
so we have a Lindy obviously doing our customer support, 00:26:36.020 |
where someone was asking Lindy for video tutorials. 00:26:39.180 |
And at the time, actually, we did not have video tutorials, 00:26:50.180 |
And so we clicked on the link and it was a recall. 00:26:53.940 |
that the customer had not yet opened the email, 00:26:57.340 |
"Oh, hey, actually, sorry, this is the right link." 00:26:59.180 |
And so the customer never reacted to the first link. 00:27:08.540 |
and we found like, I think, like three or four 00:27:18.380 |
That's like, "Hey, don't recall people, please don't recall." 00:27:22.500 |
I mean, so you can explain it retroactively, right? 00:27:29.100 |
that obviously it learned to hallucinate that. 00:27:35.260 |
- I wouldn't be surprised if that takes one token. 00:27:37.300 |
Like there's a tokenizer and it's just one token. 00:27:46.140 |
Like, and you have to basically get it exactly correct. 00:27:55.660 |
- So this is just a jump maybe into evals from here. 00:28:01.980 |
that says, "Make sure my AI does not rickroll my customer." 00:28:10.300 |
when it's such like an open-ended problem space? 00:28:21.700 |
"In one click, we can turn it into effectively a unit test." 00:28:26.500 |
This is how you're supposed to handle things like this." 00:28:33.860 |
So it's very easy for us to spin up this kind of eval. 00:28:43.900 |
We're most likely going to switch to Brain Trust. 00:28:47.060 |
It's, well, when we built it, there was nothing. 00:28:51.260 |
And we, I mean, we started this project like end of 2022. 00:28:55.740 |
I wouldn't recommend it to build your own eval tool. 00:29:02.220 |
And that's not something we want to be spending our time on. 00:29:05.500 |
'cause I think my first conversations with you about Lindy 00:29:18.780 |
I think the ecosystem has matured a fair bit. 00:29:21.300 |
- What is one thing that Braintrust has nailed 00:29:25.460 |
- Well, not using them yet, so I couldn't tell. 00:29:29.900 |
like they're doing what we do with our eval tool, but better. 00:29:34.700 |
but also like 60 other companies do it, right? 00:29:36.340 |
So I don't know how to shop apart from brand. 00:29:42.660 |
- Yeah, like evals, there's two kinds of evals, right? 00:29:45.980 |
In some way, you don't have to eval your system as much 00:29:50.340 |
because you've constrained the language model so much. 00:29:55.260 |
that the structured outputs are going to be good, right? 00:30:02.020 |
constraint grammar sampling and all that good stuff. 00:30:22.300 |
I think the bar is it needs to be better than a human. 00:30:30.180 |
- Is there a limiting factor of Lindy at the business? 00:30:41.700 |
- Yeah, the raw capabilities for sure are a big limit. 00:30:52.980 |
It's kind of insane that we started building this 00:30:54.860 |
when the context windows were like 4,000 tokens. 00:30:56.660 |
Like today our system prompt is more than 4,000 tokens. 00:31:26.420 |
ask me for confirmation before actually sending it. 00:31:28.540 |
And so today you see the email that's about to get sent 00:31:33.780 |
And we are making it so that when you make a change, 00:31:36.420 |
we are then saving this change that you're making 00:31:40.220 |
And then we are retrieving these examples for future tasks 00:31:49.540 |
It's really like good old engineering and product work. 00:32:09.820 |
I don't think it's good for agentic behavior. 00:32:12.180 |
- Yeah, 3.5 Sonnet is when I started feeling that. 00:32:14.540 |
And then with prompt caching with 3.5 Sonnet, 00:32:25.780 |
is that your prompts are kind of dynamic, right? 00:32:28.700 |
you need the front prefix portion to be stable. 00:32:32.180 |
- Yes, but we have this append-only ledger paradigm. 00:32:38.260 |
and every filled node inherits all the context 00:32:54.220 |
we just, like the prompt caching works pretty well. 00:33:04.660 |
I wonder how much more money OpenAN and Anthropic are making 00:33:07.420 |
just because people don't rewrite their prompts 00:33:17.580 |
are routinely cutting their costs by two, four, five. 00:33:22.300 |
They want people to take advantage of these innovations. 00:33:25.540 |
Do you have any other competitive commentary? 00:33:32.460 |
I think the market is, look, I mean, AGI is coming. 00:33:43.020 |
I'm adding my small brick to this giant, giant, 00:33:48.260 |
we are gonna, this entire category of software 00:33:54.420 |
that it's going to create trillions of dollars of value 00:34:07.860 |
It's like, oh, like, stop, this is a 12 people team here. 00:34:11.220 |
I guess we'll set up this Lindy for one or two days, 00:34:17.140 |
there's this immense uncapped market opportunity. 00:34:22.620 |
I'm focused on the ocean more than on the sharks. 00:34:29.460 |
What are the high order bits of good agent design? 00:34:31.740 |
- The model, the model, the model, the model. 00:34:33.620 |
I think people fail to truly, and me included, 00:34:37.420 |
they fail to truly internalize the bitter lesson. 00:34:40.060 |
So for the listeners out there who don't know about it, 00:34:42.260 |
it's basically like, you just scale the model, 00:34:46.980 |
I think it also holds for the cognitive architecture. 00:34:49.820 |
I used to be very cognitive architecture-filled, 00:34:58.740 |
I think we're seeing it a little bit right now with O1. 00:35:14.260 |
Like it's like 14 on AIME versus O1, it's like 83. 00:35:21.780 |
- Like there's no cognitive architecture on top of it. 00:35:25.460 |
- And so as a founder, how do you think about that? 00:35:32.100 |
You know, you started Lindy, it's like 4K context, 00:35:38.260 |
and just like waiting for the models to get better. 00:35:40.300 |
How do you today decide, again, what to build next? 00:35:44.300 |
- Knowing that, hey, the models are gonna get better, 00:35:45.260 |
so maybe we just shouldn't focus on improving 00:35:48.980 |
and just build the connectors instead or whatever. 00:35:54.420 |
oh, when we have a feature idea or a feature request, 00:35:56.820 |
we ask ourselves like, is this the kind of thing 00:36:02.140 |
I'm reminded again, when we started this in 2022, 00:36:09.700 |
You really can't do anything with 4,000 tokens. 00:36:13.300 |
Like now it's like it was for nothing, right? 00:36:15.540 |
Now we just assume that infinite context windows 00:36:19.420 |
a year and a half, and infinitely cheap as well. 00:36:24.620 |
Like we just assume all of these things are gonna happen. 00:36:29.780 |
is to provide the input and output to the model. 00:36:32.340 |
I really compare it all the time to the PC and the CPU, 00:36:38.500 |
Well, now actually they do build the CPU as well, 00:36:40.420 |
but leaving that aside, they're busy building a laptop. 00:36:43.140 |
It's just a lot of work to build these things. 00:36:46.620 |
another person that we're close to, Mihaly from Repl.it, 00:36:59.180 |
you do need that and what situations you don't. 00:37:01.180 |
Obviously the simple answer is for coding, it helps. 00:37:10.980 |
No, no, no, the cognitive architecture changed. 00:37:14.100 |
For you, you one-shot and you chain tools together 00:37:23.260 |
I have some of my Lindys, I've told them like, 00:37:27.300 |
But that gives you a little bump also on use cases, 00:37:36.420 |
I don't know if many people were expecting computer use 00:37:40.060 |
Do these things make you rethink how to design 00:37:46.020 |
That's just like a small thing in their like AGI pursuit 00:37:49.140 |
that like maybe they're not even gonna support. 00:37:56.940 |
"Hey, look, why am I building all these API integrations 00:38:03.180 |
No, I mean, we did take into account computer use. 00:38:05.420 |
We were talking about this a year ago or something. 00:38:07.820 |
Like we've been talking about it as part of our roadmap. 00:38:16.060 |
My philosophy about it is anything that can be done 00:38:20.900 |
or should be done by an API for a very long time. 00:38:24.260 |
I think it is dangerous to be overly cavalier 00:38:35.100 |
And I can only assume that the conversation that went down 00:38:43.460 |
"And so that's all going to go to zero eventually." 00:39:02.460 |
in terms of how smooth the interactions were. 00:39:05.620 |
Android phones were significantly slower and laggier 00:39:18.460 |
when we're talking about API use versus computer use, 00:39:41.140 |
It's making like $2 billion a year or something. 00:39:43.780 |
All they need to do is add tools to the ChatGPT, 00:40:02.340 |
because it's like we can also use the plug-ins. 00:40:05.100 |
No, again, I think it's going to be such a huge market. 00:40:11.460 |
I know they have like a huge enterprise offering and stuff, 00:40:18.300 |
this sort of use cases that we're going after, 00:40:19.980 |
which is like, we're doing a lot of like lead generation 00:40:26.020 |
like Lindy Today right now joins your Zoom meetings 00:40:30.100 |
I don't see that so far on the OpenAI roadmap. 00:40:38.620 |
Cool, I have some other questions on company building stuff. 00:40:43.220 |
- It's a fascinating way to build a business, right? 00:40:45.260 |
Like what should you, as CEO, be in charge of? 00:40:48.980 |
And what should you basically hire a mini CEO to do? 00:40:53.540 |
I think that's all something we're figuring out. 00:40:55.380 |
The GM thing was inspired from my days at Uber, 00:40:58.380 |
where we hired one GM per city or per major geo area. 00:41:02.180 |
We had like all GMs, regional GMs and so forth. 00:41:08.940 |
to own each vertical and to go to market of the vertical 00:41:18.780 |
I mean, the canonical reply here is always going to be, 00:41:21.420 |
you own the fundraising, you own the culture, 00:41:24.380 |
you own the, what's the rest of the canonical reply? 00:41:30.780 |
- Even that eventually you do have to hand out. 00:41:33.500 |
Yes, the vision, the culture and the fundraising. 00:41:37.380 |
and you've done it well, you've done your job as a CEO. 00:41:40.500 |
I mean, all day, I do a lot of product work still 00:41:45.060 |
Obviously, like you're recording and managing the team, yeah. 00:41:48.220 |
- That one feels like the most automatable part of the job, 00:41:56.260 |
- Relationship between Factorio and building Lindy. 00:42:01.220 |
how the business of the future is like a game of Factorio. 00:42:06.380 |
and you've got your Lindy instance, it's like Slack, 00:42:08.700 |
and you've got like 5,000 Lindys in the sidebar 00:42:11.060 |
and your job is to somehow manage your 5,000 Lindys. 00:42:13.740 |
And it's going to be very similar to company building 00:42:16.140 |
because you're going to look for the highest leverage way 00:42:18.980 |
to understand what's going on in your AI company 00:42:31.740 |
you could have a meeting with your team and you're like, 00:42:36.540 |
And two months later, you have a new designer. 00:42:38.300 |
Now it's like, "Okay, boom, I'm going to spin up 00:42:41.820 |
- Like actually, it's more important that you can clone 00:42:46.700 |
Because the hiring process, you cannot clone someone. 00:42:51.340 |
is going to have their own tweaks and you don't want that. 00:43:04.460 |
Apparently a whole bunch of people stopped working. 00:43:09.140 |
But the other thing was you had a tweet recently 00:43:11.100 |
about how the sort of intentional top-down design 00:43:21.060 |
- I think people read it a little bit too much 00:43:25.380 |
I did not intend it as a giant statement on life. 00:43:27.860 |
- I mean, you notice you have a pattern of this, right? 00:43:33.980 |
- I legit was just hearing an interesting story 00:43:37.060 |
And everybody was like, "Oh my God, so deep." 00:43:38.500 |
I guess this explains everything about life and companies. 00:43:46.820 |
"People underestimate the extent to which moonshots 00:43:49.900 |
are just one pragmatic step taken after the other." 00:43:52.340 |
And I think as long as you have some inductive bias 00:43:54.300 |
about like some loose idea about where you want to go, 00:43:56.500 |
I think it makes sense to follow a sort of greedy search 00:44:10.140 |
When I tried out one of your automation templates 00:44:15.980 |
So like it was not as useful to me as a small one 00:44:25.500 |
and I just upfront stuffed everything I wanted to do 00:44:29.980 |
into my prompt and expected O1 to do everything. 00:44:43.140 |
So I threw away the code base, started small, 00:44:50.100 |
And to me, that was the factorial sentiment, right? 00:44:56.180 |
of something that you just randomly tweeted out. 00:45:00.380 |
for Lindy building, for coding, I don't know. 00:45:08.740 |
- There's the, I forgot the name of this other blog post 00:45:18.140 |
and people who optimize the system for its legibility. 00:45:23.540 |
Anytime you make a system more understandable 00:45:38.300 |
It's like you are actually optimizing for legibility. 00:45:41.860 |
but in some other cases, it may not make sense. 00:45:44.660 |
Sometimes it's better to leave the system alone 00:45:47.020 |
and let it be its glorious, chaotic, organic self 00:45:50.940 |
and just trust that it is going to perform well, 00:45:53.100 |
even though you don't understand it completely. 00:45:55.020 |
- It does remind me of a common managerial issue or dilemma, 00:45:59.780 |
which you experienced in a small scale of Lindy 00:46:01.780 |
where, you know, do you want to organize your company 00:46:07.740 |
or, you know, whatever the opposite of functional is. 00:46:14.500 |
but actually it stopped working at the small level. 00:46:35.220 |
And then the team one day came to me with pitchforks 00:46:51.060 |
- Another big change you had was going away from remote work, 00:47:04.220 |
Was there kind of like a threshold of employees 00:47:06.180 |
and team size where you felt like, okay, maybe that worked. 00:47:16.460 |
The business was about building a virtual office 00:47:19.460 |
And so being remote was not merely something we did. 00:47:25.020 |
because we were helping companies to go remote, right? 00:47:27.940 |
And so, frankly, in a way it's a bit embarrassing 00:47:32.060 |
but I guess when the facts changed, I changed my mind. 00:47:41.220 |
And on paper, the gains of remote are enormous. 00:47:50.140 |
Saving on commute is huge for everyone and so forth. 00:47:52.580 |
But then, look, I'm not going to say anything original here. 00:48:02.260 |
And my conclusion is at least we couldn't figure it out 00:48:13.500 |
I don't know that software can actually solve this problem. 00:48:16.260 |
Reality of it is everyone just wants to get off 00:48:19.820 |
And it's not a good feeling to be in your home office 00:48:22.780 |
if you even are lucky enough to have a home office all day. 00:48:28.660 |
I think software is peculiar because it's like an iceberg. 00:48:31.820 |
It's like the vast majority of it is submerged under water. 00:48:35.780 |
And so the quality of the software that you ship 00:48:38.460 |
is a function of the alignment of your mental models 00:48:43.860 |
about what it is exactly fundamentally that we're building? 00:48:47.940 |
And it is so much harder to get in sync about that 00:48:52.820 |
because people are offline and you can't get ahold of them 00:48:55.820 |
It's just, it's like you feel like you're walking 00:48:59.100 |
And eventually I just, I was like, okay, this is it. 00:49:09.020 |
But I still have a big, like one of my big heroes 00:49:16.500 |
but like these people run thousand person remote businesses. 00:49:25.980 |
Because if you go from one building to two buildings, 00:49:28.180 |
you're congrats, you're now remote from the other building. 00:49:32.660 |
to like two city offices, they're remote from each other. 00:49:36.460 |
Every time anyone talks about remote success stories, 00:49:39.860 |
I mean, it's always GitLab and WordPress and Zapier. 00:49:47.660 |
And I will point out that in every one of these examples, 00:49:52.860 |
that is sometimes orders of magnitude bigger. 00:50:01.140 |
and they're like a fraction of the size of even Substack. 00:50:09.580 |
Like look, GitLab is much smaller than GitHub. 00:50:17.820 |
because they wanted to move from San Francisco to LA. 00:50:19.940 |
So I think if you're optimizing for productivity, 00:50:23.100 |
if you really know, hey, this is a support ticket, right? 00:50:32.860 |
to offshore like the Philippines or whatever, 00:50:35.780 |
If you're optimizing for cost, absolutely be remote. 00:50:46.980 |
You want the very best album that you can make. 00:50:52.780 |
So the line is that all jobs that can be remote 00:50:58.980 |
and all jobs that are not remote are in person. 00:51:01.580 |
Like there's a very, very clear separation of jobs. 00:51:13.500 |
and in product defining and how to express that with LLMs. 00:51:19.020 |
You're definitely a, what I call a temperature zero 00:51:23.100 |
You want it to be reliable, predictable, small. 00:51:27.460 |
that are more for like creativity and engines, right? 00:51:31.460 |
but I'm pretty sure no one uses Lindy for brainstorming. 00:51:36.540 |
- I use Lindy for brainstorming a lot, actually. 00:51:40.220 |
you want to have like something that's anti-fragile 00:51:48.980 |
If it is about direction, like decide what to do, 00:51:52.780 |
If it is about magnitude and just do it as fast as possible, 00:51:59.780 |
software companies are not necessarily creative. 00:52:03.580 |
And I'll say that is going to come across the wrong way, 00:52:16.860 |
I don't mean to throw shade at them, like good for them. 00:52:21.820 |
that they have work-life balance on their job description. 00:52:31.380 |
And so we're just scratching our heads all day, 00:52:41.940 |
And there's a whole bunch of tough stuff in here. 00:52:44.100 |
I don't know if you have any particular leanings. 00:52:46.500 |
Probably the biggest one I'll just congratulate you on 00:52:50.980 |
Like you, very French, but your heart was sort of 00:52:54.700 |
in the US, you eventually found your way here. 00:52:57.100 |
What are your takes for like founders, right? 00:53:00.660 |
you wrote this post on like, "Go West, young man." 00:53:02.780 |
And now you've basically completed that journey, right? 00:53:04.820 |
Like you're now here and up to the point where 00:53:07.900 |
you're kind of mystified by how Europe has been so decel. 00:53:13.620 |
'cause I was making the prediction that Europe 00:53:15.780 |
was over 14 years ago or something like that. 00:53:18.380 |
I think it's been a walking corpse for a long time. 00:53:22.900 |
that it is paying the consequences of its policies 00:53:27.700 |
I wish I could rewrite the "Go West, young man" article, 00:53:40.580 |
you either lack judgment or you lack ambition. 00:53:48.980 |
And I was like, "Like I said, judgment or ambition." 00:53:55.500 |
than your company in life or your career in life. 00:53:58.740 |
But know that that's the trade-off you're making. 00:54:00.380 |
If you prioritize your career, you've got to be here. 00:54:02.740 |
- As a fellow European escapist, I grew up in Rome. 00:54:10.100 |
Well, I started my first company in Europe 10 years ago, 00:54:13.660 |
And yeah, you can tell nobody really wants to do much. 00:54:17.900 |
It's funny, I was looking back through some old tweets 00:54:20.340 |
and I was sending all these tweets to Mark Andreessen 00:54:22.460 |
like 15 years ago, trying to learn more about 00:54:25.900 |
why are you guys putting money in these things 00:54:28.300 |
that most people here would say you're crazy to even back. 00:54:35.620 |
And I think just like so many people in Europe 00:54:38.340 |
reach out and ask, "Hey, can you talk to our team?" 00:54:42.140 |
And they just cannot comprehend the risk appetite 00:54:46.620 |
It just like so foreign to people, at least in Italy 00:54:51.700 |
I'm sure there's some great founders in Europe, 00:54:55.980 |
like why would I leave my job at the post office 00:54:59.740 |
to go work on the startup that could change everything 00:55:02.700 |
and become very successful, but might go out of business. 00:55:06.740 |
we host a hackathon and it's like 400 people show up 00:55:14.500 |
and there's no incentives from the government 00:55:19.380 |
such a deep-rooted culture of like, you know, 00:55:25.820 |
So I don't really know how it's going to change. 00:55:35.300 |
It's just like, hey, like there's no rational explanation 00:55:40.180 |
It just, if you want to do this job and do this, 00:55:56.700 |
So I should get my U.S. citizenship interview. 00:56:09.660 |
They've decided to say no to capitalism a long time ago. 00:56:17.780 |
is a little bit of a self-fulfilling prophecy 00:56:33.140 |
there was a movement of people moving to Miami. 00:56:39.860 |
You can't break the network effect, you know? 00:56:41.580 |
- It's so annoying because first principles wise, 00:56:49.100 |
- San Francisco does not want tech to be here. 00:56:54.940 |
- I think the people that are in San Francisco 00:57:01.380 |
But I would say people in Miami would hate it too 00:57:05.340 |
Like the Niki Beach crowd would also not chill. 00:57:08.380 |
- They're just rich enough and chill enough to not care. 00:57:23.860 |
I'll maybe carve out that I think the UK has done really well. 00:57:27.060 |
That's an argument for the UK not being part of Europe 00:58:25.020 |
- It remains to be seen how that move is going. 00:59:05.980 |
- I don't want you to focus on any particular thing 00:59:08.660 |
- You know, well, is that tweet in particular, 00:59:18.180 |
and I received a lot of people reaching out like, 00:59:20.740 |
I think it's coming from a few different places. 00:59:24.860 |
Like I don't feel like self-censoring all the time. 00:59:32.340 |
you never know what everyone, what anyone thinks. 00:59:34.420 |
And so it's becoming like a self-perpetuating thing. 00:59:36.860 |
It's like a public lies, private truth sort of phenomenon. 00:59:40.860 |
there's this phenomenon called a preference cascade. 00:59:44.100 |
It's like, oh, there's only one communist left in USSR. 00:59:49.340 |
because everyone else pretends to be a communist. 00:59:52.180 |
for someone to have backbone and just be like, 00:59:56.620 |
especially when you are like me in a position 00:59:58.740 |
where it's like, I don't have a boss who's going to fire me. 01:00:06.980 |
Right, like there's really not that much downside. 01:00:10.700 |
about standing up for what you think is right 01:00:19.380 |
for your political beliefs and free speech or whatever, 01:00:22.020 |
and the way that you think about business too. 01:00:27.100 |
- I think the world contrarian has become abused, 01:00:38.660 |
because you're not even thinking in your other way 01:00:42.460 |
to adopt the same beliefs as people around you. 01:00:46.380 |
and everyone else is doing it except themselves. 01:00:50.820 |
- That's right, and so I actually make it a point 01:00:53.060 |
to like look for, hey, what would be a thing right now 01:01:02.460 |
And so I think the AI safety is an example of that. 01:01:04.540 |
Like at some point, Mark Andresen blocked me on Twitter 01:01:17.420 |
- Mark Andresen was really my booster initially on Twitter. 01:01:29.100 |
that actually came out in support of the bill or something. 01:01:33.380 |
A year and a half ago, I put some tweet storms 01:01:38.220 |
And yeah, I was blocked by a bunch of A6NZ people 01:02:05.740 |
Like, look, the reality of it is 95% of the country 01:02:08.220 |
did not resist and most of it actually collaborated 01:02:20.620 |
even if I've gotten into physical fights in my life, 01:02:24.300 |
like in SF, because some people got attacked. 01:02:30.340 |
You get involved and you help the person, right? 01:02:32.260 |
And so, look, I'm not pretending we're like nowhere near 01:02:36.460 |
but I'm like exactly because we are nowhere near 01:02:41.620 |
for what you think is right when the stakes are so low, 01:02:49.500 |
But here in the US, I'm always like, "Oh man." 01:02:52.500 |
- I feel, I detect some inconsistency in your statements 01:02:55.540 |
because you simultaneously believe that AGI is very soon. 01:03:03.740 |
so why does AGI make the stakes of speaking up higher? 01:03:12.060 |
- Oh no, the stakes of AI safety couldn't be higher. 01:03:13.900 |
I meant the stakes of like speaking up about- 01:03:19.220 |
- How do you figure out who's real and who isn't? 01:03:23.700 |
for responsible AI that like hundreds of like VCs 01:03:30.300 |
- Was that the pause letter, like six month pause, or? 01:03:32.540 |
- Some, no, there was like something else too 01:03:34.340 |
that I think general catalyst and like some fun sign, 01:03:37.020 |
but, and then there's maybe the anthropic case, 01:03:41.860 |
because you guys don't take security seriously." 01:03:43.860 |
And then it's like, "Hey, what if we gave AI access 01:03:46.380 |
to a whole computer to just like go do things?" 01:03:52.860 |
I mean, you could say the same thing about Lindy. 01:03:54.700 |
It's like, if you're worried about AI safety, 01:03:59.340 |
How do you internally decide between participation 01:04:08.900 |
and building actually makes it safer because I'm involved," 01:04:11.980 |
versus just being like anti, "I think this is unsafe," 01:04:16.700 |
and just kind of remove yourself from the whole thing, 01:04:19.740 |
- Yeah, the way I think about our own involvement here 01:04:22.420 |
is I'm acutely concerned about the risks at the model layer. 01:04:26.980 |
And I'm simultaneously very excited about the upside. 01:04:32.340 |
insofar as I can quantify it, which I cannot, 01:04:50.140 |
that we live in a utopia where there's no disease 01:04:53.740 |
I think that utopia is going to happen through, 01:04:56.100 |
like, again, I'm bringing my little contribution 01:04:58.980 |
I think it would be silly to say no to the upside 01:05:07.900 |
"Oh, you know, like the downside doesn't exist 01:05:09.540 |
at my layer, it exists at like the model layer." 01:05:11.700 |
But truly, look at Lindy, look at the Apple building. 01:05:15.380 |
I struggle to see exactly how it would like get up 01:05:20.820 |
- Okay, well, this kind of discussion can go on for hours. 01:05:23.340 |
It is still daylight, so not the best time for it, 01:05:26.100 |
but I really appreciate you spending the time. 01:05:49.020 |
Yeah, we are hiring designers and engineers right now. 01:06:03.740 |
- Part of that is I don't have to know, right? 01:06:07.500 |
because we also have a teammate whose name is Lindy. 01:06:23.060 |
just because I think that there are a lot of AI companies 01:06:30.060 |
and the models and the capabilities and the benchmark. 01:06:44.780 |
One, at the lowest level, it's make this look pretty, 01:06:47.220 |
make this look like Stripe or Linear's homepage. 01:06:52.260 |
it is make this integrate seamlessly into my life. 01:06:54.300 |
Intuitive, beautiful, inspirational, maybe even. 01:07:00.260 |
this is kind of like a blog post I've been thinking about, 01:07:03.340 |
actually are going to win more than companies that don't. 01:07:09.500 |
"Design is the expression of the soul of a man-made product 01:07:16.260 |
- He was cooking, he was cooking on that one. 01:07:23.780 |
to reach alignment about that soul of the product, right? 01:07:26.980 |
like you peel the onion in those layers, right? 01:07:30.860 |
just like the user experiencing your product chronologically 01:07:33.900 |
all the way from the beginning of like the awareness stage,