(upbeat music) - Hey everyone. Welcome to the Lydian Space Podcast. This is Alessio, partner and CTO at Decibel Partners, and I'm joined by my co-host, Swix, founder of Small AI. - Hey, and today we're joined in the studio by Florent Crevello. Welcome. - Hey, yeah, thanks for having me.
- Also known as Altimore. Always wanted to ask, what is Altimore? - It was the name of my character when I was playing Dungeons and Dragons. - Always. - I was like 11 years old. - What was your classes? - I was an elf. I was a magician elf.
- Okay, all right. Well, you're still spinning magic. Right now you're a solo founder/CEO of Lindy.ai. What is Lindy? - Yeah, we are a no-code platform letting you build your own AI agents easily. So you can think of we all to link chain as air table into MySQL. Like you can just pin up AI agents super easily by clicking around and no code required.
You didn't have to be an engineer and you can automate business workflows that you simply could not automate before in a few minutes. - You've been in our orbit a few times. I think you spoke at our Latent Space anniversary. You spoke at my summit, the first summit, which was a really good keynote.
And most recently, like we actually already scheduled this podcast before this happened, but Andrew Wilkinson was like, "I'm obsessed by Lindy." He's just created a whole bunch of agents. So basically, why are you blowing up? - Well, thank you. I think we are having a little bit of a moment.
I think it's a bit premature to say we're blowing up, but why are things going well? We revamped the product majorly. We called it Lindy 2.0. I would say we started working on that six months ago. We've actually not really announced it yet. It's just, I guess, I guess that's what we're doing now.
(laughs) And so we've basically been cooking for the last six months, like really rebuilding the product from scratch. I think, Alessio, actually, the last time you tried the product, it was still Lindy 1.0. - Oh yeah, it was. - If you log in now, the platform looks very different.
There's like a ton more features. And I think one realization that we made, and I think a lot of folks in the agent space made the same realization, is that there is such a thing as too much of a good thing. I think many people, when they started working on agents, they were very LLM-peeled and chat GPT-peeled, right?
They got ahead of themselves in a way, as included, and they thought that agents were actually, and LLMs were actually more advanced than they actually were. And so the first version of Lindy was like just a giant prompt and a bunch of tools. And then the realization we had was like, hey, actually, the more you can put your agent on Rails, one, the more reliable it's going to be, obviously, but two, it's also going to be easier to use for the user, because you can really, as a user, you get, instead of just getting this big, giant, intimidating text field, and you type words in there, and you have no idea if you're typing the right words or not, here you can really click and select step-by-step and select, tell your agent what to do, and really give as narrow or as wide a guardrail as you want for your agent.
We started working on that. We called it Lindy on Rails about six months ago. And we started putting it into the hands of users over the last, I would say, two months or so. And that's, I think things really started going pretty well at that point. The agent is way more reliable, way easier to set up, and we're already seeing a ton of new use cases pop up.
FRANCESC CAMPOY: Yeah. Just a quick follow-up on that. You launched the first Lindy in November last year, and you were already talking about having a DSL, right? I remember having this discussion with you, and you were like, it's just much more reliable. Is it still the DSL under the hood?
Is this a UI-level change, or is it a bigger rewrite? FRANCESC CAMPOY: No, it is a much bigger rewrite. I'll give you a concrete example. Suppose you want to have an agent that observes your Zendesk tickets, OK? And it's like, hey, every time you receive a Zendesk ticket, I want you to check my knowledge base-- so it's like a RAG module and whatnot-- and then answer the ticket.
The way it used to work with Lindy before was you would type the prompt asking it to do that. Every time you receive a Zendesk ticket, you check my knowledge base, and so on and so forth. The problem with doing that is that it can always go wrong. Like, you're praying the LLM gods that they will actually invoke your knowledge base, but I don't want to ask it.
I want it to always, 100% of the time, consult the knowledge base after it receives a Zendesk ticket. And so with Lindy, you can actually have the trigger, which is Zendesk ticket received, have the knowledge base consult, which is always there, and then have the agent. So you can really set up your agent any way you want like that.
FRANCESC CAMPOY: This is something I think about for AI engineering as well, which is the big labs want you to hand over everything in the prompts and only code of English. And then the smaller brains, the GPU pourers, always want to write more code to make things more deterministic, and reliable, and controllable.
One way I put it is put Shoggoth in a box and make it a very small, like the minimal viable box. Everything else should be traditional, if this, then that software. FRANCESC CAMPOY: I love that characterization, put the Shoggoth in the box. Yeah, we talk about using as much AI as necessary and as little as possible.
OK. YOSSI ELKRIEF: And what was the choosing between this drag and drop, low code, whatever, super code-driven, maybe like the of the world? And maybe the flip side of it, which you don't really do, it's just text to agent. It's like build the workflow for me. Whatever you learn, actually putting this in front of users and figuring out how much do they actually want to edit versus how much-- kind of like Ruby on Rails, instead of Lindy on Rails, it's kind of like defaults over configuration.
FRANCESC CAMPOY: Yeah. I actually used to dislike when people said, oh, text is not a great interface. I was like, ah, this is such a mid-take. I think text is awesome. And I've actually come around. I actually sort of agree now that text is really not great. I think for people like you and me, because we sort of have a mental model, OK, when I type a prompt into this text box, this is what it's going to do.
It's going to map it to this kind of data structure under the hood and so forth. I guess it's a little bit blackmailing towards humans. You jump on these calls with humans, and you're like, here's a text box. This is going to set up an agent for you. Do it.
And then they type words like, I want you to help me put order in my inbox. Or actually, this is a good one. This is actually a good one. Well, that's a bad one. I would say 60% or 70% of the prompts that people type don't mean anything. Me as a human, as AGI, I don't understand what they mean.
I don't know what they mean. It is actually, I think, whenever you can have a GUI, it is better than to have just a pure text interface. And then how do you decide how much to expose? So even with the tools, you have a bunch of them. You have Slack, you have Google Calendar, you have Gmail.
Should people, by default, just turn over access to everything, and then you help them figure out what to use? I think that's the question. Because when I tried to set up Slack, it was like, hey, give me access to all channels and everything. Which, for the average person, probably makes sense, because you don't want to re-prompt them every time to add new channels.
But at the same time, for maybe the more sophisticated enterprise use cases, people are like, hey, I want to really limit what you have access to. How do you thread that balance? The general philosophy is we ask for the least amount of permissions needed at any given moment. I don't think Slack-- I could be mistaken, but I don't think Slack lets you request permissions for just one channel.
But, for example, for Google, obviously there's hundreds of scopes that you could require for Google. There's a lot of scopes. And sometimes it's actually painful to set up your Lindy, because you're going to have to ask to Google and add scopes five or six times. Like, we've had sessions like this.
But that's what we do, because, for example, the Lindy email drafter, she's going to ask you for your authorization once for, I need to be able to read your email so I can draft a reply. And then another time for, I need to be able to write a draft for them.
So we just try to do it very incrementally like that. FRANCESC CAMPOY: Yeah. Do you think OAuth is just overall going to change? I think maybe before it was like, hey, we need to set up an OAuth that humans only want to kind of do once. So we try to jam-pack things all at once, versus what if you could, on-demand, get different permissions every time from different parts?
Like, do you ever think about designing things knowing that maybe AI will use it instead of humans will use it? FRANCESC CAMPOY: Yeah, for sure. One pattern we've started to see is people provisioning accounts for their AI agents. And so in particular, Google Workspace accounts. So, for example, Lindy can be used as a scheduling assistant.
And so you can just cc her to your emails when you're trying to find time with someone. And just like a human assistant, she's going to go back and forth in a flow of abilities and so forth. Very often, people don't want the other party to know that it's an AI.
So it's actually funny, they introduce delays. They ask the agent to wait before replying, so it's not too obvious that it's an AI. And they provision an account on Google Suite, which costs them like $10 a month or something like that. So we're seeing that pattern more and more.
I think that does the job for now. I'm not optimistic on us actually patching OAuth, because I agree with you ultimately. We would want to patch OAuth, because the new account thing is kind of a kludge. It's really a hack. You would want to patch OAuth to have more granular access control and really be able to put your showbiz in the box.
I'm not optimistic on us doing that before AGI, I think. FRANCESC CAMPOY: That's a very close timeline. I'm mindful of talking about a thing without showing it. And we already have the setup to show it. We jump into a screen share. For listeners, you can jump on to YouTube and like and subscribe.
But also, let's have a look at how you show off Lindy. ALAIN VONGSOUVANH: Yeah, absolutely. I'll give an example of a very simple Lindy, and then I'll graduate to a much more complicated one. A super simple Lindy that I have is I unfortunately bought some investment properties in the south of France.
It was a really, really bad idea. And I put them on the Holydew, which is like the French Airbnb, if you will. And so I received these emails from time to time telling me, like, oh, hey, you made $200. Someone booked your place, OK? When I receive these emails, I want to log this reservation in a spreadsheet.
Doing this without an AI agent or without AI in general is a pain in the butt, because you must write an HTML parser for this email. And so it's just hard. You may not be able to do it, and it's going to break the moment the email changes. By contrast, the way it works with Lindy, it's really simple.
It's two steps. It's like, OK, I receive an email. If it is a reservation confirmation-- I have this filter here-- then I append a row to this spreadsheet. And so this is where you can see the AI part, where the way this action is configured here-- you see these purple fields on the right.
Each of these fields is a prompt. And so I can say, OK, you extract from the email the day the reservation begins on. You extract the amount of the reservation. You extract the number of travelers of the reservation. And now you can see, when I look at the task history of this, Lindy, it's really simple.
It's like, OK, you do this. And boom, I'm appending this row to this spreadsheet. And this is the information extracted. So effectively, this node here, this append row node, is a mini-agent. It can see everything that just happened. It has context over the task, and it's appending the row.
And then it's going to send a reply to the thread. That's a very simple example of an agent. Quick follow-up question on this one while we're still on this page. Is that one call? Is that a structured output call? Yeah. OK, nice. Yeah. And you can see here, for every node, you can configure which model you want to power the node.
Here, I use Cloud. For this, I use GPT for Turbo. Much more complex example, my meeting recorder. It looks very complex, because I've added to it over time. But at a high level, it's really simple. It's like, when a meeting begins, you record the meeting. And after the meeting, you send me a summary.
And it sends me coaching notes. So I receive-- like, my Lindy is constantly coaching me. And so you can see here, in the prompt of the coaching notes, I've told it, hey, was I unnecessarily confrontational at any point? I'm French, so I have to watch out for that. Or not confrontational enough.
Should I have double-clicked on any issue? So I can really give it exactly the kind of coaching that I'm expecting. And then the interesting thing here is, you can see, the agent here, after it sends me these coaching notes, moves on. And it does a bunch of other stuff.
So it goes on Slack. It disseminates the notes on Slack. It does a bunch of other stuff. But it's actually able to backtrack and resume the automation at the coaching notes email if I responded to that email. So I'll give a super concrete example. This is an actual coaching feedback that I received from Lindy.
She was like, hey, this was a sales call I had with a customer. And she was like, I found your explanation of Lindy too technical. And I was able to follow up and just ask a follow-up question in the thread here. And I was like, why did you find too technical about my explanation?
And Lindy restored the context. And so she basically picked up the automation back up here in the tree. And she has all of the context of everything that happened, including the meeting in which I was. So she was like, oh, you used the words "deterministic" and "context window" and "agent state." And that concept exists at every level for every channel and every action that Lindy takes.
So another example here is, I mentioned, she also disseminates the notes on Slack. So this was a meeting where I was not, right? So this was a teammate. He's an Indie meeting recorder, posts the meeting notes in this customer discovery channel on Slack. So you can see, okay, this is the onboarding call we had.
This was the use case. Look at the questions. How do I make Lindy slower? How do I add delays to make Lindy slower? And I was able, in the Slack thread, to ask follow-up questions like, oh, what did we answer to these questions? And it's really handy because I know I can have this sort of interactive Q&A with these meetings.
It means that very often now, I don't go to meetings anymore. I just send my Lindy, and instead of going to a 60-minute meeting, I have a five-minute chat with my Lindy afterwards. And she just replied. She was like, well, this is what we replied to this customer, and I can just be like, okay, good job, Jack.
No notes about your answers. So that's the kind of use cases people have with Lindy. It's a lot of, there's a lot of sales automations, customer support automations, and a lot of this, which is basically personal assistance automations, like meeting scheduling and so forth. - Yeah, and I think the question that people might have is memory.
So as you get coaching, how does it track whether or not you're improving? You know, if these are like mistakes you made in the past, like, how do you think about that? - Yeah, we have a memory module. So I'll show you my meeting scheduler, Lindy, which has a lot of memories because by now I've used her for so long.
And so every time I talk to her, she saves a memory. If I tell her, you screwed up, please don't do this. So you can see here, it's, oh, it's got a double memory here. This is the meeting link I have, or this is the address of the office.
If I tell someone to meet me at home, this is the address of my place. This is the code. I guess we'll have to edit that out. (all laughing) This is not the code of my place. No doxing. (all laughing) Yeah, so Lindy can just like manage her own memory and decide when she's remembering things between executions.
- Okay, I mean, I'm just gonna take the opportunity to ask you, since you are the creator of this thing, how come there's so few memories, right? Like if you've been using this for two years, there should be thousands of thousands of things. - That is a good question.
Agents still get confused if they have too many memories, to my point earlier about that. So I just am out of a call with a member of the Lama team at Meta, and we were chatting about Lindy, and we were going into the system prompt that we sent to Lindy and all of that stuff.
And he was amazed, and he was like, it's a miracle that it's working, guys. He was like, this kind of system prompt, this does not exist either pre-training or post-training. Like these models were never trained to do this kind of stuff. Like it's a miracle that they can be agents at all.
And so what I do, I actually prune the memories. - You know, it's actually something I've gotten into the habits of doing from back when we had GPT 3.5 being Lindy agents. I suspect it's probably not as necessary in like the cloud 3.5 sunnet days, but I prune the memories, yeah.
- Yeah, okay. The reason is 'cause I have another assistant that also is recording and trying to come up with facts about me. It comes up with a lot of like trivial, useless facts that I, so I spend most of my time pruning. It actually is not super useful.
I'd much rather have high quality facts that it accepts. Or maybe I was even thinking like, were you ever tempted to add a wake word to only memorize this when I say memorize this? - Yeah. - And otherwise don't even bother. - I have a Lindy that does this.
So this is my inbox processor Lindy. It's kind of beefy because there's a lot of different emails, but somewhere in here, there is a rule where I'm like, aha, I can email my inbox processor Lindy. It's really handy. So she has her own email address. And so when I process my email inbox, I sometimes forward an email to her and it's a newsletter, or it's like a cold outreach from a recruiter that I don't care about or anything like that.
And I can give her a rule and I can be like, hey, this email I want you to archive moving forward. Or I want you to alert me on Slack when I have this kind of email, it's really important. And so you can see here, the prompt is, if I give you a rule about a kind of email, like archive emails from X, save it as a new memory.
And I give it to the memory saving skill. And yeah. - One thing that just occurred to me. So I'm a big fan of virtual mailboxes. I recommend that everybody have a virtual mailbox. You could set up a physical mail receive thing for Lindy. And so then people can just, then Lindy can process your physical mail.
- That's actually a good idea. I actually already have something like that. I use like Else Class Mail. Yeah. So yeah, most likely I can process my physical mail. - And then the other products idea I have looking at this thing is people want to brag about the complexity of their Lindys.
So this would be like a 65 point Lindy, right? - What's a 65 point? - Complexity counting. Like how many nodes, how many things, how many conditions, right? - Yeah, this is not the most complex one. I have another one. This designer recruiter here is kind of beefy as well.
- Right, right, right. So I'm just saying like, let people brag. Let people like be super users. - Oh, right. Give them a score or something. - Give them a score. Then they'll just be like, okay, how high can you make this score? - Yeah, that's a good point.
And I think that's against the beauty of this on-rails phenomenon. It's like, think of the equivalent, the prompt equivalent of this Lindy here, for example, that we're looking at. It'd be monstrous. And the odds that it gets it right are so low. But here, because we're really holding the agent's hand step-by-step-by-step, it's actually super reliable.
- Yeah. And is it all structured output base? - Yeah. - As far as possible? - Basically. - Like there's no non-structured output. - There is. So for example, here, this like AI agent step, right? Or this like send message step. Sometimes it gets to- - That's just plain text.
- That's right. - Yeah. - So I'll give you an example. Maybe it's TMI. I'm having blood pressure issues these days. And so I'm, this Lindy here, I give it my blood pressure readings and it updates a log that I have of my blood pressure that it sends to my doctor.
- Oh, so this is a, every Lindy comes with a to-do list? - Yeah. Every Lindy has its own task history. - Huh. - Yeah. And so you can see here, this is my main Lindy, sort of like my personal assistant. And I've told it, where is this? There is a point where I'm like, if I am giving you a health-related fact, right here.
I'm giving you a health information. So then you update this log that I have in this Google Doc, and then you send me a message. And you can see, I've actually not configured this send message node. I haven't told it what to send me a message for. Right? And you can see, it's actually lecturing me.
It's like, I'm giving it my blood pressure readings. It's like, hey, it's a bit high. Like, here are some lifestyle changes you may want to consider. - I think maybe this is the most confusing or new thing for people. So even I use Lindy and I didn't even know you could have multiple workflows in one Lindy.
I think the mental model is kind of like the Zapier. Workflows is like, it starts and it ends. It's not, doesn't choose between. How do you think about what's a Lindy versus what's a sub-function of a Lindy? Like, what's the hierarchy? - Yeah, frankly, I think the line is a little arbitrary.
It's kind of like when you code, like when do you start to create a new class versus when do you overload your current class? I think of it in terms of like jobs to be done. And I think of it in terms of who is the Lindy serving. This Lindy is serving me personally.
It's really my day-to-day Lindy. I give it a bunch of stuff, like very easy tasks. And so this is just a Lindy I go to. Sometimes when a task is really more specialized, so for example, I have this like summarizer Lindy or this designer recruiter Lindy, these tasks are really beefy.
I wouldn't want to add this to my main Lindy, so I just created a separate Lindy for it. Or when it's a Lindy that serves another constituency, like our customer support Lindy, I don't want to add that to like my personal assistant. These are two very different Lindys. - Yeah.
And you can call a Lindy from within another Lindy. - That's right. - You can kind of chain them together. - Lindys can work together, absolutely. - A couple more things for the video portion. I noticed you have a podcast follower. We have to ask about that. What is that?
- So this one wakes me up every, so wakes herself up every week. And she sends me, so she woke up yesterday actually, and she searches for Lenny's podcast. And she looks for like the latest episode on YouTube. And once she finds it, she transcribes the video. And then she sends me the summary by email.
I don't listen to podcasts as much anymore. I just like read these summaries. - Yeah, yeah. - We should make a "Late in Space" Lindy. Marketplace. - Okay, so, and then, you know, you have a whole bunch of connectors. I saw the list briefly. Any interesting one, complicated one that you're proud of?
Anything that you want to just share? - Yeah. - Connector stories. - So many of our workflows are about meeting scheduling. So we had to build some very open unity tools around meeting scheduling. So for example, one that is surprisingly hard is this find available times action. You would not believe, this is like a thousand lines of code or something.
It's just a very beefy action. And you can pass it a bunch of parameters about how long is the meeting? When does it start? When does it end? What are the meeting, like the weekdays in which I meet? There's like, how many time slots do you return? What's the buffer between my meetings?
It's just a very, very, very complex action. I really like our GitHub action. So we have like a Lindy PR reviewer. And it's really handy because anytime any bug happens, so the Lindy reads our guidelines on a Google Docs. By now the guidelines are like 40 pages long or something.
And so every time any new kind of bug happens, we just go to the guideline and we add the lines like, "Hey, this has happened before. Please watch out for this category of bugs." And it's saving us so much time every day. - There's companies doing PR reviews. Where does a Lindy start?
When does a company start? Or maybe how do you think about the complexity of these tasks when it's going to be worth having kind of like a vertical standalone company versus just like, "Hey, a Lindy is going to do a good job 99% of the time." - That's a good question.
We think about this one all the time. I can't say that we've really come up with a very crisp articulation of when do you want to use a vertical tool versus when do you want to use a horizontal tool. I think of it as very similar to the internet.
I find it surprising the extent to which a horizontal search engine has won. But I think that Google, right? But I think the even more surprising fact is that the horizontal search engine has won in almost every vertical, right? You go through Google to search Reddit. You go through Google to search Wikipedia.
I think maybe the biggest exception is e-commerce. Like you go to Amazon to search e-commerce, but otherwise you go through Google. And I think that the reason for that is because search in each vertical has more in common with search than it does with each vertical. And search is so expensive to get right, and Google is a big company, that it makes a lot of sense to aggregate all of these different use cases and to spread your R&D budget across all of these different use cases.
I have a thesis, which is a really cool thesis for Lindy, is that the same thing is true for agents. I think that by and large, in a lot of verticals, agents in each vertical have more in common with agents than they do with each vertical. I also think there are benefits in having a single agent platform because that way your agents can work together.
They're all under one roof. That way you only learn one platform, and so you can create agents for everything that you want, and you don't have to pay for a bunch of different platforms and so forth. So I think ultimately it is actually going to shake out in a way that is similar to search in that search is everywhere on the internet.
Every website has a search box, right? So there's going to be a lot of vertical agents for everything. I think AI is going to completely penetrate every category of software, but then I also think there are going to be a few very, very, very big horizontal agents that serve a lot of functions for people.
- Yeah, that is actually one of the questions that we had about the agent stuff. So I guess we can transition away from the screen and I'll just ask the follow-up, which is, that is a hot topic. You're basically saying that the current VC obsession of the day, which is vertical AI-enabled SaaS, is mostly not going to work out, and then there are going to be some super giant horizontal SaaS?
- Oh no, I'm not saying it's either or. Like SaaS today, vertical SaaS is huge, and there's also a lot of horizontal platforms. If you look at like Airtable or Notion, basically the entire no-code space is very horizontal. I mean, Loom and Zoom and Slack, like there's a lot of very horizontal tools out there.
- Okay. (laughs) I was just trying to get a reaction out of you for hot takes. - Trying to get a hot take. No, I also think it is natural for the vertical solutions to emerge first, 'cause it's just easier to build. It's just much, much, much harder to build something horizontal.
- Cool, some more Lindy-specific questions. So we covered most of the top use cases, and you have an Academy. That was nice to see. I also see some other people doing it for you for free. So like Ben Spites is doing it, and then there's some other guy who is also doing like lessons.
- Yeah. - Which is kinda nice, right? - Yeah, absolutely. - You don't have to do any of that. - Oh, we're even seeing it more and more on like LinkedIn and Twitter, like people posting their Lindys and so forth. - Yeah, I think that's the flywheel, that you built the platform where creators see value in aligning themselves to you, and so then your incentive is to make them successful so that they can make other people successful, and then it just drives more and more engagement that you are, like it's earned media, like you don't have to do anything.
- Yeah, yeah, I mean, community is everything. - Are you doing anything special there, any big wins? - We have a Slack community that's pretty active. I can't say we've invested much more than that so far. - I would say from having, so I have some involvement in the no-code community.
I would say that Webflow going very hard after no-code as a category got them a lot more allies than just the people using Webflow. So it helps you to grow the community beyond just Lindy. And I don't know what this is called. Maybe it's just no-code again. Maybe you want to call it something different, but there's definitely an appetite for this, and you are one of a broad category, right?
Like just before you, we had Duston, and they're also kind of going after a similar market. Zapier obviously is not going to try to also compete with you. - Yeah. - There's no question there, it's just like a reaction about community. Like I think a lot about community, Linspace is growing the community of AI engineers, and I think you have a slightly different audience of, I don't know what.
- Yeah, I think the no-code tinkerers is the community. Yeah, it is going to be the same sort of community as what Webflow, Zapier, Airtable, Notion to some extent. - Yeah, the framing can be different if you were, so I think tinkerers has this connotation of not serious or like small.
And if you framed it to like no-code EA, we're exclusively only for CEOs with a certain budget, then you just have, you tap into a different budget. - That's true. The problem with EA is like the CEO has no willingness to actually tinker and play with the platform. - Maybe Andrew's doing that.
Like a lot of your biggest advocates are CEOs, right? - A solopreneur, you know, small business owners, I think Andrew is an exception, yeah. - Yeah, yeah, he is. He's an exception in many ways. - Yep. - Just before we wrap on the use cases, is Rick rolling your customers, like a officially supported use case, or maybe tell that story?
- It's one of the main jobs to be done, really. Yeah, we woke up recently, so we have a Lindy obviously doing our customer support, and we do check after the Lindy. And so we cut this email exchange where someone was asking Lindy for video tutorials. And at the time, actually, we did not have video tutorials, we do now on the Lindy Academy.
And Lindy responded to the emails like, "Oh, absolutely, here's a link." And we were like, "What, like we don't, what kind of link did you send?" And so we clicked on the link and it was a recall. We actually reacted fast enough that the customer had not yet opened the email, and so we reacted immediately like, "Oh, hey, actually, sorry, this is the right link." And so the customer never reacted to the first link.
And so, yeah, I tweeted about that, it went surprisingly viral. And I checked afterwards in the logs, we did like a database query, and we found like, I think, like three or four other instances of it. - That's surprisingly low. - Yeah, it is, it is low. And we fixed it across the board by just adding a line to the system prompt.
That's like, "Hey, don't recall people, please don't recall." - Yeah, yeah, yeah, yeah. I mean, so you can explain it retroactively, right? Like that YouTube slug has been pasted in so many different corpuses that obviously it learned to hallucinate that. - And it pretended to be so many things.
That's the thing, it's like everybody- - I wouldn't be surprised if that takes one token. Like there's a tokenizer and it's just one token. - That's the idea of a YouTube video. - Because it's used so much, right? Like, and you have to basically get it exactly correct. It's probably not.
I mean, that's a long- - It would have been so good. It is not a single token. (both laughing) - So this is just a jump maybe into evals from here. How could you possibly come up for an eval that says, "Make sure my AI does not rickroll my customer." I feel like when people are writing evals, that's not something that they come up with.
So how do you think about evals when it's such like an open-ended problem space? - Yeah, it is tough. We built quite a bit of infrastructure for us to create evals in one click from any conversation history. So we can point to a conversation and we can be like, "In one click, we can turn it into effectively a unit test." It's like, "This is a good conversation.
This is how you're supposed to handle things like this." Or if it's a negative example, then we modify a little bit the conversation after generating the eval. So it's very easy for us to spin up this kind of eval. - Do you use an off-the-shelf tool, which is at Brain Trust on the podcast, or did you just build your own?
- We built, we unfortunately built our own. We're most likely going to switch to Brain Trust. It's, well, when we built it, there was nothing. Like, there was no eval tool, frankly. And we, I mean, we started this project like end of 2022. It was like, it was very, very, very early.
I wouldn't recommend it to build your own eval tool. There's better solutions out there and our eval tool breaks all the time and it's a nightmare to maintain. And that's not something we want to be spending our time on. - I was going to ask that basically, 'cause I think my first conversations with you about Lindy was that you had a strong opinion that everyone should build their own tools.
And you were very proud of your evals. You're kind of showing off to me like how many evals you were running, right? - Yeah, I think that was before all of these tools came around. I think the ecosystem has matured a fair bit. - What is one thing that Braintrust has nailed that you always struggled to do?
- Well, not using them yet, so I couldn't tell. But from what I've gathered from the conversations I've had, like they're doing what we do with our eval tool, but better. - Yeah, and like they do it, but also like 60 other companies do it, right? So I don't know how to shop apart from brand.
- Yeah. - Word of mouth. - Same here. - Yeah, like evals, there's two kinds of evals, right? In some way, you don't have to eval your system as much because you've constrained the language model so much. And you can rely on open AI to guarantee that the structured outputs are going to be good, right?
We had Michelle sit where you sit and she explained exactly how they do constraint grammar sampling and all that good stuff. So actually, I think it's more important for your customers to eval their Lindys than you evaling your Lindy platform 'cause you just built the platform. You don't actually need to eval that much.
- Yeah, in an ideal world, our customers don't need to care about this. And I think the bar is not like, look, it needs to be at 100%. I think the bar is it needs to be better than a human. And for most use cases we serve today, it is better than a human, especially if you put it on Rails.
- Is there a limiting factor of Lindy at the business? Like, is it adding new connectors? Is it adding new node types? Like how do you prioritize what is the most impactful to your company? - Yeah, the raw capabilities for sure are a big limit. It is actually shocking the extent to which the model is no longer the limit.
It was the limit a year ago. It was too expensive. The context window was too small. It's kind of insane that we started building this when the context windows were like 4,000 tokens. Like today our system prompt is more than 4,000 tokens. So yeah, the model is actually very much not a limit anymore.
It almost gives me pause because I'm like, I want the model to be a limit. And so no, the integrations are ones, the core capabilities are ones. So for example, we are investing in a system that's basically, I call it like the, it's a Jayhack, gave me these names, like the Paul Mann's RLHF.
So you can turn on a toggle on any step of your Lindy workflow to be like, ask me for confirmation before you actually execute this step. So it's like, hey, I receive an email, you send a reply, ask me for confirmation before actually sending it. And so today you see the email that's about to get sent and you can either approve, deny, or change it and then approve.
And we are making it so that when you make a change, we are then saving this change that you're making or embedding it in a vector database. And then we are retrieving these examples for future tasks and injecting them into the context window. So that's the kind of capability that like makes a huge difference for users.
That's the bottleneck today. It's really like good old engineering and product work. I assume you're hiring. What's the call for hiring at the end? - Any other comments on the model side? When did you start feeling like the model was not a bottleneck anymore? Was it 4.0? Was it 3.5?
- 3.5 Sonnet, definitely. I think 4.0 is overhyped, frankly. We don't use 4.0. I don't think it's good for agentic behavior. - Yeah, 3.5 Sonnet is when I started feeling that. And then with prompt caching with 3.5 Sonnet, like that fills the cost, cut the cost again. - Just cut it by half.
- Yeah. - Your prompts are, some of the problems with agentic uses is that your prompts are kind of dynamic, right? Like from caching to work, you need the front prefix portion to be stable. - Yes, but we have this append-only ledger paradigm. So every node keeps appending to that ledger and every filled node inherits all the context built up by all the previous nodes.
And so we can just decide like, hey, every X thousand nodes, we trigger prompt caching again. - Oh, so you do it like programmatically, not all the time. - No, sorry. Anthropic manages that for us. But basically it's like, because we keep appending to the prompt, we just, like the prompt caching works pretty well.
- We have this like small podcaster tool that I built for the podcast and I rewrote all of our prompts because I noticed, you know, I was inputting stuff early on. I wonder how much more money OpenAN and Anthropic are making just because people don't rewrite their prompts to be like static at the top and like dynamic at the bottom, but.
- I think that's the remarkable thing about what we're having right now is it's insane that these companies are routinely cutting their costs by two, four, five. Like they basically just apply constraints. They want people to take advantage of these innovations. - Very good. Do you have any other competitive commentary?
Dust, WordWare, Gumloop, Zapier? If not, we can move on. - No comment. I think the market is, look, I mean, AGI is coming. - All right, that's what I'm talking about. - I think you're helping. Like you're paving the road to AGI. - I'm playing my small role. I'm adding my small brick to this giant, giant, giant castle.
Yeah, look, when it's here, we are gonna, this entire category of software is going to create, it's going to sound like an exaggeration, but it is a fact that it's going to create trillions of dollars of value in a few years, right? It's going to, for the first time, we're actually having software directly replace human labor.
I see it every day in sales calls. It's like Lindy is today replacing, like we talk to even small teams. It's like, oh, like, stop, this is a 12 people team here. I guess we'll set up this Lindy for one or two days, and then we'll have to decide what we do with this 12 people team.
And so, yeah, to me, there's this immense uncapped market opportunity. It's just such a huge ocean. And there's like three sharks in the ocean. I'm focused on the ocean more than on the sharks. - Cool. So we're moving on to hot topics, like kind of broadening out from Lindy, but obviously informed by Lindy.
What are the high order bits of good agent design? - The model, the model, the model, the model. I think people fail to truly, and me included, they fail to truly internalize the bitter lesson. So for the listeners out there who don't know about it, it's basically like, you just scale the model, like GPUs go brrr, it's all that matters.
I think it also holds for the cognitive architecture. I used to be very cognitive architecture-filled, and I was like, ah, and I was like a critic, and I was like a generator, and all this, and then it's just like GPUs go brrr, like just like let the model do its job.
I think we're seeing it a little bit right now with O1. I'm seeing some tweets that say that the new 3.5 SONET is as good as O1, but with none of all the crazy- - It beats O1 on some measures. - On some reasoning tasks. - On AIME, it's still a lot lower.
Like it's like 14 on AIME versus O1, it's like 83. - Got it. - So. (laughs) - Right. - But even O1 is still the model. - Yeah. - Like there's no cognitive architecture on top of it. You can just like wait for O1 to get better. - And so as a founder, how do you think about that?
Right, because now knowing this, wouldn't you just wait to start Lindy? You know, you started Lindy, it's like 4K context, the models are not that good. It's like, but you're still kind of like going along and building and just like waiting for the models to get better. How do you today decide, again, what to build next?
- Yeah. - Knowing that, hey, the models are gonna get better, so maybe we just shouldn't focus on improving our prompt design and all that stuff and just build the connectors instead or whatever. - Yeah, I mean, that's exactly what we do. Like all day, we always ask ourselves, oh, when we have a feature idea or a feature request, we ask ourselves like, is this the kind of thing that just gets better while we sleep because models get better?
I'm reminded again, when we started this in 2022, we spent a lot of time because we had to around the context pruning, 'cause 4,000 tokens is really nothing. You really can't do anything with 4,000 tokens. All that work was throwaway work. Like now it's like it was for nothing, right?
Now we just assume that infinite context windows are gonna be here in a year or something, a year and a half, and infinitely cheap as well. And dynamic compute is gonna be here. Like we just assume all of these things are gonna happen. And so we really focus, our job to be done in the industry is to provide the input and output to the model.
I really compare it all the time to the PC and the CPU, right? Apple is busy all day. They're not like a CPU wrapper. They have a lot to build, but they don't. Well, now actually they do build the CPU as well, but leaving that aside, they're busy building a laptop.
It's just a lot of work to build these things. - It's interesting 'cause like, for example, another person that we're close to, Mihaly from Repl.it, he often says that the biggest jump for him was having a multi-agent approach, like the critique thing that you just said that he doesn't need.
And I wonder when, in what situations you do need that and what situations you don't. Obviously the simple answer is for coding, it helps. And you're not coding except for... Are you still generating code? - In Lindy? - Yeah. - No, we do. - Not really, right? - Oh, right.
No, no, no, the cognitive architecture changed. We don't, yeah. - Yeah, yeah, okay. For you, you one-shot and you chain tools together and that's it. - And if the user really wants to have this kind of critique thing, you can also edit the prompt. You're welcome to. I have some of my Lindys, I've told them like, "Hey, be careful, think step-by-step "about what you're about to do." But that gives you a little bump also on use cases, but yeah.
- What about unexpected model releases? So Anthropic released computer use today. I don't know if many people were expecting computer use to come out today. Do these things make you rethink how to design like your roadmap and things like that? Or are you just like, "Hey, look, whatever." That's just like a small thing in their like AGI pursuit that like maybe they're not even gonna support.
And like, it's still better for us to build their own integrations into systems and things like that. Because maybe people would say, "Hey, look, why am I building all these API integrations "when I can just do computer use "and never go to the product?" - Yeah. No, I mean, we did take into account computer use.
We were talking about this a year ago or something. Like we've been talking about it as part of our roadmap. It's been clear to us that it was coming. Like we've read reports of OpenAI working on something like that for a very long time. My philosophy about it is anything that can be done with an API must be done by an API or should be done by an API for a very long time.
I think it is dangerous to be overly cavalier about improvements of model capabilities. I'm reminded of iOS versus Android. Android was built on the JVM. There was a garbage collector. And I can only assume that the conversation that went down in the engineering meeting room was, "Oh, who cares about the garbage collector?
"Anyway, Moore's law is here. "And so that's all going to go to zero eventually." Sure, but in the meantime, you are operating on a 400 megahertz CPU, was like the first CPU on the iPhone 1. And it's really slow. And the garbage collector is introducing a tremendous overhead on top of that, especially like a memory overhead.
And so for the longest time, and it's really only been recently that Android caught up to iOS in terms of how smooth the interactions were. But for the longest time, Android phones were significantly slower and laggier and just not feeling as good as iOS devices. And so, look, when you're talking about all those magnitude of differences in terms of performance and reliability, which is what we are talking about when we're talking about API use versus computer use, then you can't ignore that, right?
And so I think we're going to be in an API use world for a while. - O1 doesn't have API use today. It will have it at some point. It's on the roadmap. There is a future in which OpenAI goes much harder after your business, your market, than it is today.
Like ChatGPT, it's its own business. It's making like $2 billion a year or something. All they need to do is add tools to the ChatGPT, and now they're suddenly competing with you. And by the way, they have a GPT store where a bunch of people have already configured their tools to fit with them.
Is that a concern? - I think even the GPT store in a way, like the way they architect it, for example, the plug-in systems are actually grateful because it's like we can also use the plug-ins. It's very open. No, again, I think it's going to be such a huge market.
I think there's going to be a lot of different jobs to be done. Today, at least, ChatGPT, I know they have like a huge enterprise offering and stuff, but today, ChatGPT is a consumer app, right? And so the sort of flow detail I showed you, this sort of workflow, this sort of use cases that we're going after, which is like, we're doing a lot of like lead generation and lead outreach and all of that stuff.
That's not something like meeting recording, like Lindy Today right now joins your Zoom meetings and takes notes, all of that stuff. I don't see that so far on the OpenAI roadmap. - Yeah, but they do have an enterprise team that we talked to for a Decibel Summit. Cool, I have some other questions on company building stuff.
You're hiring GMs? - We did. - It's a fascinating way to build a business, right? Like what should you, as CEO, be in charge of? And what should you basically hire a mini CEO to do? - Yeah, that's a good question. I think that's all something we're figuring out.
The GM thing was inspired from my days at Uber, where we hired one GM per city or per major geo area. We had like all GMs, regional GMs and so forth. And yeah, Lindy is so horizontal that we thought it made sense to hire GMs to own each vertical and to go to market of the vertical and the customization of the Lindy templates for these verticals and so forth.
What should I own as a CEO? I mean, the canonical reply here is always going to be, you own the fundraising, you own the culture, you own the, what's the rest of the canonical reply? The culture, the fundraising. - I don't know, products. - Even that eventually you do have to hand out.
Yes, the vision, the culture and the fundraising. And it's like, if you just do these things and you've done it well, you've done your job as a CEO. In practice, obviously, yeah. I mean, all day, I do a lot of product work still and I want to keep doing product work for as long as possible.
Obviously, like you're recording and managing the team, yeah. - That one feels like the most automatable part of the job, the recruiting stuff. - Well, yeah. You saw my design your recruiter here, yeah. - Relationship between Factorio and building Lindy. - We actually very often talk about how the business of the future is like a game of Factorio.
It's like, you just wake up in the morning and you've got your Lindy instance, it's like Slack, and you've got like 5,000 Lindys in the sidebar and your job is to somehow manage your 5,000 Lindys. And it's going to be very similar to company building because you're going to look for the highest leverage way to understand what's going on in your AI company and understand what levels do you have to make impact in that company.
So I think it's going to be very similar to like a human company, except it's going to go infinitely faster. Today in a human company, you could have a meeting with your team and you're like, "Oh, I guess we need one more designer. Okay, I guess I'll kick off a search." And two months later, you have a new designer.
Now it's like, "Okay, boom, I'm going to spin up 50 designers." - That is going to go away. - Yeah. - Like actually, it's more important that you can clone an existing designer that you know works. Because the hiring process, you cannot clone someone. - Yeah. - Because every new person you bring in is going to have their own tweaks and you don't want that.
- Yeah, yeah, that's true. - You want an army of mindless drones that all work the same way. The reason I bring Factorio up as well is one, Factorio Space just came out. Apparently a whole bunch of people stopped working. I tried out Factorio. I never really got that much into it.
But the other thing was you had a tweet recently about how the sort of intentional top-down design was not as effective as just build. - Yeah. - Just ship. - I think people read it a little bit too much into that tweet. It went weirdly viral. I did not intend it as a giant statement on life.
- I mean, you notice you have a pattern of this, right? Like you've done this for eight years now. You should know. (both laughing) - I legit was just hearing an interesting story about the Factorio game I had. And everybody was like, "Oh my God, so deep." I guess this explains everything about life and companies.
And there is something to be said certainly about focusing on the constraint. And I think it is Patrick Collison who said, "People underestimate the extent to which moonshots are just one pragmatic step taken after the other." And I think as long as you have some inductive bias about like some loose idea about where you want to go, I think it makes sense to follow a sort of greedy search along that path.
I think planning and organizing is important and having order is important. - I'm wrestling with that. There's two ways I encountered it recently. One with Lindy. When I tried out one of your automation templates and one of them was quite big and I just didn't understand it, right? So like it was not as useful to me as a small one that I can just plug in and see all of.
And then the other one was me using Cursor. I was very excited about O1 and I just upfront stuffed everything I wanted to do into my prompt and expected O1 to do everything. And it got itself into a huge jumbled mess and it was stuck. It was really... There was no amount, I wasted like two hours on just like trying to get out of that hole.
So I threw away the code base, started small, switched to a class on it and build off something working and just added over time and it just worked. And to me, that was the factorial sentiment, right? Maybe I'm one of those fanboys that's just like obsessing over the death of something that you just randomly tweeted out.
But I think it's true for company building, for Lindy building, for coding, I don't know. - I think it's fair. And I think like you and I talked about there's the Tuft and Metal principle and there's this other. - Yes, I love that. - There's the, I forgot the name of this other blog post but it's basically about this book, "Seeing Like a State" that talks about the need for legibility and people who optimize the system for its legibility.
And anytime you make a system, so legible is basically more understandable. Anytime you make a system more understandable from the top down, it performs less well from the bottom up. And it's fine if that's what you want, but you should at least make this trade off with your eyes wide open.
You should know I am sacrificing performance for understandability, for legibility. And in this case for you, it makes sense. It's like you are actually optimizing for legibility. You do want to understand your code base, but in some other cases, it may not make sense. Sometimes it's better to leave the system alone and let it be its glorious, chaotic, organic self and just trust that it is going to perform well, even though you don't understand it completely.
- It does remind me of a common managerial issue or dilemma, which you experienced in a small scale of Lindy where, you know, do you want to organize your company by functional sections or by products or, you know, whatever the opposite of functional is. And you tried it one way and it was more legible to you as CEO, but actually it stopped working at the small level.
- Yeah, I mean, one very small example, again, at a small scale is we used to have everything on Notion. And for me as founder, it was awesome because everything was there. The roadmap was there, the tasks were there, the postmortems were there. And so the postmortem was linked to a task.
- Yeah, it's optimized for you. - It was exactly. And so I had this like one pane of glass and everything was on Notion. And then the team one day came to me with pitchforks and they really wanted to implement Linear. And I had to bite my fist so hard.
I was like, fine, do it, implement Linear. 'Cause I was like, at the end of the day, the team needs to be able to self-organize and pick their own tools. Yeah, but it did make the company slightly less legible for me. - Another big change you had was going away from remote work, bringing people back in person.
I think there's obviously every other month the discussion comes up again. What was that discussion like? How did your feelings change? Was there kind of like a threshold of employees and team size where you felt like, okay, maybe that worked. Now it doesn't work anymore. And how are you thinking about the future as you scale the team?
- Yeah, so for context, I used to have a business called TeamFlow. The business was about building a virtual office for remote teams. And so being remote was not merely something we did. I was banging the remote drum super hard because we were helping companies to go remote, right?
And so, frankly, in a way it's a bit embarrassing for me to do like a 180 like that, but I guess when the facts changed, I changed my mind. What happened? Well, I think at first, like everyone else, we went remote by necessity. It was like COVID and you got to go remote.
And on paper, the gains of remote are enormous. In particular, from a founder standpoint, being able to hire from anywhere is huge. Saving on rent is huge. Saving on commute is huge for everyone and so forth. But then, look, I'm not going to say anything original here. It's like it is really making it much harder to work together.
And I spent three years of my youth trying to build a solution for this. And my conclusion is at least we couldn't figure it out and no one else could. Zoom didn't figure it out. We had like a bunch of competitors, like Gathertown was one of the bigger ones.
We had dozens and dozens of competitors. No one figured it out. I don't know that software can actually solve this problem. Reality of it is everyone just wants to get off the darn Zoom call. And it's not a good feeling to be in your home office if you even are lucky enough to have a home office all day.
It's harder to build culture. It's harder to get in sync. I think software is peculiar because it's like an iceberg. It's like the vast majority of it is submerged under water. And so the quality of the software that you ship is a function of the alignment of your mental models about what is below that waterline.
Can you actually get in sync about what it is exactly fundamentally that we're building? What is the soul of a product? And it is so much harder to get in sync about that when you're remote. And then you waste time in a thousand ways because people are offline and you can't get ahold of them or like you can't share your screen.
It's just, it's like you feel like you're walking in more or less is all day. And eventually I just, I was like, okay, this is it. Like, we're not gonna do this anymore. - Yeah. I think that is the current builder San Francisco consensus here. But I still have a big, like one of my big heroes as a CEO is Sid Subbanj from GitLab.
Matt Molleweg used to be a hero, but like these people run thousand person remote businesses. The main idea is that at some company size, your company is remote anyway. Because if you go from one building to two buildings, you're congrats, you're now remote from the other building. Like if you won't go from one city office to like two city offices, they're remote from each other.
- But the teams are co-located. Every time anyone talks about remote success stories, they always talk about this real force. I mean, it's always GitLab and WordPress and Zapier. - Zapier. - It used to be InVision. (laughs) And I will point out that in every one of these examples, you have a co-located counterfactual that is sometimes orders of magnitude bigger.
Look, I like Matt Molleweg a lot, but WordPress is a commercial failure. They run 60% of the internet and they're like a fraction of the size of even Substack. Right? Or-- - They're trying to get more money. (laughs) - Yeah, that's my point, right? Like look, GitLab is much smaller than GitHub.
InVision, you know, is no more. And Figma like completely took off. And Figma was like very in person. Figma let go of people because they wanted to move from San Francisco to LA. So I think if you're optimizing for productivity, if you really know, hey, this is a support ticket, right?
And I want to have my support tickets for a buck 50 per support ticket. And next year, I want it for like a buck 20. Then sure, send your support ticket team to offshore like the Philippines or whatever, and just optimize for cost. If you're optimizing for cost, absolutely be remote.
If you're optimizing for creativity, which I think that's software and product building is a creative endeavor. If you're optimizing for creativity, it's kind of like composing an album. You can't do it on the cheap. You want the very best album that you can make. And you have to be in person and hear the music to do that.
- Yeah. So the line is that all jobs that can be remote should be AI or Lindy's, and all jobs that are not remote are in person. Like there's a very, very clear separation of jobs. - Sure. Well, I think over the long term, every job is going to be AI anyway.
(all laughing) - It would be curious to break down what you think is creativity in coding and in product defining and how to express that with LLMs. I think that is underexplored for sure. You're definitely a, what I call a temperature zero use case of LLMs. You want it to be reliable, predictable, small.
And then there's other use cases of LLMs that are more for like creativity and engines, right? I haven't checked, but I'm pretty sure no one uses Lindy for brainstorming. Actually, probably they do. - I use Lindy for brainstorming a lot, actually. - Yeah, yeah, yeah. But like, you know, you want to have like something that's anti-fragile to hallucination.
Like hallucinations are good. - By creativity, I mean, is it about direction or magnitude? If it is about direction, like decide what to do, then it's a creative endeavor. If it is about magnitude and just do it as fast as possible, as cheap as possible, then it's magnitude. And so sometimes, you know, software companies are not necessarily creative.
Sometimes you know what you're doing. And I'll say that is going to come across the wrong way, but Linear, I look up to a huge amount, like such amazing product builders, but they know what they're building. They're building a task tracker. And so Linear is remote, right? Linear is building a task tracker, right?
I don't mean to throw shade at them, like good for them. I think they're aware that they're not like- - They recently got shit for saying that they have work-life balance on their job description. They were like, "What do you mean by this?" (laughing) - We're building a new kind of product that no one's ever built before.
And so we're just scratching our heads all day, trying to get in sync about like, what exactly is it that we're building? What does it consist of? - Inherently creative struggle. - Yeah. - Dare we ask about San Francisco? And there's a whole bunch of tough stuff in here.
I don't know if you have any particular leanings. Probably the biggest one I'll just congratulate you on is becoming American, right? Like you, very French, but your heart was sort of in the US, you eventually found your way here. What are your takes for like founders, right? Like a few years ago, you wrote this post on like, "Go West, young man." And now you've basically completed that journey, right?
Like you're now here and up to the point where you're kind of mystified by how Europe has been so decel. - In a way though, I feel vindicated 'cause I was making the prediction that Europe was over 14 years ago or something like that. I think it's been a walking corpse for a long time.
I think it is only now becoming obvious that it is paying the consequences of its policies from 10, 20, 30 years ago. I think at this point, I wish I could rewrite the "Go West, young man" article, but really even more extreme. I think at this point, if you are in tech, especially in AI, but if you're in tech and you're not in San Francisco, you either lack judgment or you lack ambition.
It's one of the two. It's funny, I recently told that to someone and they were like, "Oh, like not everyone wants to be like a unicorn founder." And I was like, "Like I said, judgment or ambition." It's fine to not have ambition. It's fine to want to prioritize other things than your company in life or your career in life.
That's perfectly okay. But know that that's the trade-off you're making. If you prioritize your career, you've got to be here. - As a fellow European escapist, I grew up in Rome. - Yeah, how do you feel? We never talked about your feelings. - Yeah, I've been in the US now six years.
Well, I started my first company in Europe 10 years ago, something like that. And yeah, you can tell nobody really wants to do much. And then you're like, "Okay." It's funny, I was looking back through some old tweets and I was sending all these tweets to Mark Andreessen like 15 years ago, trying to learn more about why are you guys putting money in these things that most people here would say you're crazy to even back.
And eventually I started doing venture, yeah, six, five years ago. And I think just like so many people in Europe reach out and ask, "Hey, can you talk to our team?" And like, blah, blah, blah. And they just cannot comprehend the risk appetite that people have here. It just like so foreign to people, at least in Italy and like in some parts of Europe.
I'm sure there's some great founders in Europe, but like the average European founders, like why would I leave my job at the post office to go work on the startup that could change everything and become very successful, but might go out of business. Instead in the US, you have like, you know, we host a hackathon and it's like 400 people show up and it's like, where can I go work that it's like no job security?
You know? - Yeah. - It's just like completely different and there's no incentives from the government to change that. There's no way you can like change such a deep-rooted culture of like, you know, going in wine and April spritz and all of that early in the afternoon. So I don't really know how it's going to change.
- It's quality of life. - Yeah, totally. That's why I left. (all laughing) The quality's so high that I left. But again, I agree with you. It's just like, hey, like there's no rational explanation as to why it's better to move here. It just, if you want to do this job and do this, you should be here.
If you don't want to, that's fine. But like, don't cope him. - Right. - Don't be like, oh no, you can also be successful doing this and knees or like whatever. No, probably not, you know? So yeah, I've already done my N400. So I should get my U.S. citizenship interview.
- Hell yeah. - Damn. - Soon. - Yeah. And I think to be fair, I think what's happening right now to Europe is largely self-inflicted. I think that it's just completely, again, they've said no to capitalism. They've decided to say no to capitalism a long time ago. They've completely over-regulated.
Taxation is much too high and so forth. But I also think some of this is a little bit of a self-fulfilling prophecy or it's a self-perpetuating phenomenon. Because, look, to your point, once there is a network effect that's just so incredibly powerful, they can't be broken, really. And we tried with San Francisco.
I tried with San Francisco. Like during COVID, there was a movement of people moving to Miami. - You and I both moved there. (laughing) - How did that pan out? You can't break the network effect, you know? - It's so annoying because first principles wise, tech should not be here.
Like tech should be in Miami 'cause it's just better city. (laughing) - San Francisco does not want tech to be here. - San Francisco hates tech. - 100%. - This is the thing I actually wrote down. Like San Francisco hates tech. - It is true. - I think the people that are in San Francisco that were here before, tech hated and then there's kind of like this passed down thing.
But I would say people in Miami would hate it too if there were too much of it, you know? Like the Niki Beach crowd would also not chill. - They're just rich enough and chill enough to not care. - Yeah, I think so too. Like, oh, crypto kids. Okay, cool.
- Yeah. (laughing) - Yeah, Miami celebrates success, which is one thing I loved about it. - A little bit too much. (laughing) Maybe the last thing I'll mention, I just wanted a little bit of EUAC talk. I think that's good. I'll maybe carve out that I think the UK has done really well.
That's an argument for the UK not being part of Europe is that, you know, the AI institutions there at least have done very well, right? - Sure. - The economy of Britain is in the gutter. - Yeah, exactly. - They've been stagnating at best. And then France has a few wins.
- Who? - Mistral. - Who uses Mistral? - Hugging face. A few wins. (laughing) I'm just saying. They disappointed their first AI minister. - You know the meme with the guy who's celebrating with his trophy and then he's like, "No, it's France." Right? To me, that's France. It's like, "Aha, look, we've got Mistral!" And it's like, "Champagne!" And it's like maybe 1% of market share.
And by the way, I said that I love Mistral. I love the guys over there. And it's not a critic of them. It's a critic of France and of Europe. And by the way, I think I've heard that the Mistral guys were moving to the U.S. - Yeah, they're opening an office here.
- They're opening an office here. - But I mean, they're very French, right? (laughing) You can't really avoid it. There's one interesting counter move which is Jason Warner and ISOCAT moving to Paris for poolside. - I don't know. - It remains to be seen how that move is going.
Maybe the last thing I'll say, you know, that's the Europe talk. We try not to do politics so much, but you're here. One thing that you do a lot is you test your Overton windows, right? Like far more than any founder I know. You know it's not your job.
Someone, for sure, you're just indulging, but also I think you consciously test. And I just want to see what drives you there and why do you keep doing it? (laughing) 'Cause some of you tweet very spicy stuff, especially for like the San Francisco sort of liberal dynasty. - I don't know because, so I assume you're referring to recently, I posted something about pronouns and how non-sense.
- Just in general. - I don't want you to focus on any particular thing unless you want to. - You know, well, is that tweet in particular, when I was tweeting it, I was like, "Oh, this is kind of spicy. "Should I do this?" And then I just did it.
And I, you know, I received zero pushback and the tweet was actually pretty successful and I received a lot of people reaching out like, "Oh my God, so true." I think it's coming from a few different places. One, life is more fun this way. Like I don't feel like self-censoring all the time.
You know, it's just, it's like, you know, that's number one. Number two, if everyone always self-censors, you never know what everyone, what anyone thinks. And so it's becoming like a self-perpetuating thing. It's like a public lies, private truth sort of phenomenon. Or like, you know, it's like, there's this phenomenon called a preference cascade.
It's like, there's this joke. It's like, oh, there's only one communist left in USSR. The problem is no one knows which one it is. So everyone pretends to be communist because everyone else pretends to be a communist. And so I think there's a role to be played for someone to have backbone and just be like, "Hey, I'm thinking this." And actually everyone thinks the same, especially when you are like me in a position where it's like, I don't have a boss who's going to fire me.
It's like, look, if I don't speak up and if founders don't speak up, I'm like, why, what are you afraid of? Right, like there's really not that much downside. And I think there's something to be said about standing up for what you think is right and being real and owning your opinions.
- I think there's a correlation there between having that level of independence for your political beliefs and free speech or whatever, and the way that you think about business too. Like I see that it helps, I think. - I think the world contrarian has become abused, but I think there's such a powerful insight that it's cool, which is groupthink is real and pervasive and really problematic.
Like your brain constantly shuts down because you're not even thinking in your other way or you're not thinking. You just look around you and you decide to adopt the same beliefs as people around you. And everyone thinks they're immune and everyone else is doing it except themselves. - I'm a special snowflake, I have free will.
- That's right, and so I actually make it a point to like look for, hey, what would be a thing right now that I can't really say? And then I think about it and I'm like, "Do I believe this thing?" And very often the answer is yes. And then I just say it.
And so I think the AI safety is an example of that. Like at some point, Mark Andresen blocked me on Twitter and it hurt, frankly. I really look up to Mark Andresen and I knew he would block me. - It means you're successful on Twitter. That's just a right of passage.
- Mark Andresen was really my booster initially on Twitter. He really made my account. And I was like, "Look, I'm really concerned "about AI safety. "It is an unpopular view amongst my peers." - I remember you were one of the few that actually came out in support of the bill or something.
- I came out in support of SB1047. A year and a half ago, I put some tweet storms about how I was really concerned. And yeah, I was blocked by a bunch of A6NZ people and I don't like it, but it's funny. Maybe it's my French education. But look, in France, World War II is very present in people's minds.
And the phenomenon of people collaborating with the Nazis during World War II is really present in people's minds. And there is always this sort of debate that people have at dinner and say, "Ah, would you really have resisted "during World War II?" And everybody is always saying, "Oh yeah, I would totally have resisted." It's like, "Yeah, but no." Like, look, the reality of it is 95% of the country did not resist and most of it actually collaborated actively with the Nazis.
And so 95% of y'all are wrong. You would actually have collaborated, right? I've always told myself I will stand for what I think is right, even if I've gotten into physical fights in my life, like in SF, because some people got attacked. And the way I was brought up is like if someone gets attacked before you, you get involved.
Like, it doesn't matter. You get involved and you help the person, right? And so, look, I'm not pretending we're like nowhere near like a World War II phenomenon, but I'm like exactly because we are nowhere near this kind of phenomenon. Like the stakes are so low. And if you're not going to stand up for what you think is right when the stakes are so low, are you going to stand up when it matters?
- Italian education is that in Italy, people don't have guns when you fight them, so you can always get in a fight. But here in the US, I'm always like, "Oh man." - I feel, I detect some inconsistency in your statements because you simultaneously believe that AGI is very soon.
And you also say stakes are low. You can't believe both are real. - Well, the stakes, so why does AGI make the stakes of speaking up higher? - Sorry, the stakes of like safety. - Oh yeah, no, the stakes of AIs, or like physical safety? - No, AI safety.
- Oh no, the stakes of AI safety couldn't be higher. I meant the stakes of like speaking up about- - Pronouns or whatever. - Oh, okay, okay. - Yeah, yeah, yeah. - How do you figure out who's real and who isn't? Because there was the whole like manifesto for responsible AI that like hundreds of like VCs and people signed, and I don't think anybody actually, any of them thinks about it anymore.
- Was that the pause letter, like six month pause, or? - Some, no, there was like something else too that I think general catalyst and like some fun sign, but, and then there's maybe the anthropic case, which is like, "Hey, we're leaving open AI because you guys don't take security seriously." And then it's like, "Hey, what if we gave AI access to a whole computer to just like go do things?" Like, how do you reconcile like, okay, I mean, you could say the same thing about Lindy.
It's like, if you're worried about AI safety, why are you building AI, right? That's kind of like the extreme thinking. How do you internally decide between participation and talking about it and saying, "Hey, I think this is important, but like I'm still gonna build towards that and building actually makes it safer because I'm involved," versus just being like anti, "I think this is unsafe," but then not do anything about it and just kind of remove yourself from the whole thing, if that makes sense.
- Yeah, the way I think about our own involvement here is I'm acutely concerned about the risks at the model layer. And I'm simultaneously very excited about the upside. Like for the record, my PDoOM, insofar as I can quantify it, which I cannot, but if I had to, like my vibe is like 10% or something like that.
And so there's like a 90% chance that we live in like a pure utopia, right? And that's awesome, right? So like, let's go after the utopia, right? And let's talk about the 10% chance that things go terribly wrong, but I do believe there's a 90% chance that we live in a utopia where there's no disease and it's like a post-scarcity world.
I think that utopia is going to happen through, like, again, I'm bringing my little contribution to the movement. I think it would be silly to say no to the upside because you're concerned about the downside. At the same time, we want to be concerned about the downside. I know that it's very self-serving to say, "Oh, you know, like the downside doesn't exist at my layer, it exists at like the model layer." But truly, look at Lindy, look at the Apple building.
I struggle to see exactly how it would like get up and start doing crazy stuff. I'm concerned about the model layer. - Okay, well, this kind of discussion can go on for hours. It is still daylight, so not the best time for it, but I really appreciate you spending the time.
Any other last calls to actions or thoughts that you feel like you want to get off your chest? - AGI is coming. (all laugh) - Are you hiring for any roles? - Oh yeah, I guess that should be the... (all laugh) - Don't bother. - No, can you stop saying AGI is coming and just talk about it?
- We are also hiring. Yeah, we are hiring designers and engineers right now. - Yeah. - So hit me up at a flow@lindy.ai. - And then go talk to my Lindy. - That's right. - You're not actually going to read it. - Actually, I have wondered how many times when I talk to you, I'm talking to a bot.
- I want that as a discussion. - Part of that is I don't have to know, right? - That's right. Well, it's actually doubly confusing because we also have a teammate whose name is Lindy. - Yes, I was wondering, I met her. I was like, "Wait, did you hire her first?" - Marketing is fun.
- No, she was an inspiration. We named the company both after her. - Okay, interesting, interesting. Yeah, wonderful. I'll comment on the design piece just because I think that there are a lot of AI companies that very much focus on the functionality and the models and the capabilities and the benchmark.
But I think that increasingly I'm seeing people differentiate with design and people want to use beautiful products and people who can figure that out and integrate the AI into their human lives. Design at the limit. One, at the lowest level, it's make this look pretty, make this look like Stripe or Linear's homepage.
That's design. But at the highest level of design, it is make this integrate seamlessly into my life. Intuitive, beautiful, inspirational, maybe even. And I think that companies that, this is kind of like a blog post I've been thinking about, companies that emphasize design actually are going to win more than companies that don't.
- Yeah, I love this Steve Jobs quote and I'm going to butcher it. It's something like, "Design is the expression of the soul of a man-made product "through successive layers of design." - Jesus. - He was good. - He was cooking, he was cooking on that one. - It starts with the soul of the product, which is why I was saying it is so important to reach alignment about that soul of the product, right?
It's like an onion, like you peel the onion in those layers, right? And you design an entire journey, just like the user experiencing your product chronologically all the way from the beginning of like the awareness stage, I think it is also the job of the designer to design that part of the experience.
It's like, okay, what, you know? And that's brand basically. So yeah, I agree with you. I think design is immensely important. - Okay, lovely. - Yeah, thanks for coming on, Flo. - Yeah, absolutely. Thanks for having me. (upbeat music) (upbeat music) (upbeat music)