back to indexPowering your Copilot for Data - with Artem Keydunov from Cube.dev
Chapters
0:0 Introductions
1:35 History of Statsbot - Slack bot for querying data in Slack
4:45 Building Cube to power Statsbot due to limitations in natural language processing at the time
6:50 Open sourcing Cube as a standalone product
8:34 Explaining the concept of a semantic layer and OLAP cubes
10:27 Using semantic layers to provide context to AI models
11:54 Challenges of using tabular vs. language data with AI models
13:11 Workflow of natural language to SQL query using semantic layer
16:1 Ensuring AI agents have proper data context and make correct queries
18:20 Treating metrics definitions in the semantic layer as a codebase with collaboration
22:55 Natural language capabilities becoming a commodity baseline for BI tools
24:37 Recommendations for building data-driven AI applications
28:26 Predictions on the consolidation of modern data stack tools/companies
30:14 AI assistance augmenting but not fully automating data workflows
34:20 Using external Python scripts to handle limitations of models with math
36:15 Embedded analytics challenges and natural language commoditization
39:4 Lightning round
00:00:09.640 |
This is Swigx, writer, editor of Latent Space 00:00:17.540 |
And today we have Artem Ketunov on the podcast, 00:00:32.640 |
And Cube is actually a spin out of his previous company, 00:00:37.040 |
And this kind of feels like going both backward 00:00:42.020 |
So the premise of Statsbot was having a Slack bot 00:00:54.480 |
and you see startups trying to do that today. 00:01:02.140 |
And Cube then evolved from an embedded analytics product 00:01:09.580 |
I think you have over 16,000 stars on GitHub today. 00:01:13.480 |
You have a very active open source community. 00:01:24.080 |
You know, what got you interested in like text to SQL 00:01:27.200 |
and what were some of the limitations that you saw then, 00:01:30.960 |
the limitations that you're also seeing today 00:01:45.320 |
based off my initial project that I did at a company 00:01:59.880 |
you know, like Slack apps, Chatsbots in general. 00:02:04.420 |
like another wave of, you know, bots and all of that. 00:02:10.760 |
So we were like living through one of those waves. 00:02:18.440 |
from the different places where like a data lives to Slack. 00:02:26.400 |
you know, maybe some marketing data, Google Analytics, 00:02:31.400 |
like a production databases or even Salesforce sometimes. 00:02:52.680 |
and people started to use it even before Slack came up 00:03:15.040 |
So they featured me on this application directory front page, 00:03:21.640 |
It was a lot of fun, I think, you know, like, 00:03:26.600 |
in terms of how you can process natural language, 00:03:29.260 |
because the original idea was to let people ask questions 00:03:37.340 |
like opportunities closed last week or something like that. 00:03:49.360 |
but it was, you know, we didn't have LLMs, right, 00:03:55.880 |
especially the systems that can kind of, you know, 00:04:03.520 |
and like it was a lot of hit and miss, right? 00:04:05.320 |
If you know how to construct a query in natural language, 00:04:28.680 |
So right now I'm kind of bullish that, you know, 00:04:32.860 |
probably would have a much better shot at it, 00:04:41.380 |
as we were working on Statsbot because we needed it. 00:04:50.560 |
like a natural language understanding models? 00:04:53.160 |
Like were there open source models that were good 00:04:57.080 |
- I think it was mostly combination of a bunch of things, 00:05:02.140 |
The first version which I built, like was RegApps. 00:05:13.520 |
and I had a natural language pricing tool thing, 00:05:27.000 |
And he started to like do some stuff that was like, 00:05:33.800 |
And, you know, like he started to do like some models 00:05:38.280 |
like look at what we had on the market back then, 00:05:41.240 |
or, you know, like try to build a different sort of, 00:05:45.920 |
Again, we didn't have any foundation back in place, right? 00:05:48.680 |
We wanted to build something that, you know, like we, 00:05:51.260 |
okay, we wanted to try to use existing math, obviously, 00:05:54.200 |
right, but it was not something that we can take the model 00:05:58.940 |
I think in 2019, we started to see more like of stuff, 00:06:07.160 |
like resulted in all this LLM, like what we have right now. 00:06:10.120 |
But back then in 2016, it was not much, you know, 00:06:13.920 |
like available for just the people to build on top. 00:06:18.280 |
kind of been happening, but it was like very, very early, 00:06:21.680 |
you know, like for something to actually be able to use. 00:06:28.160 |
which was started just as an open source project. 00:06:30.480 |
And I think I remember going on a walk with you 00:06:33.120 |
in San Mateo in like 2020, something like that. 00:06:36.440 |
And you were like, you have people reaching out to you 00:06:38.840 |
who are like, "Hey, we use Kube in production." 00:06:49.760 |
- We built a Kube at Statsbot because we needed it. 00:07:07.680 |
okay, people wanted to get active opportunities, right? 00:07:16.320 |
you always, you know, like try to reduce everything down 00:07:19.160 |
to the sort of, you know, like a multidimensional framework. 00:07:24.080 |
And that's where, you know, like it didn't really work well 00:07:39.680 |
So we built a framework where you would be able 00:07:42.760 |
to map your data into this concept, into this metrics. 00:07:49.440 |
they were bringing their own datasets, right? 00:07:52.040 |
And the big question was, how do we tell the system 00:07:55.520 |
what is active opportunities for that specific users? 00:07:58.080 |
How we kind of, you know, like provide that context, 00:08:10.860 |
But at some point we saw people started to see more value 00:08:24.540 |
it feels like it might be a standalone product 00:08:31.440 |
So we took it out of Statsbot and open sourced. 00:08:39.240 |
The concept of a Kube is not something that you invented. 00:08:42.280 |
I think, you know, not everyone has the same background 00:08:44.440 |
in analytics and data that all three of us do. 00:08:47.520 |
Maybe you want to explain like OLAP Kube, Hyper Kube, 00:08:50.320 |
you know, anything, whatever the brief history of Kubes. 00:08:56.560 |
I'll try, you know, like there's a lot of like 00:08:58.980 |
Wikipedia pages and like a lot of like a blog post 00:09:08.180 |
so when we think about just a table in a database, 00:09:10.860 |
the problem with the table, it's not a multidimensional, 00:09:13.740 |
meaning that in many cases, if we want to slice the data, 00:09:17.420 |
we kind of need to result with a different table, right? 00:09:20.900 |
Like think about when you're writing a SQL query 00:09:29.300 |
Then you write to answer a different question, 00:09:37.100 |
bring all this tables together into multidimensional table. 00:09:43.520 |
So it's just like the way that we can have measures 00:09:46.340 |
and dimension that can potentially be kind of, you know, 00:09:49.040 |
like used at the same time from a different angles. 00:09:58.220 |
but you recently released a link chain integration. 00:10:01.640 |
There's obviously more and more interest in, again, 00:10:07.340 |
So you've seen the chat GPT code interpreter, 00:10:09.740 |
which is renamed as like advanced data analysis. 00:10:18.460 |
what are like some of the use cases that you're seeing 00:10:27.540 |
- Yeah, so, I mean, you know, when it started to happen, 00:10:34.000 |
They just have a better technology for, you know, 00:10:37.800 |
So it kind of, it made sense to me, you know, 00:10:46.700 |
And that's, chat bot is one of the use cases. 00:10:49.780 |
I think, you know, like if you try to generalize it, 00:10:52.940 |
the use case would be how do we use a structured 00:10:56.980 |
or tabular data with, you know, like AI models, right? 00:11:06.260 |
and then model can, you know, like give you answers, 00:11:11.260 |
But the question is like how we go from just the data 00:11:18.180 |
Like in a SQL based warehouses to some sort of, you know, 00:11:26.980 |
It's like no way you can get away around not doing this. 00:11:32.700 |
or you come up with some framework or something else. 00:11:35.420 |
So our take is that, and my take is that semantic layer 00:11:38.020 |
is just really good place for this context to live 00:11:41.460 |
because you need to give this context to the humans. 00:11:43.620 |
You need to give that context to the AI system anyway, right? 00:11:48.620 |
And then, you know, like you teach your AI system 00:11:59.420 |
and some of the ways that having the semantic layer 00:12:03.460 |
- I feel like, imagine you're a human, right? 00:12:05.820 |
And you going into like your new data analyst at a company 00:12:19.780 |
without any, you know, like additional context 00:12:24.100 |
like in many cases they might have a weird names. 00:12:27.180 |
Sometimes, you know, if they follow some kind of 00:12:30.100 |
like a star schema or like a Kimball style dimensions, 00:12:34.180 |
because you would have facts and dimensions column, 00:12:48.820 |
to give context to the data so people can understand that. 00:12:53.140 |
And I think the same applies to the AI, right? 00:13:00.180 |
it doesn't have this sort of context that it can read. 00:13:07.220 |
and give that book to the system so it can understand it. 00:13:10.940 |
- Can you run through the steps of how that works today? 00:13:14.940 |
So the initial part is like the natural language query, 00:13:18.100 |
like what are the steps that happen in between 00:13:26.900 |
- The first key step is to do some sort of indexing. 00:13:36.220 |
Like describe in a text format what your data is about, 00:13:50.180 |
So sort of, you know, like build a really good indexed 00:13:52.940 |
as a text representation and then turn it into embeddings 00:14:00.700 |
Once you have that, then you can sort of, you know, 00:14:30.300 |
Because what usually happens is that your query 00:14:37.780 |
by that dimension and maybe that filter should be applied. 00:14:46.100 |
a lot of different, you know, like techniques, 00:14:56.260 |
the more room that the model can make an error, right? 00:14:59.660 |
Like even sometimes it could be a small error 00:15:01.940 |
and then, you know, like your numbers is going to be off. 00:15:11.580 |
and then it executes us against semantic layer. 00:15:14.340 |
That's semantic layer executes us against your warehouse 00:15:17.540 |
and then sends result all the way back to your application. 00:15:24.140 |
because what we were missing was just about this ability 00:15:27.140 |
to have a conversation, right, with the model. 00:15:31.580 |
and then system can do a follow-up questions, 00:15:34.620 |
you know, like then do a query to get some information, 00:15:37.540 |
additional information based on this information, 00:15:40.820 |
And sort of, you know, like it can keep doing this stuff 00:15:42.940 |
and then eventually maybe give you a big report 00:15:48.220 |
But the whole flow is that it knows the system, 00:15:53.380 |
because you already kind of did the indexing. 00:16:03.340 |
for people that haven't used a semantic layer before, 00:16:24.020 |
and then it turns into a bigger and bigger query. 00:16:27.820 |
- One of the biggest difficulties around semantic layer 00:16:30.380 |
for people who've never thought about this concept before, 00:16:39.180 |
who all have different concepts of what a revenue is, 00:16:44.900 |
And then so they'll have like revenue revision one 00:16:48.540 |
by the sales team and then revenue revision one, 00:16:54.380 |
I feel like I always want semantic layer discussions 00:16:57.980 |
to talk about the not so pretty parts of the semantic layer 00:17:02.260 |
because this is where effectively you ship your org chart 00:17:11.900 |
and in Qubit, it's essentially a code base, right? 00:17:19.540 |
We know that, we're like software engineers, right? 00:17:23.220 |
You will have a lot of, you know, like revisions of code. 00:17:30.740 |
and we are in semantic layer as a code, right? 00:17:36.100 |
You know, like if there are like a multiple teams 00:17:43.540 |
Like why they think that should be calculated differently. 00:17:48.780 |
You know, like when everyone can just discuss it 00:17:54.140 |
why that code is written the way it's written, right? 00:17:58.620 |
And then hopefully at some point you can come up, 00:18:03.220 |
Now, if you still have multiple versions, right? 00:18:11.420 |
is that like we really need to treat it as a code base. 00:18:15.860 |
not as spreadsheets, you know, like a hidden Excel files. 00:18:21.820 |
then having the definition spread in the organization, 00:18:24.940 |
you know, like versus everybody trying to come up 00:18:28.980 |
But yeah, I'm sure that when you talk to customers, 00:18:31.420 |
there's people that, you know, have issues with the product 00:18:44.780 |
How important is the natural language to people? 00:18:51.140 |
in modern data stack companies either now or before. 00:19:05.580 |
and having non-technical folks do more of the work. 00:19:08.300 |
Are you seeing that as a big push too with these models, 00:19:12.100 |
like allowing everybody to interact with the data? 00:19:21.660 |
like where you have a lot of inside the question. 00:19:32.140 |
We have a company that built as internal Slack bot 00:19:52.460 |
I think it's really hard to tell them apart at this point 00:20:02.500 |
you know, like kind of even at least a pet project. 00:20:05.260 |
So that's what happened in Krizawa community as well. 00:20:07.420 |
We see a lot of like people building a lot of cool stuff 00:20:13.980 |
and kind of to see like what are real, the best use cases. 00:20:23.940 |
So they essentially connect into Q semantic layer 00:20:52.740 |
But other dimension of your question is like, 00:21:09.300 |
but it's more like a co-pilot for a data analyst, 00:21:13.980 |
where you develop something, you develop a model 00:21:16.260 |
and it can help you to write a SQL or something like that. 00:21:19.100 |
So, you know, it can create a boilerplate SQL 00:21:23.900 |
which is fine because you know how to edit SQL, right? 00:21:28.580 |
but it will help you to just generate, you know, 00:21:30.860 |
like a bunch of SQL that you write again and again, right? 00:21:37.660 |
I think that's great and we'll see more of it. 00:21:39.740 |
I think every platform that is building for data engineers 00:21:43.660 |
will have some sort of a co-pilot capabilities 00:21:54.820 |
to have some sort of, you know, like a co-pilot capabilities. 00:22:02.460 |
how do we enable access to data for non-technical people 00:22:05.860 |
through the natural language as an interface to data, right? 00:22:11.780 |
it's always has been an interface to data in every BI. 00:22:15.580 |
Now, I think we will see just a second interface 00:22:22.900 |
many BI's will add it as a commodity feature. 00:22:25.620 |
It's like Tableau will probably have a search bar 00:22:28.900 |
at some point saying like, "Hey, ask me a question." 00:22:31.420 |
I know that some of the, you know, like AWS QuickSight, 00:22:37.900 |
in their like BI and I think Power BI will do that, 00:22:45.540 |
some sort of a search capabilities built in inside their BI. 00:22:48.820 |
So I think that's just going to be a baseline feature 00:22:51.060 |
for them as well, but that's where Kube can help 00:23:08.420 |
Yeah, do you just see everything will look the same 00:23:28.260 |
And every major vendor and most of the vendors 00:23:31.100 |
will try to have some sort of natural language capabilities 00:23:40.540 |
Some of them will just have them as a checkbox, right? 00:23:43.980 |
So we'll see, but I don't think it's going to be something 00:23:54.700 |
rather than, you know, like what we have right now. 00:23:59.660 |
- Let's talk a bit more about application use cases. 00:24:03.620 |
So people also use Kube for kind of like analytics 00:24:06.900 |
on their product, like dashboards and things like that. 00:24:14.420 |
especially like when it comes to like agents, you know, 00:24:16.540 |
so there's like a lot of people trying to build agents 00:24:23.700 |
you need to know everything about the purchasing history 00:24:33.900 |
think about when implementing data into agents? 00:24:41.740 |
One is how to make sure that agents or LLM model, right, 00:24:46.740 |
has enough context about, you know, like a tabular data. 00:24:50.740 |
And also, you know, like how do we deliver updates 00:25:04.540 |
And how do you make sure that the queries are correct? 00:25:09.940 |
in this all, you know, like AI kind of, you know, 00:25:13.940 |
that we don't, you know, provide our own counselors? 00:25:16.940 |
But I think, you know, like kind of be able to reduce 00:25:24.100 |
like to try to like minimize potential damage. 00:25:28.740 |
And then, yeah, I feel like our use case, you know, 00:25:40.460 |
So I don't think that much is going to change 00:25:43.560 |
is that I feel like, again, more and more products 00:25:58.140 |
but also some sort of, you know, like summaries, 00:26:02.900 |
you're going to open the page with some surface stats 00:26:06.540 |
and you will have a smart summary kind of generated by AI. 00:26:09.580 |
And that summary can be powered by Kube, right? 00:26:11.820 |
Like, because the rest is already being powered by Kube. 00:26:14.660 |
- You know, we had Linus from Notion on the pod 00:26:18.180 |
and one of the ideas he had that I really like 00:26:23.500 |
kind of like how do you like compress knowledge 00:26:27.700 |
A lot of that comes into dashboards, you know, 00:26:34.060 |
hey, this is like the three lines summary of it. 00:26:36.820 |
Yeah, and yeah, makes sense that you would want to power that. 00:26:42.220 |
So are you, how are you thinking about, yeah, 00:26:49.900 |
what's like the future of what people are going to do? 00:26:53.260 |
What's the future of like what models and agents 00:27:03.420 |
I mean, it's obviously a big crossover between AI, 00:27:06.460 |
you know, like ecosystem, AI infrastructure ecosystem 00:27:15.900 |
like I'm looking at a lot of like what's happening 00:27:23.020 |
we use BI's, you know, different like transformation tools, 00:27:26.780 |
catalogs, like data quality tools, ETLs, all of that. 00:27:30.500 |
I don't see a lot of being compacted by AI specifically. 00:27:35.020 |
I think, you know, that space is being compacted 00:27:40.700 |
yes, we'll have all those copilot capabilities, 00:27:45.500 |
but I don't think see anything sort of dramatically, 00:27:48.860 |
you know, being sort of, you know, a change or shifted 00:27:57.220 |
I think, you know, like in the last two, three years, 00:28:09.900 |
And, you know, like, I mean, if Fivetran and DBT merge, 00:28:13.900 |
they can be Alteryx of a new generation or something like, 00:28:18.100 |
- And, you know, probably some ETL too there, 00:28:23.460 |
I mean, it just natural waves, you know, like in cycles. 00:28:26.940 |
- I wonder if everybody is gonna have their own copilot. 00:28:29.660 |
The other thing I think about these models is like, 00:28:31.940 |
you know, SWIX was at AirByte and yeah, there's Fivetran. 00:28:41.980 |
- There's the, you know, a lot of times these companies 00:28:48.660 |
of like building the integration between your data store 00:28:53.340 |
I feel like now these models are pretty good at coming up 00:28:58.500 |
and like using the docs to then connect the two. 00:29:07.580 |
I mean, you think about DBT and some of these tools 00:29:10.620 |
and right now you have to create rules to normalize 00:29:23.380 |
But yeah, I think it all needs a semantic layer 00:29:28.460 |
as far as like figuring out what to do with it, you know, 00:29:34.900 |
like workflows will be augmented by, you know, 00:29:40.300 |
You know, you can describe what transformation 00:29:43.100 |
you want to see and it can generate a boilerplate, right, 00:29:47.340 |
Or even, you know, like kind of generate a boilerplate 00:30:10.260 |
can be augmented quite significantly with all that stuff. 00:30:14.500 |
- I think the other important thing with data too 00:30:20.420 |
the big thing with machine learning before was like, 00:30:26.180 |
And I think like now at least with these models, 00:30:31.940 |
and they can also tell you if your data is bad, you know, 00:30:34.500 |
which I think is like something that before you didn't, 00:30:37.420 |
Any cool apps that you've seen being built on, on Cube, 00:30:51.140 |
it's definitely like, they all remind me of Statsbot, 00:31:03.580 |
It's just that use case that you really want, you know, 00:31:06.180 |
think about your data engineer in your company, 00:31:15.540 |
You know, like, so they will ping that bot instead. 00:31:18.820 |
So it's like, that's why a lot of people doing this. 00:31:23.620 |
But I think inside that use case, people get creative. 00:31:26.580 |
So I see bots that can actually have a dialogue with you. 00:31:29.500 |
So, you know, like you would come to that bot and say, 00:31:32.220 |
And the bot would be like, "What kind of metrics? 00:31:40.460 |
You want to see active users, you know, like sort of cohort, 00:31:54.660 |
and that's how many data analysts work, right? 00:32:00.300 |
you always try to understand what exactly do you mean? 00:32:18.380 |
and you can write exact query to your data warehouse, right? 00:32:21.940 |
So many people like say a little bit in the middle, 00:32:27.460 |
they usually have the knowledge about your data. 00:32:30.260 |
And that's why they can ask follow-up questions 00:32:34.620 |
And I saw people building bots who can do that. 00:32:39.860 |
I mean, like generating SQL, all of that stuff, 00:32:44.300 |
but when the bot can actually act like they know your data 00:32:57.180 |
You know, one of the big complaints people have 00:32:59.420 |
is like GPD, at least three and a half cannot do math. 00:33:02.500 |
You know, have you seen any limitations and improvement? 00:33:11.060 |
because it's like the best at this kind of analysis? 00:33:13.500 |
- I think I saw people use all kinds of models. 00:33:19.220 |
I mean, inside GPT, it could be 3.5 or 4, right? 00:33:22.620 |
But it's not like I see a lot of something else, 00:33:27.100 |
maybe know like some open source alternatives, 00:33:32.220 |
like it feels like the market is being dominated 00:33:40.540 |
I think I've been chatting about it with a few people. 00:33:47.780 |
you know, like outside of, you know, like chat GPT itself. 00:33:50.580 |
So it would be like some additional Python scripts 00:33:53.940 |
When we're talking about production level use cases, 00:33:57.660 |
it's quite a lot of Python code around, you know, 00:33:59.860 |
like your model to make it work, to be honest. 00:34:06.820 |
For like a toy use cases, the one we have on a, you know, 00:34:09.180 |
like our demo page or something, it works fine. 00:34:12.340 |
But you know, like if you want to do like a lot 00:34:16.020 |
you probably need to code it in Python anyway. 00:34:21.220 |
We heard the same from Harrison and Langstream 00:34:30.340 |
And it was funny to like just see the reaction 00:34:40.780 |
You're kind of like at the cutting edge of this, you know, 00:34:43.300 |
if I'm looking to build a data-driven AI application, 00:34:47.580 |
I'm trying to build data into my AI workflows. 00:35:00.700 |
I think a lot of people feel that MySQL can be a warehouse, 00:35:10.540 |
So just kind of have it starting with a good warehouse, 00:35:14.740 |
that's probably like something I would recommend 00:35:34.620 |
I feel it's a very interesting space, you know, 00:35:40.140 |
I see a lot of people using link chain right now, 00:35:45.500 |
but I'm sure the space will continue to evolve. 00:35:51.620 |
like some tools would be a better fit for a job. 00:35:57.020 |
but it's always interesting to see how it evolves. 00:36:05.620 |
documenting all that stuff will kind of evolve too. 00:36:09.100 |
But yeah, again, it's just like really interesting 00:36:15.380 |
- Okay, so before we go to the lightning round, 00:36:17.700 |
I wanted to ask you on your thoughts on embedded analytics. 00:36:39.020 |
embedded analytics is basically user facing dashboards 00:36:48.420 |
it's an individual user seeing their own data 00:36:53.220 |
that is owned by the platform that they're using. 00:36:58.940 |
but actually overwhelmingly the observation that I've had 00:37:02.420 |
is that people who try to build in this market 00:37:21.940 |
like a chatbots for their internal data consumption 00:37:28.660 |
because it's historically been dominated by the BI vendors. 00:37:33.660 |
And we still, you know, like see a lot of, you know, 00:37:39.020 |
like organizations are using BI tools as a vendors. 00:37:49.340 |
to the embedded analytics capabilities as well, right? 00:37:56.620 |
Also, you know, if you look at the embedded analytics market 00:38:03.940 |
they're really more custom, you know, like it becomes. 00:38:10.500 |
and they just kind of build most of the stuff from scratch, 00:38:13.860 |
which probably, you know, like the right way to do it. 00:38:16.020 |
So it's sort of, you know, like you got a market 00:38:21.660 |
and then you also in that middle and small segment 00:38:32.900 |
embedded analytics therefore is fragmented also. 00:38:36.340 |
So you're really going after the mid-market slice 00:38:39.980 |
and then with a lot of other vendors competing for that. 00:38:43.060 |
So that's why it's historically been hard to monetize, right? 00:38:58.140 |
So it's going to be more like a commodity feature 00:39:07.820 |
One is about acceleration, one on exploration 00:39:14.380 |
that already happened in AI or maybe, you know, 00:39:17.660 |
in data that you thought would take much longer 00:39:22.140 |
- To be honest, all this foundational models, 00:39:26.820 |
we had a lot of models that had been in production 00:39:29.940 |
for like quite, you know, maybe decade or so. 00:39:39.020 |
And even when we're building stats bot back then in 2016, 00:39:49.940 |
like a Google Translate or something that was, 00:39:53.940 |
But it was very customized with a specific use case. 00:39:56.820 |
So I thought that would continue for like many years. 00:40:00.660 |
We'll use AI, we'll have all this customized niche models. 00:40:07.900 |
They like, they can serve many, many different use cases. 00:40:17.420 |
- And the next question is about exploration. 00:40:22.340 |
is the most interesting unsolved question in AI? 00:40:24.780 |
- I think AI is a subset of software engineering in general. 00:40:29.780 |
And it's sort of connected to the data as well. 00:40:33.100 |
And in software, because software engineering 00:40:46.620 |
And now AI, I don't think it's completely different, 00:40:51.620 |
You know, like it's quite much not idempotent, right? 00:40:59.580 |
So which kind of may require a different methodologies, 00:41:03.820 |
may require different approaches in a different toolkit. 00:41:10.100 |
I think many sort of, you know, like tools and practices 00:41:19.780 |
But it's might be a very interesting subfield, 00:41:27.580 |
So now like AI is kind of feels like it's shaping into that 00:41:42.100 |
How do we test, you know, like what is the best practices? 00:41:45.780 |
So I think that would be an interesting to see. 00:41:52.180 |
you have a big audience of engineers and technical folks. 00:41:56.180 |
What's something you want everybody to remember, 00:42:00.380 |
- It says being who could try to build a chatbot, 00:42:25.780 |
or so that we actually now can build a smart agents. 00:42:28.460 |
I think that's sort of, you know, like a takeaways. 00:42:30.540 |
And yeah, we are, as you know, like as humans in general, 00:42:35.340 |
we're like, we really move technology forward 00:42:38.340 |
and it's fun to see, you know, like it's just a firsthand. 00:42:41.420 |
- Ah, well, thank you so much for coming on Artem.