back to indexYour Personal Open-Source Humanoid Robot for $8,999 — JX Mo, K-Scale Labs

00:00:19.280 |
We build open source human robots from hardware 00:00:29.360 |
JXX LIU: Yeah, so humanoids have been getting a lot of hype 00:00:38.300 |
They're quite proprietary, and they're quite expensive. 00:00:42.820 |
is because of very big problems, like physical labor shortage, 00:00:46.700 |
consumer household, and also like off-world exploration 00:00:55.540 |
is to really solve general purpose robotics for everyone, 00:00:58.720 |
and open sourcing the entire stack to the entire world. 00:01:02.240 |
So everyone will be benefiting from this really, really 00:01:04.760 |
useful technology instead of a few different companies. 00:01:10.900 |
Yeah, and our team is about 15 people in Palo Alto. 00:01:15.900 |
We're launching some robots in the next coming month. 00:01:19.020 |
And yeah, I'll be demoing some robots in the slides. 00:01:25.940 |
So we're currently working on two robots, the K-Bot and Z-Bot. 00:01:35.060 |
It has a full aluminum body, runs an RL controller 00:01:46.300 |
And this will also be one of the cheapest humanoid robots 00:01:49.640 |
And it's ready for pre-order right now, delivered by October. 00:02:08.240 |
basically, you can do VR teleoperation with building hands. 00:02:26.760 |
we thought about how do we make humanoids scale in the future? 00:02:42.460 |
And it's going to be the most affordable developer 00:02:48.880 |
The next cheapest option is probably at $40,000, 00:02:53.920 |
This is roughly the cost of most robot arms today. 00:02:57.800 |
Like if you buy UR5, that's, I think, $15,000, et cetera. 00:03:02.320 |
So the entire robot is going to be open source. 00:03:04.800 |
What that means is we're going to have open bomb. 00:03:07.160 |
Every single piece of the hardware, CAD design, electronics, 00:03:10.100 |
PCB, software, machine learning models will be fully open sourced. 00:03:17.920 |
And it's also going to be very, very modular, which means 00:03:20.700 |
that you can basically change out end effectors. 00:03:23.880 |
We have different mechanical designs that you can easily 00:03:32.040 |
for a parallel gripper instead of a five fingers hand. 00:03:35.920 |
or you can use whatever end effector you want to use, 00:03:41.000 |
And that means we can also easily upgrade and also fix the robot. 00:03:44.800 |
So our goal with selling this to developers is that when we have new hardware updates, 00:03:49.200 |
you can easily just re-screw the robot in with brand new legs or brand new arms 00:03:56.760 |
So, you know, compute improves, like you get new NVIDIA chips, 00:04:00.440 |
you can easily just add a new head onto the robot. 00:04:03.240 |
Yeah, we're also building--we have built an entire Python/Rust SDK for people to use. 00:04:11.120 |
If you come visit us, you can start programming this robot basically immediately. 00:04:14.920 |
It's like pip install package and you can start working on it. 00:04:17.880 |
And it's capable of running the latest state-of-the-art ML algorithms. 00:04:22.840 |
So in terms of local motion, you can use like NVIDIA, Isaacson to train a PPO policy. 00:04:28.680 |
We use MJX, which I'll explain a bit later, but you can use all kinds of different frameworks. 00:04:34.160 |
You can also run like different VLMs, like language models on the robot. 00:04:37.760 |
Maybe not locally directly, but you can run it through cloud. 00:04:45.160 |
So this robot, by default, it's going to be five DOF arms, 00:04:48.960 |
but you can easily interchange to a seven DOF arm. 00:04:51.880 |
So that'll suit most of the research need, research labs need. 00:04:55.400 |
And we'll make continuous like model and software improvements with OTA rollouts. 00:05:00.680 |
So every week, basically, we'll make new changes to the software as we go. 00:05:09.240 |
I can't release the full spec yet because we're launching soon. 00:05:11.720 |
But yeah, you see, it uses MIT cheetah actuators, pretty standard components with like MUs, 00:05:20.520 |
just different audio modules, displays, cameras, and up to 250 TOPS compute currently. 00:05:30.680 |
Yeah, and we really started this project in about October last year. 00:05:39.480 |
And we have a new, basically, design that we completed that I can't show you now, 00:05:45.800 |
Yeah, so we started off this -- the KBOT is like the KScale Stompy project, 00:05:50.600 |
like a full-sized humanoid robot that's 3D printable. 00:05:53.480 |
Then we move to like a prototype, and we work with different manufacturing partners to actually make 00:06:00.840 |
Yeah, it's launching soon if you're interested in the KBOT. 00:06:17.960 |
Yeah, and so what if you can't spend $9,000 on a cool humanoid robot? 00:06:22.680 |
What if your, like, wife or husband doesn't allow you to do it? 00:06:26.520 |
Well, introducing the ZBOT, which is a 1.5 feet humanoid robot that we also made at KScale Labs. 00:06:36.280 |
So this started from a hackathon project we did. 00:06:38.600 |
It became really popular on Twitter and also WeChat. 00:06:42.120 |
And so we're bringing this robot to mass manufacturing as well. 00:06:47.000 |
It runs the same locomotion and software stack. 00:06:49.480 |
Like, that means you can basically program stuff for the small robot, 00:06:56.840 |
So, you know, if you make, like, a voice chatting app, 00:07:06.200 |
Yeah, we really got inspired by the Google DeepMinds robot soccer paper, 00:07:16.600 |
So, yeah, we have a pretty good-- the launch went really well for the 3D printing one. 00:07:25.400 |
I think a few hundred people have actually made a 3D printed one on the orange one on the bottom. 00:07:31.400 |
Yeah, so we also started this project in November, and we are already bringing it into mass manufacturing, 00:07:41.160 |
Some people-- yeah, we also run, like, monthly hackathons, so you can just come try out the robot. 00:07:53.800 |
We talked about the hardware components we just open sourced. 00:07:57.800 |
Oh, yeah, also Zbot will be fully open sourced as well. 00:08:00.840 |
And so we also open sourced our entire ML and software stack. 00:08:05.960 |
So really, like, our core angle is basically to make this-- make the kbot autonomous. 00:08:13.320 |
Well, so, you know, it's a pretty standard dual policy. 00:08:17.080 |
You have the high-level controller, which is a VLA. 00:08:19.320 |
Then you have the RL whole body locomotion policy. 00:08:25.400 |
So what we really want right now is to basically finish-- we're currently working on both, basically. 00:08:32.760 |
And we also made our own firmware/software architecture to power these robots in Rust. 00:08:39.480 |
Our end goal is basically to make the robot so easy to use. 00:08:46.200 |
So, you know, Python application that you can re-share with people. 00:08:49.240 |
You make the robot do some very specific use cases that can be reused by other people. 00:08:56.200 |
And to do that, basically, we offer a lot of really cool developer tools we've been working 00:09:03.160 |
So we open sourced the library for, basically, GPU-accelerated robot learning. 00:09:08.440 |
Well, it's mostly like locomotion manipulation training. 00:09:13.720 |
And, yeah, the video of you saw us kicking the robot, it runs the controller-- 00:09:18.840 |
RL controller-- RL model that we trained in this training framework. 00:09:25.480 |
We also are working on to basically being able to integrate and fine-tune all the different VLA 00:09:31.240 |
and generalist policies that you see from Pi Zero and also NVIDIA group 00:09:38.520 |
you know, we're trying to make infrastructure very easy to run any cool models that you see 00:09:47.240 |
We also made this operating system, which is like a software framework plus like a Python 00:09:52.040 |
interface that you can use to program the robot using Python or Rust. 00:09:57.800 |
I don't know, if you guys use Rust 1, Rust 2, they're pretty hard to set up. 00:10:01.080 |
But using our system, you can just install Python package, like pip install KOS, 00:10:09.960 |
And we also have a digital twin in simulation. 00:10:14.120 |
It has the same gRPC interface you can use for controlling robot in simulation. 00:10:18.920 |
And all you have to do between programming something in simulation and real is by changing the IP address. 00:10:24.680 |
So you can prototype really, really rapidly without having to worry about breaking the robot, 00:10:33.960 |
You can actually program the robot just by using KOS-SIM. 00:10:38.200 |
Oh, I don't know what happened to the images. 00:10:44.040 |
And then, basically, what the machine learning and the operating system layers enables for us 00:10:49.720 |
to run different policies, VLA models, and on our robot hardware. 00:10:54.280 |
And for people to develop really cool applications with. 00:10:58.840 |
I'm just going to go through like a very, very quick RL training and deployment examples 00:11:02.840 |
of how researchers and developers could use our robot to train a local manipulation policy 00:11:09.080 |
for the robot to, you know, grab different things or walk around or even dance. 00:11:19.560 |
And then, all you have to do is run python-m train. 00:11:22.920 |
And in this train.py, you effectively have all the training code you need abstracted. 00:11:32.920 |
And then, basically, you can run this on like, you know, using run-power your local GPU. 00:11:38.200 |
So, it's like, you know, it's accelerated, accelerated compute. 00:11:42.520 |
And training and walking policy roughly takes one hour to two hours. 00:11:47.480 |
So, we're just going to run through like millions of different, not examples. 00:11:52.360 |
Iterations of the robot performing tasks you want. 00:11:54.600 |
And you can tune the reward functions and et cetera. 00:11:57.560 |
And you can see the loss and reward functions in our observability. 00:12:05.480 |
And afterwards, when the robots finish training, you can easily evaluate it in KOSM. 00:12:11.080 |
So, all you have to do is like, k-inverse-sim, like, you run the policy. 00:12:20.200 |
If it's walking, if you can see if it's doing the thing you want it to be doing. 00:12:24.280 |
For example, like walking, standing, picking up objects. 00:12:28.440 |
And if that's really good in simulation, then you can just easily change the IP address. 00:12:40.840 |
Like, you can basically get a robot to work in like, one-tenth of the time of like, 00:12:45.800 |
what you would take to set up most training libraries right now. 00:12:53.480 |
We do everything from hardware to software to ML. 00:12:56.040 |
So, how we're able to do this is actually by working with our open source community. 00:13:01.080 |
Currently, we have about like 5,000 active Discord members in a few servers. 00:13:07.000 |
We have a lot of public open bounties that people are tackling. 00:13:10.760 |
And because of our software's MIT lessons, yeah, a lot of people are coming to help us. 00:13:16.440 |
And we also run hackathons almost on a bi-monthly basis that a lot of people come to participate. 00:13:24.120 |
So, we're also hiring electrical firmware and ML engineers. 00:13:27.800 |
So, if you're interested, feel free to ask me. 00:13:31.800 |
And then, also, go on the kscout.dev/joint website. 00:13:35.160 |
Yeah, we're trying to hire a lot more cracked people to join us. 00:13:42.040 |
So, we're launching the robots and software stack in about two, three weeks. 00:14:08.600 |
I don't know if I can show you in the picture. 00:14:16.840 |
Yeah, the weight of the battery versus longevity. 00:14:34.440 |
Yeah, walking so far, our test is about two hours. 00:14:41.800 |
So, it's, yeah, you can just keep letting it charge and also run at the same time. 00:14:53.000 |
What are the use cases that you guys are imagining to start off with? 00:14:58.920 |
Is this going to be more for like commercial, like, you know, factory kind of use cases? 00:15:02.680 |
Or do you envision this more in the home, like helping out? 00:15:05.640 |
Yeah, so basically, our bet, so a lot of companies, like especially in the US, are betting on like B2B. 00:15:12.760 |
So, like Figure, for example, are selling to factories. 00:15:17.560 |
For us, our really bet is becoming the first US consumer robotics company. 00:15:24.920 |
Yeah, so we're really selling it to anyone that's interested in developing robotics. 00:15:28.920 |
So, a lot of our current customers, we accidentally launched our robots. 00:15:33.320 |
Like, people accidentally started buying our robots through our Shopify page. 00:15:39.000 |
But a lot of people that bought were just people genuinely interested in using for household tasks, 00:15:47.000 |
And also, there are a lot of companies also interested in working with us to make B2B businesses. 00:15:52.840 |
Like, for example, the food demo that we just saw. 00:15:59.080 |
Like, what household chores do you think it would be well suited for? 00:16:06.200 |
I mean, right now, we don't have any, I don't think anyone really has a fully working VLA model yet. 00:16:11.720 |
So, right now, it's pretty limited to teleoperating system, sorry, teleoperation. 00:16:17.000 |
But soon, we hope to be able to have, like, this navigation VLA stack for you to do, like, 00:16:22.040 |
you know, folding clothes or, like, doing dishwashing as the model capabilities improve. 00:16:37.720 |
So, you kind of alluded to complexity of ROS2 in terms of the setup. 00:16:44.280 |
I'm wondering if there are other, like, benefits and trade-offs that you considered 00:16:48.440 |
for foregoing something like ROS and ROS2 in that ecosystem? 00:16:57.320 |
So, ROS is really good because, like, the nodes and stuff, right? 00:17:00.120 |
So, like, the async, like, the communication. 00:17:02.440 |
But for our robot, we really don't have that many sensors. 00:17:05.720 |
And we really want to do this, like, model-based, like, policy-based robot. 00:17:09.720 |
So, we don't have many complicated sensors that we need to, like, async-communicate at all times. 00:17:14.600 |
The other part is, like, where I'm pretty opinionated. 00:17:22.520 |
I've just had a pretty bad experience using it, having to set up Ubuntu, you know. 00:17:27.640 |
Like, I just want a robot that I can just buy, open the box, it stands or walks, 00:17:33.240 |
and then I can just start programming it using my computer. 00:17:36.360 |
What kind of AI accelerator is on each of the robots? 00:17:46.840 |
So, basically, for the KBOT, it's going to be JSON, Nano, and AGX. 00:17:53.720 |
So, yeah, there are different compute options you'll be able to choose when we launch. 00:18:09.400 |
So, the preferred option for a lot of people is VR headset. 00:18:11.960 |
We have this, like, we also turn, like, a pseudo-IK. 00:18:15.880 |
So, basically, it's, like, going to position using, like, an RL model, 00:18:25.000 |
So, you can move the hand gestures, you can click button to open and close gripper, 00:18:43.480 |
In terms of, like, mechanical powerness, like, you know, the Tesla's way more powerful. 00:18:49.160 |
It has, like, linear actuators and et cetera. 00:18:51.480 |
But, in terms of, like, actual use cases, I don't think it's really that different. 00:18:56.040 |
I mean, Tesla is actually built for, like, a factory type of use cases. 00:19:01.000 |
But, in terms of, like, you want to, for people to actually buy and use this robot, it's not very different. 00:19:05.400 |
I think the last time I heard Tesla Optimus is about 60k, at least. 00:19:14.920 |
But, our robot's $9,000 before mass production.