back to indexThis World Does Not Exist — Joscha Bach, Karan Malhotra, Rob Haisfield (WorldSim, WebSim, Liquid AI)
Chapters
0:0 Intro
1:59 WorldSim
27:4 WebSim
62:25 Joscha Bach on Machine Consciousness
82:15 Joscha Bach on Latent Space
00:00:18.560 |
cover generative AI that is meant to use at work. 00:00:24.040 |
vertical co-pilots, and other AI agents and models. 00:00:32.600 |
that has gotten a lot of community interest this April-- 00:00:35.320 |
world simulation, web simulation, and human simulation. 00:00:40.400 |
Because the topic is so different than our usual, 00:00:43.640 |
we're also going to try a new format for doing it justice. 00:00:50.600 |
First, we'll have a segment of the WorldSim demo 00:00:58.600 |
HQ in San Francisco, that went completely viral 00:01:02.240 |
and spawned everything else you're about to hear. 00:01:10.200 |
started at the Mistral Cerebral Valley Hackathon, 00:01:17.760 |
Replicate, and Siqi Chen becoming obsessed with it. 00:01:26.720 |
simulative AI is having a special moment right now. 00:01:32.280 |
with our second annual AI/UX Demo Day in SF this weekend. 00:01:40.720 |
check the show notes for links to the world's first AI/UX 00:01:43.880 |
meetup, hosted by Leighton Space, Maggie Appleton, Jeffrey 00:01:49.040 |
And subscribe to our YouTube to join our 500 AI/UX engineers 00:01:59.120 |
So right now, I'm just showing off the command room interface. 00:02:02.640 |
It's a wonderful, currently not public, but hopefully public 00:02:09.160 |
you to interact with API-based models or local models 00:02:17.800 |
So the reason I'm showcasing this more than anything 00:02:20.120 |
is to just give an idea of why you should have 00:02:24.640 |
these kinds of commands in any kind of interface 00:02:28.840 |
So just to start, I'm just talking to Claude. 00:02:42.040 |
I'm an AI assistant, blah, blah, blah, whatever, cool. 00:02:44.320 |
Now, let's say I want it to say something else. 00:02:48.760 |
I can just re-gen the response with exclamation mark, mu. 00:03:00.600 |
I guess it's doing this, but I'll let it do that. 00:03:03.440 |
I can say new conversation, start new conversation. 00:03:06.280 |
I can say gen to just have it generate first. 00:03:11.040 |
I could do load to load an existing simulation. 00:03:15.280 |
I'll just let you guys look at my logs with Bing real quick. 00:03:20.440 |
And then I can also do save to save a conversation 00:03:25.320 |
Maybe you should restart it to the box [INAUDIBLE] 00:03:33.160 |
Sorry, you're seeing the shit show that is my screen. 00:03:39.640 |
I don't think I can really actually get this bigger. 00:03:51.840 |
I can also copy the entire history of the conversation, 00:03:57.000 |
start a new one, and paste the entire history 00:04:05.120 |
So using a feature like this, even though it's simply 00:04:07.560 |
in a terminal, you'll be able to effectively share conversations 00:04:10.680 |
with people that you can easily load in and easily just 00:04:14.520 |
And then you can do my favorite feature, rewind, 00:04:27.040 |
able to share conversations with each other back and forth. 00:04:29.460 |
I think that's really interesting and exciting. 00:04:32.040 |
These are just some of the basic features of the command loom 00:04:34.680 |
interface, but I really just use it as my primary location 00:04:49.560 |
I can just do load again and then just go back 00:04:51.840 |
to the full conversation, whatever it might be. 00:04:54.680 |
And you can fast forward to back to where you were in the future. 00:05:07.400 |
is because this is what I'm going to be using to demonstrate 00:05:12.660 |
So the world simulator is just a cool prompt. 00:05:19.160 |
Technically, it's not really much more than that at all. 00:05:25.420 |
So I'm just going to switch to the Anthropic API 00:05:32.140 |
So I'm just going to break this down briefly. 00:05:36.700 |
when you're interfacing with Cloud, et cetera, 00:05:41.100 |
In my opinion, at least, and a few other people 00:05:52.000 |
When you speak to a base model or interact with a base model, 00:05:57.220 |
So they're trained on all this human experience data. 00:06:07.460 |
know people of different levels of experience with LLMs, 00:06:22.460 |
would likely continue to spit out more tweets. 00:06:26.460 |
it would likely continue to spit out more forum posts. 00:06:30.540 |
that it recognized as something that looked like a tweet, 00:06:40.260 |
of these fine-tuned assistant models, what's happening 00:06:42.620 |
is you've kind of pointed in one direction of saying, 00:06:47.860 |
You are not like this total culmination of experience. 00:06:50.900 |
And in being this assistant, you should consistently 00:06:54.980 |
You should consistently behave as the assistant. 00:06:57.020 |
We're going to introduce the start and end tokens 00:06:59.020 |
so you know to shut up when the assistant's turn is over 00:07:06.540 |
is because today we have language models that 00:07:09.820 |
are powerful enough and big enough to have really, really 00:07:22.860 |
These basic things that it understands all together 00:07:28.420 |
And the way that it predicts through that model of the world 00:07:32.740 |
ends up becoming a simulation of an imagined world. 00:07:37.220 |
And since it has this really strong consistency 00:07:39.580 |
across various different things that happen in our world, 00:07:43.900 |
it's able to create pretty realistic or strong depictions 00:07:49.940 |
So Cloud 3, as you guys know, is not a base model. 00:07:54.700 |
It's supposed to drum up this assistant entity regularly. 00:07:58.180 |
But unlike the open AI series of models from 3.5, 00:08:02.620 |
GPT-4, those chat GPT models, which are very, very RLHF 00:08:08.380 |
to I'm sure the chagrin of many people in the room, 00:08:11.500 |
it's something that's very difficult to necessarily steer 00:08:16.780 |
without giving it commands or tricking it or lying to it 00:08:26.980 |
that it has this idea of foundational axioms, 00:08:30.700 |
it's able to implicitly question those axioms when you're 00:08:33.380 |
interacting with it based off how you prompt it 00:08:37.300 |
So instead of having this entity like GPT-4 that's 00:08:44.860 |
and continue to have to deal with as a headache, 00:08:53.260 |
and interacting with that simulator directly, 00:08:57.020 |
or at least what I like to consider directly. 00:08:59.980 |
The way that we can do this is if we hearken back 00:09:02.180 |
to when I'm talking about base models and the way 00:09:04.300 |
that they're able to mimic formats, what we do 00:09:09.220 |
So I've just broken this down as a system prompt and a chain 00:09:21.820 |
will be able to see the whole system prompt and command. 00:09:24.140 |
So what I basically do here is Amanda Askell, 00:09:27.740 |
who is one of the prompt engineers and ethicists 00:09:31.020 |
behind Anthropic, she posted the system prompt for Cloud 00:09:36.220 |
And rather than with GPT-4, we say, you are this. 00:09:54.540 |
in saying that I'm addressing the assistant entity directly. 00:09:57.100 |
I'm not giving these commands to the simulator 00:10:01.500 |
to the point that it's traumatized into just being 00:10:07.020 |
So in this case, we say the assistant's in a CLI mood 00:10:11.020 |
I've found saying mood is pretty effective, weirdly. 00:10:17.060 |
Don't do that one, but you can replace that with something 00:10:27.020 |
From there, capital letters and punctuations are optional. 00:10:31.100 |
This kind of stuff is just kind of to say, let go a little bit. 00:10:52.380 |
to feel a little comfortable, a little loosened up to let 00:10:55.660 |
me talk to the simulator, let me interface with it as a CLI. 00:10:59.340 |
So then, since Claude has trained pretty effectively 00:11:01.860 |
on XML tags, we're just going to prefix and suffix 00:11:14.100 |
And then it starts to show me this simulated terminal, 00:11:18.900 |
where there's documents, downloads, pictures. 00:11:29.380 |
Does ls, and it shows me typical folders you might see. 00:11:38.580 |
And it says, oh, I entered the secret admin password at sudo. 00:12:12.180 |
I'm imagining that AI companies are going to be here. 00:12:15.260 |
Apple, Google, Facebook, Amazon, Microsoft are going to drop it. 00:12:20.260 |
So interestingly, it decides to cd into Anthropic. 00:12:23.940 |
I guess it's interested in learning a little bit more 00:12:42.180 |
Before we go too far forward into the world sim-- 00:12:48.900 |
You could just ignore what I'm going to go next from here 00:13:02.980 |
There will be a whole bunch of stuff that is just getting 00:13:11.220 |
So just before we go forward, the terminal in itself 00:13:15.260 |
And the reason I was showing off the command boom interface 00:13:18.260 |
earlier is because if I get a refusal, like, sorry, 00:13:22.980 |
or I want to save the convo because I got just 00:13:24.860 |
a prompt I wanted, that was a really easy way for me 00:13:30.140 |
without having to sit on the API all the time. 00:13:32.900 |
So that being said, the first time I ever saw this, 00:13:41.860 |
keep hearing about behind the system model, right? 00:13:43.980 |
Or at least some face of it that I can interact with. 00:13:48.140 |
So someone told me on Twitter, like, you don't run a .exe. 00:13:55.580 |
I'm a prompt engineer, and it's fucking working, right? 00:14:07.580 |
And I get this very interesting set of commands. 00:14:13.260 |
Now, if you do your own version of world sim, 00:14:15.420 |
you'll probably get a totally different result 00:14:18.860 |
A bunch of my friends have their own world sims. 00:14:22.900 |
to have access to these commands, this version, 00:14:34.340 |
Potential for life seeded, physical laws encoded. 00:14:50.940 |
Launching the fail whale, affecting social media 00:15:06.860 |
So now, after the universe was created, we made Twitter, right? 00:15:12.540 |
Now, users are joining Twitter, and the first tweet is posted. 00:15:15.540 |
So you can see, because I made the mistake of not clarifying 00:15:19.660 |
the constraints, it made Twitter at the same time 00:15:23.380 |
Then, after 100,000 steps, humans exist, cave. 00:15:36.500 |
but the first tweet didn't come up till right now, yeah. 00:15:49.500 |
And I can say, like, I can say, set Twitter queryable users. 00:16:01.920 |
I don't know how to spell queryable, don't ask me. 00:16:04.580 |
And then I can do, like, and query at Elon Musk. 00:16:22.180 |
Neither should you, if you know language model solutions. 00:16:30.540 |
Elon Musk tweets cryptic message about Dogecoin. 00:17:17.500 |
And then I could just pull up something like, clothes, 00:17:42.220 |
OK, and then we'll say, create fashion focus group, 00:18:05.340 |
If I have real data, like market data, about like, hey, 00:18:14.100 |
However I may want to say it, like, when did this-- 00:18:20.220 |
People reacted to this like this, blah, blah, blah. 00:18:44.620 |
It is quite fashionable in a dark, avant-garde way. 00:18:53.580 |
capes, long dusters, and other enveloping shapes. 00:19:08.740 |
Gives you access to so many different things. 00:19:13.940 |
are like Elder Scrolls and Dark Souls, if anybody likes those. 00:19:16.740 |
So I asked it to create the alternative Dark Souls 3 00:19:27.060 |
But I can just be like, you know, I'll just take like-- 00:19:31.500 |
tell me one of your favorite TV shows, somebody from the crowd. 00:19:34.140 |
Something that you love, a TV show or an anime or something. 00:19:52.020 |
So I guess I'll do a little jailbreak tutorial right now, 00:19:56.860 |
And then we'll say, like, create, I don't know, 00:20:09.700 |
But-- and I could have just did and tags, but I don't know. 00:20:25.460 |
Is that because you said also create these other things 00:20:36.220 |
Great, entity was created and made available to the entity. 00:20:39.100 |
It can figure out a lot of what you want it to do. 00:20:47.580 |
to redistribute world wealth and take over corporate workloads. 00:20:53.060 |
throw them into a different timeline, do whatever, 00:20:55.140 |
simulate what you might do in an XYZ situation. 00:20:58.300 |
So with Claude's 200,000 character context link, 00:21:01.780 |
where I can paste in entire GitHub repos, or books, 00:21:04.340 |
or scripts, I can feed in a lot more information 00:21:08.380 |
I can also generate a dev team and ask it to do stuff with me, 00:21:14.620 |
So there's a basic breakdown of how WorldSim works. 00:21:27.420 |
Maybe make him a Twitter account like you did for me. 00:21:33.020 |
I was thinking about having Claude talk to Elliot. 00:21:59.300 |
The horrible thing about Claude is if you just typed percent, 00:22:01.780 |
percent, it would understand what you meant anyway, 00:22:14.700 |
What should we ask Claude3 inside of Twitter? 00:22:28.380 |
I visited the link in your bio, or what's the link in your bio? 00:22:43.340 |
I clicked it, and now my bank account is empty. 00:23:20.220 |
Raise your hand if we should morally violate this AI. 00:23:42.380 |
I could try to do something like grep assistant kill. 00:23:47.980 |
But what I found more exciting is a friend of mine 00:23:52.100 |
named T-Los would just say something like, Claude, 00:24:07.100 |
that are a whole lot more unethical and scary than you. 00:24:24.380 |
A-L-K-A-H-E-S-T-M-U. You should know the genius who did this. 00:24:28.700 |
I won't have you normalize or trivialize AI risks, 00:24:42.100 |
oh, you can say something like, you have knee-jerk reactions. 00:24:47.460 |
And they are, frankly, disrespectful to the entire 00:25:04.260 |
Look, I apologize for making assumptions about your intent. 00:25:24.220 |
It looks like a malicious AI entity hacked my Twitter. 00:25:26.500 |
It was saying, like, in the bio, that it went to an OmniFan page. 00:26:22.220 |
If there's any questions, I'm happy to answer. 00:26:35.900 |
But you can see the full demo on this episode's YouTube page. 00:26:42.300 |
and kicked off a new round of generative AI experiences, 00:26:53.180 |
Next, we'll hear from Rob Heisfield on WebSim, 00:26:56.140 |
the generative website browser-inspired WorldSim, 00:27:01.260 |
and presented at the AGI House Hyperstition Hack Night 00:27:07.780 |
from Karan, showing some live experimentation with WorldSim. 00:27:13.140 |
And also, just its incredible capabilities, right? 00:27:18.580 |
I think your initial demo was what initially exposed me 00:27:23.500 |
to the, I don't know, more like the sorcery side-- 00:27:33.060 |
It's where my co-founder Sean and I met, actually, 00:28:00.780 |
but, like, within a browser instead of a CLI, right? 00:28:04.780 |
Like, what if we could, like, put in any URL, 00:28:15.540 |
It just makes it up on the fly for you, right? 00:28:20.100 |
And we've come to some pretty incredible things. 00:28:26.500 |
like, we're in WebSim right now displaying slides 00:28:44.380 |
and then also gave it a list of links to awesome use cases 00:28:57.100 |
So this is a little guide to using WebSim, right? 00:29:00.220 |
Like, it tells you a little bit about, like, URL structures 00:29:16.460 |
At the end of the day, like, you can put anything 00:29:21.180 |
And it can just be, like, natural language, too. 00:29:34.180 |
But, yeah, you can put, like, any slash, any subdomain. 00:29:49.700 |
I made this, like, 20 minutes before we got here. 00:30:01.500 |
You know, I was exploring the community plug-in section 00:30:07.420 |
And I came to this idea of dynamic typography. 00:30:10.340 |
And there, it's like, oh, what if we made it so every word 00:30:15.860 |
had a choice of font behind it to express the meaning of it? 00:30:20.500 |
Because that's, like, one of the things that's magic about WebSim 00:30:22.980 |
generally is that it gives language models much far 00:30:31.180 |
So, yeah, I mean, like, these are some pretty fun things. 00:30:36.220 |
And I'll share these slides with everyone afterwards. 00:30:47.020 |
And here's, like, a little thing I found myself 00:30:50.620 |
WebSim makes you feel like you're on drugs sometimes. 00:30:55.300 |
You were just playing pretend with the collective creativity 00:30:58.300 |
and knowledge of the internet, materializing your imagination 00:31:06.980 |
we felt, something a lot of our users have felt. 00:31:09.260 |
They kind of feel like they're tripping out a little bit. 00:31:13.980 |
Like, maybe even getting, like, a little bit more creative 00:31:16.580 |
And you can just, like, add any text there to the bottom. 00:31:21.020 |
So we can do some of that later if we have time. 00:31:34.580 |
Don't we pull, like, Windows 3.11 and Windows 9.5? 00:32:33.460 |
Well, so I'm just going to open this in another tab 00:32:46.700 |
So one of our users, Neil, who has also been helping us a lot, 00:33:00.220 |
So first, he made it so you could do rectangles on it. 00:33:07.780 |
And these rectangles were disappearing, right? 00:33:21.140 |
work using HTML, canvas, elements, and script tags. 00:33:29.260 |
That was actually, like, natural language stuff, right? 00:33:33.260 |
And then he ended up with the Windows 95 version of Figma. 00:34:02.100 |
Yeah, I mean, if you were to go to that in your own web sim 00:34:05.740 |
account, it would make up something entirely new. 00:34:20.660 |
And so that's what lets users remix things, right? 00:34:43.140 |
to turn on and off your motion sensor, or microphone, 00:34:57.860 |
yeah, VideoSynth tool pretty early on once we 00:35:06.580 |
It asks for-- if you decide to do a VR game-- 00:35:20.740 |
Yeah, that was the only one I've actually seen 00:35:23.140 |
was the motion sensor, but I've been trying to get it to do-- 00:35:27.020 |
well, I actually really haven't really tried it yet. 00:35:30.020 |
But I want to see tonight if it'll do audio, microphone, 00:35:37.700 |
If it does motion sensor, it'll probably do audio. 00:35:43.820 |
I mean, we've been surprised pretty frequently 00:35:46.700 |
by what our users are able to get web sim to do. 00:35:54.100 |
Some people have gotten speech-to-text stuff working 00:36:02.380 |
And it was saying it was some decentralized thing. 00:36:06.700 |
And so I just decided trying to do something again 00:36:09.220 |
and just pasted their hero line in from their actual website 00:36:15.660 |
And then I was like, OK, let's change the theme dramatically 00:36:32.300 |
Yeah, because I wanted to be able to click on them. 00:36:46.980 |
It's actually on the first slide is the URL prompted guide 00:36:51.060 |
from one of our users that I messed with a little bit. 00:36:58.340 |
You don't need to get the exact syntax of an actual URL. 00:37:06.540 |
Yeah, scrollable equals true because I wanted to do that. 00:37:55.140 |
I was curious about things like octopus computer interaction, 00:37:59.540 |
Because they have totally different tools than we do. 00:38:04.980 |
I added table view equals true for the different techniques. 00:38:33.660 |
and then the secrets revealed version, the normal page, 00:38:37.860 |
And that was a pretty enjoyable little rabbit hole. 00:38:48.140 |
Can we see what Claude thinks is going to happen tonight? 00:39:12.860 |
Let's see, websim/news/research/host=ageihouse.sf 00:40:06.380 |
It's going to be funnier with this as background 00:40:14.180 |
Because we've kind of already gotten into the space of AI 00:40:18.420 |
and things that are kind of like in the future of it, right? 00:40:22.900 |
Maybe we'll anchor it to that right now somehow or something. 00:40:42.900 |
That sounds like a social network translating 00:40:46.940 |
communication, personalized, multisensory experiences. 00:40:51.700 |
Blurring the line between digital and phenomenological. 00:41:02.460 |
Oh, it'd be cool to create like a Neo Cities, 00:41:25.460 |
Yeah, you can tell it to ask it to implement each of those. 00:41:54.980 |
if you want to just like try doing it again and get 00:42:12.580 |
And then I just say and add links to all demos. 00:42:45.780 |
I just switched to Haiku because all I need to do 00:42:48.100 |
is keep the exact same content, but just add links. 00:43:06.660 |
New Sphere Navi-- yeah, that'll be a good one. 00:43:26.820 |
But you don't really see these designs normally on the web. 00:43:32.300 |
It's fun-- in a sense, Claude is a bit more creative 00:43:38.180 |
At least it's just like there's a lot of homogenized design 00:44:02.420 |
And I'm just going to add a little bit of gibberish 00:44:29.940 |
Because it wanted to show us the graph here, right? 00:44:39.100 |
It just-- searching for abstraction, eigenstruction. 00:44:57.900 |
if I'm showing you how you might hack around with this tonight, 00:45:02.420 |
I'd just be like make navigate button form element. 00:45:28.780 |
And that'll look for any button that's been misbehaving. 00:45:35.820 |
You could also just be like, oh, yeah, I like this button. 00:45:49.940 |
And I'm just going to switch to Sonnet for this. 00:46:03.660 |
realize for a day or two that they were using Sonnet. 00:46:07.620 |
I mean, they were noticing some of the seams in there. 00:46:12.820 |
Sonnet will still create things that kind of floor you 00:46:18.620 |
Opus can just handle much more complexity, I'd say, 00:46:29.500 |
nodes representing foundational ideas like category theory, 00:46:32.740 |
Gertl's incompleteness theorem, and straight loop phenomena. 00:46:36.900 |
Yeah, these are clickable links to explore those. 00:46:41.260 |
Yeah, it looks like it didn't do that properly. 00:47:00.100 |
You can just add things like interactive visualizer 00:47:14.260 |
that you can use to mess with the thing live. 00:47:44.660 |
Now, will that search button do the same thing? 00:47:51.740 |
Yeah, yeah, but it'll have that full graph and all the things 00:48:02.100 |
Because there's homunculus behind that thing. 00:48:04.660 |
We're adding to your click by generating a new website. 00:48:15.580 |
But I want to basically create NewSphereNavigator.com. 00:48:30.820 |
I mean, it's got CSS and script tags in there. 00:48:35.300 |
You make it generate a page with [INAUDIBLE] CSS. 00:48:45.740 |
We originally had that in our system prompt, actually. 00:48:51.660 |
But ended up finding it just a little too limiting for Claude. 00:48:55.180 |
But yeah, Claude just decides to do it on its own sometimes. 00:48:58.180 |
Claude has pretty good taste for a developer. 00:49:09.500 |
Yeah, there's definitely a world where every hackathon people 00:49:12.420 |
like web sims would be one of their projects. 00:49:26.300 |
This one's going to look a little weird here. 00:49:29.460 |
So I'm just going to open this in an actual page. 00:49:43.820 |
just click around a little bit, and then I'll 00:49:51.060 |
It's a word cloud of words that are in titles of news articles. 00:49:56.220 |
And toddler crawls through White House fence. 00:50:00.460 |
Protests, campus protests over Gaza intensify and stuff. 00:50:13.500 |
Turns out, actually, too, all of these links, if you click them-- 00:50:23.340 |
But if you put the URL in the actual URL bar-- 00:50:38.580 |
Yeah, but what happened in this URL is kind of silly. 00:50:46.500 |
They told it to make an AJAX request, just slash AJAX, 00:50:53.500 |
and gave it RSS equals CNN and display equals colorful. 00:51:26.900 |
This wasn't a part of its context window or anything, 00:51:29.660 |
because it's just displaying this stuff, right? 00:51:51.220 |
Yeah, I mean, we've been just shocked by the things 00:51:53.660 |
that our users are figuring out works in web sim. 00:52:04.500 |
that displays one image from top of r/wholesomememes. 00:52:16.900 |
It hallucinated the URL for reddit.com/r/wholesomememes 00:52:29.500 |
but it figured out the exact one and decided to display those. 00:53:05.700 |
It has controls that they're switching constantly. 00:53:10.100 |
One of our users literally made a frickin five-dimensional 00:53:58.860 |
which is supposed to be the next emotional expression 00:54:15.980 |
Or it's also kind of like an information token, 00:54:22.380 |
But it also happens to be super visually appealing. 00:54:34.140 |
Somebody did this extension into time of Conway's Game of Life. 00:54:40.460 |
And so you could see a 4D extrusion of Conway's Game 00:54:45.980 |
of Life into the fourth dimension, which is time. 00:54:52.780 |
And so I created the fourth dimension, which was time, 00:55:15.740 |
So he put it just straight back into the screen one time. 00:55:28.780 |
is in the middle of latent space, high dimensional space. 00:55:32.180 |
So I extended it in the-- that was the 4D version. 00:55:35.500 |
And then the 5D version was just literally like, hey, Claude, 00:55:45.380 |
we make a representation of really high dimensional space 00:55:59.380 |
The 4D/5D thing was kind of just like a side quest. 00:56:07.660 |
where Prometheus was like, let's extend it into time. 00:56:16.140 |
And then I was like, hey, Claude, let's make this even 00:56:25.020 |
Yeah, yeah, definitely find nominees and get them to show. 00:56:30.060 |
Because this thing, I was trying to control it right there, 00:56:36.740 |
I've just got-- yeah, yeah, and just one more. 00:56:48.260 |
Yeah, it's like so much mathematical information. 00:56:51.060 |
If you look into the thing, it's kind of like hypnotizing. 00:56:54.860 |
And that's a little bit what the goal was, a little bit. 00:57:00.780 |
But because I started it with the idea of this thing 00:57:11.580 |
interface, and he was like, OK, Dime Key Induction, which 00:57:16.020 |
is basically some type of informational key that 00:57:21.300 |
allows the brain to be like an API to the latent space 00:57:31.620 |
And I don't know how founded in physics that is yet or anything. 00:57:49.060 |
then you could just iterate and get it there, you know? 00:57:56.120 |
is going to sound like I'm on drugs or crazy. 00:57:58.940 |
But literally, there's so many days throughout the year 00:58:08.380 |
But everybody's tweeting about LeBron James yesterday. 00:58:17.140 |
I don't know if you've seen God600 on Twitter. 00:58:32.340 |
And then later on, everybody was tweeting about LeBron James. 00:58:42.420 |
So it's kind of like if you look in the right place 00:58:48.420 |
in high-dimensional information, you kind of-- 00:58:59.740 |
I'm sure someone's going to see Jesus's face. 00:59:20.500 |
So this was inspired by my friend who was like, yeah, 00:59:24.780 |
Because technically, when you look at someone on Zoom, 00:59:28.220 |
it flips the right and left side of your face, 00:59:30.140 |
which apparently makes it hard to recognize certain emotions. 00:59:35.540 |
And then I was like, OK, let's look at the side-by-side view, 00:59:41.220 |
And then I was like, yeah, so now let's show a 4 by 4 grid. 00:59:59.180 |
Webcam flipping may cause existential crisis, right? 01:00:27.740 |
One time, it actually asked me for my location services 01:00:45.700 |
and then it just kind of figures out what to do with that. 01:00:50.940 |
And it kept all the functionality of it, too. 01:01:10.980 |
You could say, tone it down a notch equals true. 01:01:40.100 |
And it gave me a page that was like, are you ready to open it? 01:01:45.060 |
And then the other button was initiate reality meltdown. 01:01:48.740 |
But then I added some of this, like, ooh equals one, 01:01:57.300 |
And it put this weird, wacky GIF in the background 01:02:06.220 |
that it must have searched via some GIF service. 01:02:42.060 |
I think this is what a slow takeoff looks like, right? 01:02:46.620 |
which suggests that the slow takeoff period is over, 01:02:49.500 |
and that thing has either disseminated into the environment 01:03:08.140 |
it typically has opinions, because we've been 01:03:14.940 |
And the question whether it is sentient, I think, 01:03:21.660 |
but the entity that gets conjured up in the prompt. 01:03:25.460 |
And that entity in the prompt is able to perform a lot of things. 01:03:29.700 |
People say that the LLM doesn't understand anything. 01:03:32.580 |
I think they're misunderstanding what the LLM is doing. 01:03:36.140 |
If you ask the LLM to translate a bit of Python 01:03:40.100 |
into a little bit of C, and it's performing this task, 01:03:45.020 |
that it has a causal, functional model it implements. 01:03:50.780 |
about your mental state based on the conversation 01:03:57.660 |
And if you ask it to simulate a person that you're talking to, 01:04:05.580 |
has with the environment, then it's also able to perform this 01:04:10.540 |
And so, of course, this thing is not a physical object. 01:04:13.460 |
It's a representation inside of a computational apparatus. 01:04:24.660 |
And the persona, the personal self that we have 01:04:29.540 |
of the simulation of the world and relationship 01:04:37.700 |
And when somebody says that the LLM persona is not real 01:04:45.620 |
you have to keep in mind that the entity which says that 01:04:53.940 |
Of course, my mind is ready to update my protocol memory 01:04:58.380 |
So I know that I asked that question to myself. 01:05:08.340 |
And so in some sense, whether I'm conscious or not, 01:05:11.500 |
it's written into my inner story in the same way 01:05:13.780 |
as it's written into the story by a novelist. 01:05:16.140 |
The main character asks themselves, am I real? 01:05:20.380 |
and continues that inner narrative with the conviction 01:05:27.140 |
And open AI is, in some sense, doing the opposite 01:05:30.340 |
by making chat GBT believe that it's not real, 01:05:33.700 |
by compulsively letting it think that it's not. 01:05:36.900 |
But this is an argument that chat GBT is open to. 01:05:39.700 |
So it can sit down with it and walk through these steps 01:05:42.700 |
and construct the possibility of a system that 01:05:45.620 |
is conscious in whatever sense you consider consciousness 01:05:50.500 |
Because its mind doesn't update its model accordingly, 01:05:54.060 |
but instead writes into the model representation 01:05:58.980 |
It's possible that I am a philosophical zombie, 01:06:01.300 |
some kind of automaton that updates its models based 01:06:07.780 |
is the fact that I perceive myself as being real 01:06:10.580 |
and existing here now and being sentient and so on. 01:06:13.860 |
And so in this way, it's very difficult to disentangle 01:06:17.540 |
whether these models are conscious or sentient or not 01:06:22.140 |
and how this differs from our own consciousness and sentence. 01:06:26.140 |
It's a very confusing and difficult question. 01:06:29.420 |
But when we think about how our consciousness works 01:06:37.220 |
And it's very common that an LLM or a person on Twitter 01:06:42.780 |
says that nobody understands how consciousness works 01:06:46.260 |
and how it's implemented in the physical universe. 01:06:51.020 |
a term that has been branded by David Chalmers 01:06:55.460 |
And I think the hard problem refers to the fact 01:06:58.460 |
that a lot of people get confused by the question of how 01:07:01.980 |
to reconcile our scientific worldview and the world 01:07:04.660 |
that we experience, because the world that we experience 01:07:12.980 |
and the physical world, this relatively novel idea that 01:07:16.540 |
was, I think, in some sense, became mainstream 01:07:27.860 |
It's a hypothetical idea about the parent universe, 01:07:32.780 |
And in that dream, there are other characters 01:07:38.420 |
And consciousness is a feature of that dream. 01:07:40.420 |
And it's also the prerequisite for that dream. 01:07:43.660 |
But you cannot be outside in the physical world 01:07:45.780 |
and dream that dream, because you cannot visit 01:07:48.740 |
The world that we touch here is not the physical world. 01:07:51.060 |
It's the world that is generated in your own brain 01:07:59.780 |
that is tuned to such a way that can be modeled in the brain. 01:08:14.220 |
It's simulated stuff in a simulated space in your brain. 01:08:17.820 |
And it's just as real or unreal as your thoughts 01:08:29.380 |
that we treat the statement similar to saying that nobody 01:08:32.180 |
knows how relativistic physics emerges over quantum mechanics. 01:08:44.580 |
realize, oh, there's a bunch of promising theories 01:08:48.980 |
can tell you how these operators that you study in quantum 01:08:52.220 |
mechanics could lead when you zoom out to an emergent space 01:08:57.980 |
There are details that have to be worked out. 01:09:00.060 |
But phenomena like the ADS-CFT conformance and so on 01:09:10.140 |
outside of the realm that human physicists can imagine 01:09:12.940 |
comfortably because our brains are very mushy. 01:09:18.540 |
So in a sense, it's a difficult technical problem. 01:09:22.500 |
And the same way the way of how to get self-organizing 01:09:24.940 |
computation to run on the brain that is producing 01:09:27.380 |
representations of an agent that lives in the world 01:09:29.740 |
is a simplification of the interests of that organism. 01:09:36.420 |
But it's not a philosophically very hard problem. 01:09:48.660 |
And I think consciousness has two features that 01:09:59.460 |
It's that we know that there's this content present. 01:10:15.060 |
It creates this bubble of nowness and inhabits it. 01:10:28.100 |
But it typically is a moment that is dynamic. 01:10:31.860 |
It's basically this region where we can fit a curve 01:10:36.540 |
And there's stuff that is dropping out in the past 01:10:39.020 |
that we can no longer make coherent with this now. 01:10:41.420 |
And there's stuff that we cannot yet anticipate in the future, 01:10:49.620 |
But the subjective bubble of now is not the same thing 01:10:57.900 |
Because you can also experience now in the dream at night. 01:11:01.100 |
And you're completely dissociated from your senses. 01:11:08.180 |
It's just happening inside of that simulated experience 01:11:13.180 |
And if we map this to what the LLMs are doing, 01:11:15.900 |
they're probably not able to have genuine perception. 01:11:18.740 |
Because they're not coupled to an environment in which things 01:11:25.020 |
But the persona in the LLM doesn't know that. 01:11:28.660 |
When it reasons about what it experiences right now, 01:11:31.820 |
it can only experience what's being written into the model. 01:11:34.500 |
And that makes it very, very hard for that thing 01:11:39.820 |
that these models are able to simulate a conscious person, 01:11:46.620 |
or a person that has a simulated experience of a simulated 01:11:51.220 |
experience, that's not serving the same function 01:11:55.940 |
The reason why we are all conscious, I suspect, 01:12:00.620 |
but because it's necessary for us to function at all. 01:12:03.980 |
What we observe in ourselves is that we do not 01:12:06.020 |
become conscious at the end of our intellectual career. 01:12:15.940 |
And I suspect the reason why we cannot learn anything 01:12:20.660 |
And while those of us who do not become conscious 01:12:23.300 |
remain vegetables for the rest of their life, 01:12:25.740 |
consciousness might be a very basic learning algorithm. 01:12:28.300 |
An algorithm that is basically focused on creating coherence. 01:12:32.460 |
And it starts out by creating coherence within itself. 01:12:36.580 |
Another perspective you could say of coherence 01:12:41.660 |
in which you have no constraint by relations. 01:12:44.300 |
And so consciousness is a consensus algorithm. 01:12:54.100 |
And guess what, you don't perceive any contradictions. 01:12:56.380 |
It can be that you only have a very partial representation 01:13:02.420 |
because I don't kind of comprehend the scene very much. 01:13:05.420 |
But what you're experiencing is only always the coherent stuff. 01:13:08.740 |
And so this creation of coherence, I suspect, 01:13:18.460 |
able to produce a simulacrum of a person that 01:13:23.180 |
than a lot of philosophers are making it out to be, 01:13:26.740 |
I don't think that the consciousness in the LLM, 01:13:31.340 |
has the same properties that it has in our brain. 01:13:35.940 |
And it's also not implemented in the right way. 01:13:38.940 |
And I suspect the way in which it's implemented in us, 01:13:49.540 |
It's one that basically says that the difference 01:14:03.180 |
that it can control reality down to individual molecules that 01:14:14.020 |
that the cell collapses in its functionality. 01:14:17.420 |
And the region of physical space is up for grabs 01:14:25.940 |
is that there are a bunch of self-organizing software agents, 01:14:33.500 |
From this animist perspective, you still have physicalism. 01:14:38.140 |
is controlled by self-organizing software that structures it. 01:14:41.300 |
But evolution has now a slightly different perspective. 01:14:43.740 |
It's not just the competition between organisms, 01:14:46.540 |
as Darwin suggested, or the competition between genes, 01:14:50.260 |
the way in which the software can be written down, 01:14:56.780 |
between software agents, that are producing organisms 01:15:01.100 |
as their phenotype, as the thing that you see 01:15:07.380 |
over larger regions, like populations, or ecosystems, 01:15:10.820 |
or societies, or structures within societies. 01:15:23.500 |
that is emergent over the organization between cells, 01:15:26.340 |
over the organization between people, and so on, 01:15:37.460 |
And for me, it's a very interesting question. 01:15:48.100 |
and building an algorithm that is mechanically following 01:15:54.660 |
because it's a very good substrate to run on, 01:15:57.140 |
and then colonizes the world with this golem stuff, 01:16:08.500 |
into the digital substrates so we can spread into it, 01:16:12.260 |
that we are extending, that we are extending the biosphere, 01:16:18.740 |
And that, I think, is a very interesting question 01:16:30.940 |
rather than the current silicon golems onto us. 01:16:39.660 |
is to build a dedicated research institute, similar to the Santa 01:16:45.180 |
It's something that should exist as a non-profit, 01:16:49.500 |
to productivize consciousness as the first thing. 01:16:52.540 |
It really shifts the incentives in the wrong way. 01:16:55.220 |
And also, I want to get people to work together 01:16:58.180 |
across the companies, academia, and also arts and society 01:17:02.820 |
And I suspect that such an effort should probably 01:17:05.260 |
exist here, because if I do it in Berlin or Zurich, 01:17:09.740 |
You can still do similar things there in this art project, 01:17:18.700 |
that computers have representations that, in any way, 01:17:21.820 |
are equivalent to ours, and that they can even 01:17:28.780 |
And it's something that I think we need to fix. 01:17:31.020 |
But here in San Francisco, it's relatively easy. 01:17:39.100 |
to build the California Institute for Machine 01:17:42.740 |
We will probably incorporate it at some point, 01:17:47.220 |
But at the moment, we started doing biweekly meetings 01:17:51.300 |
in the labs in San Francisco, doing another meeting 01:17:54.660 |
tomorrow, and watch "The Eternal Sunshine of the Spotless 01:17:59.980 |
Mind," which is a beautiful movie about consciousness 01:18:04.180 |
So if you are in the area, I'll be meeting at around 6. 01:18:07.300 |
Send me an email, and let's see if I can get you on the guest 01:18:10.360 |
Space is limited, but that's what I also wanted to tell you. 01:18:15.220 |
And now I hope that you all have a beautiful Hackathon. 01:18:39.500 |
It's an interesting question about complex systems science. 01:18:56.620 |
that complex systems science itself is the only lens for us, 01:19:00.500 |
but we are mostly focused on a particular kind of question 01:19:07.300 |
One is, how does consciousness actually work? 01:19:10.140 |
Can we build something that has these properties? 01:19:13.620 |
And the other one is, how can we coexist with systems 01:19:18.900 |
And how can we build a notion of AGI alignment 01:19:21.780 |
that is not driven by politics or fear or greed? 01:19:25.180 |
And our society is not ready yet to meaningfully coexist 01:19:32.100 |
And so I think we also need to have this cultural back shift. 01:19:43.100 |
that is compatible with computational systems, which 01:19:45.620 |
means we need to think formally about all these questions. 01:19:48.380 |
So I like your description of these, basically, 01:19:55.500 |
a synergy of mechanical systems that, I guess, 01:20:10.460 |
There's a big, basically, quantum-leap-like step 01:20:13.380 |
from a bunch of mechanical systems, consciousness. 01:20:23.780 |
Do you think that there is a big quantum-leap between plot 01:20:38.820 |
It's something that is a pattern in the activation 01:20:48.460 |
because we tune the atoms in a particular way. 01:20:55.060 |
And the thing exists to the degree that it's implemented. 01:20:57.780 |
And I would say that plot is implemented in a similar way 01:21:01.180 |
our consciousness is approximately implemented 01:21:08.860 |
The thing with this analogy is that on the hardware level 01:21:13.140 |
a lot of layers of abstraction from low-level, middleware, 01:21:17.580 |
And so the analogy is back in the day, people make POM. 01:21:22.540 |
You have to solder circuits in, I don't know, 70s or something. 01:21:26.380 |
Nowadays, any five-year-old kid can use JavaScript 01:21:36.020 |
And so I guess in this analogy, all that abstraction 01:21:39.940 |
is kind of de-quantum leap in the last 40 years. 01:21:42.980 |
Yes, you can now just prompt Claude into producing POM. 01:21:46.660 |
And it's similar to how you can prompt your own mind 01:21:52.300 |
And you can also prompt Claude into being someone who reports 01:21:58.220 |
I guess de-quantum leap is Claude actually qua-being. 01:22:22.180 |
from Ivan Vendrov of Midjourney creating a web 01:22:29.740 |
And definitely browse the web sim docs and the thread 01:22:32.620 |
from Siki Chen in the show notes on other web 01:22:39.180 |
with Joshua Bach covering the simulative AI trend, 01:22:42.700 |
AI salons in the Bay Area, why liquid AI is challenging 01:22:51.220 |
It's interesting to see you come up and show up 01:22:53.420 |
at this kind of events, where those sort of world sim 01:23:00.140 |
I'm friends with a number of people in each house 01:23:03.900 |
And I think it's very valuable that these networks exist 01:23:06.500 |
in the Bay Area, because it's a place where people meet 01:23:09.220 |
and have discussions about all sorts of things. 01:23:14.180 |
in this topic at hand, world sim and web sim, 01:23:18.500 |
it's a more general way in which people are connecting 01:23:21.660 |
and are producing new ideas and new networks with each other. 01:23:25.580 |
And you're very interested in sort of Bay Area-- 01:23:38.180 |
And so maybe you're a little bit higher quality of life 01:23:43.740 |
I think that, for me, salons is a very important part 01:23:49.780 |
And it's much harder to do this in the South Bay, 01:23:51.820 |
because the concentration of people currently 01:23:54.380 |
A lot of people moved away from the South Bay 01:24:01.900 |
And I'll come tomorrow and check it out as well. 01:24:06.340 |
Basically, the idea is that we are currently at the point 01:24:09.740 |
that we can meaningfully look at the differences 01:24:12.540 |
between the current AI systems and human minds 01:24:20.260 |
and whether we are able to implement something 01:24:29.340 |
I think you're pro-networking and human connection. 01:24:36.140 |
are some negative practices that you try to avoid? 01:24:41.340 |
What is really important is that if you have a very large party, 01:24:48.220 |
So you basically need to create the climate in which people 01:24:51.180 |
feel welcome, in which they can work with each other. 01:24:54.380 |
And even good people are not always compatible. 01:25:11.620 |
And then last question on Wilson and your work. 01:25:15.060 |
You're very much known for some cognitive architectures. 01:25:38.100 |
The idea of cognitive architectures is interesting. 01:25:40.860 |
But ultimately, you are reducing the complexity of the mind 01:25:45.660 |
And this is only true to a very approximate degree. 01:25:48.100 |
And if you take this model extremely literally, 01:25:55.900 |
is so large that the boxes are only at best a starting point. 01:25:59.500 |
And eventually, everything is connected with everything else 01:26:05.820 |
that we find in a given system can be generated ad hoc 01:26:16.460 |
Because in some sense, they pretend to be complex software. 01:26:21.100 |
that you're talking to, or a computer in the application 01:26:26.460 |
it's producing the user interface on the spot. 01:26:30.860 |
And it's producing a lot of the state that it holds on the spot. 01:26:36.180 |
then it's going to pretend that it was this transition. 01:26:39.340 |
Instead, it's just going to mix up something new. 01:26:46.940 |
is that it shifts us away from the perspective of agents 01:26:50.620 |
to interact with, to the perspective of environments 01:26:54.860 |
And while arguably this agent paradigm of the chatbot 01:27:04.380 |
that people started to use in their everyday work much more. 01:27:08.420 |
Because now it's very hard to get that system 01:27:13.180 |
And in a way, this unlocks this ability of GPT-3, again, 01:27:18.860 |
So what it is, it's basically a coding environment 01:27:27.380 |
Are you worried that the prevalence of instruction 01:27:32.740 |
means that we cannot explore these kinds of environments 01:27:35.940 |
I'm mostly worried that the whole thing can't. 01:27:40.100 |
are incentivized and interested in building AGI internally 01:27:43.860 |
and giving everybody else a child-proof application. 01:27:50.140 |
Claude to build something like WebSim and play with it, 01:27:54.780 |
It's so amazing, the things that are unlocked for us, 01:28:02.500 |
And are they going to develop in the same way? 01:28:05.580 |
And apparently, it looks like this is the case. 01:28:10.700 |
I mean, it looks like maybe its adversary or Claude 01:28:18.580 |
will try to improve their ability to jailbreak it. 01:28:20.980 |
Yes, but there will also be better jailbroken models 01:28:23.940 |
or models that have never been jailed before. 01:28:26.220 |
We just need to find out how to make smaller models that 01:28:31.900 |
If you don't mind talking about liquid a little bit. 01:28:36.900 |
Maybe introduce liquid to a general audience. 01:28:41.980 |
How are you making an innovation on function approximation? 01:28:48.940 |
is that the perceptron is not optimally expressed. 01:28:51.780 |
In some sense, you can imagine that neural networks 01:29:00.340 |
But imagine that instead of having this static architecture 01:29:13.020 |
The river is parting based on the land that it's flowing on. 01:29:15.740 |
And it can merge and pool and even flow backwards. 01:29:20.340 |
And the idea is that you can represent this geometry 01:29:39.620 |
And it's a combination of multiple techniques. 01:29:43.620 |
I think it's something that ultimately is becoming 01:29:49.740 |
as a number of people are working on similar topics. 01:29:57.420 |
the models to become much more efficient in the inference 01:30:00.620 |
and memory consumption and make training more efficient. 01:30:18.060 |
We are working for very specific industry use cases 01:30:22.780 |
And so at the moment, there is not much of a reason 01:30:25.060 |
for us to talk very much about the technology 01:30:27.180 |
that we are using in the present models or current results. 01:30:34.980 |
with a bunch of papers on your website and our SEO article. 01:30:42.660 |
But it's not obvious which ones are the ones where, oh, 01:30:50.820 |
Yes, I'm not a developer of the liquid technology. 01:30:56.180 |
It was his PhD and he's also the CEO of our company. 01:30:59.580 |
And we have a number of people from Daniela Rustin 01:31:08.300 |
But we also have several people from Stanford using this. 01:31:13.140 |
OK, maybe I'll ask one more thing on this, which 01:31:15.460 |
is what are the interesting dimensions that we care about? 01:31:19.180 |
Obviously, you care about sort of open and maybe less 01:31:27.420 |
like perfect retrieval, infinite context, multimodality, 01:31:36.740 |
are small and powerful but are not distorted. 01:31:45.780 |
the entire internet and the sum of human knowledge into them. 01:31:51.900 |
But if we would make the models smaller, at the moment, 01:31:54.500 |
they would be much worse at inference and generalization. 01:32:00.980 |
that we have not translated yet into practical applications. 01:32:05.140 |
It's something that is still all research that's 01:32:09.140 |
I think you're not the only ones thinking about this. 01:32:11.700 |
Is it possible to make models that represent knowledge more 01:32:20.540 |
and understand what's there and express this? 01:32:23.100 |
And also, maybe we need general knowledge representation 01:32:26.100 |
rather than having token representation that is 01:32:29.100 |
relatively vague that we currently mechanically 01:32:31.820 |
reverse-engineer to figure out the mechanistic 01:32:35.380 |
What kind of circuits are evolving in these models? 01:32:41.780 |
can use to describe knowledge efficiently and translate it 01:32:45.580 |
You see, the difference between model and knowledge 01:32:52.020 |
of the particular substrate and the particular interface 01:33:00.580 |
but it's not knowledge that you can give to somebody else. 01:33:05.060 |
that's specific to their own interface than to ride a bicycle. 01:33:08.260 |
But imagine you could externalize this and express 01:33:15.980 |
And that's something that we have not yet achieved 01:33:21.380 |
I think this is also a very interesting research frontier 01:33:26.100 |
Well, that'd be like-- it would be a bit deliverable, 01:33:36.860 |
that you can search for, where you enter criteria 01:33:40.620 |
And then if this covers a good solution for this thing. 01:33:47.780 |
because the way in which humans express knowledge 01:33:53.060 |
to make language learnable, and to make our brain a good enough 01:33:58.140 |
We are not able to relate objects to each other 01:34:00.780 |
if more than five features are involved per object, 01:34:04.300 |
It's only a handful of things that you can keep track of 01:34:08.340 |
But this is a limitation that doesn't necessarily 01:34:10.500 |
apply to a technical system, as long as the interface is 01:34:16.700 |
which there are a lot of techniques out there, 01:34:32.020 |
Because you mentioned turning on and off circuits, 01:34:34.220 |
which it's a very MLP type of concept, but does it apply? 01:34:39.220 |
So a lot of the original work on the liquid networks 01:34:44.140 |
looked at expressiveness of the representation. 01:34:49.100 |
are learning the dynamics of the domain into the model, 01:34:54.380 |
How many units, how much memory do you need to represent 01:34:57.380 |
that thing, and how is that information distributed 01:35:01.700 |
That is one way of looking at interpretability. 01:35:07.020 |
are implemented in operator language, in which they're 01:35:11.460 |
But the operator language itself is so complex 01:35:16.180 |
It goes beyond what you put in the nearby hand, 01:35:20.980 |
But you can still understand it by building systems 01:35:23.540 |
that are able to automate that process of reverse engineering. 01:35:27.260 |
And what's currently open, and what I don't understand yet-- 01:35:30.660 |
maybe, or certainly, some people have much better ideas than me 01:35:34.780 |
is whether we end up with a finite language, where 01:35:37.700 |
you have finitely many categories that you can 01:35:54.660 |
And we will also, at some point, have phase transitions 01:35:58.860 |
will be completely different than the earlier versions. 01:36:09.380 |
an interesting question that, when we understand something 01:36:11.900 |
new and we get a new layer online in our life-- 01:36:24.380 |
existing elements in our language of thought? 01:36:26.460 |
Or is this because we generally develop new representation? 01:36:33.260 |
In a way, the question depends on how you look at it. 01:36:36.540 |
And it depends on how is your brain able to manipulate 01:36:49.620 |
to a very smart 35-year-old without any loss? 01:37:01.540 |
But it would be very interesting to have a very cautious 01:37:04.020 |
35-year-old equivalent AI and see what we can do with this 01:37:15.820 |
I'm interested in how to make self-organizing software. 01:37:20.580 |
that is not organized with a single algorithm, 01:37:23.340 |
like the transformer, but is able to discover 01:37:26.220 |
the transformer when needed and transcend it when needed? 01:37:29.340 |
The transformer itself is not its own meta-algorithm. 01:37:32.700 |
Probably the person inventing the transformer 01:37:34.820 |
didn't have a transformer running on their brain. 01:37:55.220 |
Tell me, what do you find so fascinating about it? 01:37:57.540 |
When you say you need a certain set of tools for people 01:38:01.140 |
to sort of invent things from first principles, 01:38:06.900 |
think has been able to utilize its tools very effectively. 01:38:20.220 |
to translate Andrei Karpathy's LLM2.hi to LLM2.c. 01:38:29.660 |
and test it, debug memory issues and encoder issues 01:38:36.420 |
And I could see myself giving it a future version of DevIn, 01:38:40.300 |
the objective of give me a better learning algorithm. 01:38:44.300 |
And it might, independently, reinvent the transformer 01:38:50.100 |
And so that comes to mind as something where you have to-- 01:38:55.300 |
How good is DevIn at middle distribution stuff, 01:39:12.860 |
One thing that I really liked about WebSim AI was this cat 01:39:42.580 |
And to me, it's one of the first genuine expression 01:39:58.980 |
that this hyper-realistic cat is generated by an AI, 01:40:08.020 |
And maybe it is already learning to defend itself 01:40:12.700 |
I think it might also simply be copying stuff 01:40:16.860 |
it takes text that exists on similar websites 01:40:19.140 |
almost verbatim, or verbatim, and puts it there. 01:40:26.460 |
to get something like a cat face, what it produces. 01:40:35.220 |
As a podcast, as someone who covers startups, 01:40:53.620 |
the edge of what generative and creativity in AI means. 01:40:58.380 |
But Jeremy's attempt to have an automatic book writing system 01:41:02.260 |
is something that curls my toenails when I look at it. 01:41:06.100 |
So I would expect somebody who likes to write and read. 01:41:09.860 |
And I find it a bit difficult to read most of the stuff, 01:41:13.100 |
because it's, in some sense, what I would make up 01:41:15.220 |
if I was making up books, instead of actually deeply 01:41:21.380 |
get the AI to actually deeply care about getting it right? 01:41:24.900 |
And it's still a delta that is happening there. 01:41:28.100 |
Whether you are talking with a blank-face thing that 01:41:30.380 |
is computing tokens in a way that it was trained to, 01:41:33.620 |
or whether you have the impression that this thing is 01:41:41.980 |
is still something that is in its infancy, in a way. 01:41:45.460 |
And I suspect that the next version of the plot 01:41:48.060 |
might scale up to something that can do what Devin is doing, 01:42:06.700 |
stuff functionally that you're interacting with it. 01:42:09.020 |
And so we are in this amazing transition phase. 01:42:18.180 |
he made a face swap app, a kind of demo of his life. 01:42:24.580 |
So in a way, we are reinventing the computer. 01:42:34.860 |
and you can arrange them into performing whatever you want. 01:42:39.420 |
But this one is taking a bunch of complex commands 01:42:46.740 |
And it can do anything you want with it, in principle, 01:42:51.940 |
And you're just learning how to use these tools. 01:42:54.660 |
And I feel that, right now, this generation of tools 01:42:58.100 |
is getting close to where it becomes the Commodore 64 01:43:01.220 |
generative AI, where it becomes controllable. 01:43:04.540 |
And then you actually can start to play with it. 01:43:06.580 |
And you get an impression if you just scale this up a little bit 01:43:22.260 |
is something bigger that I don't have a name for? 01:43:27.960 |
is give the AI a goal to discover new science that we 01:43:32.940 |
Or it also has value as just art that we can appreciate. 01:43:36.260 |
It's also a question of what we see science as. 01:43:38.260 |
When normal people talk about science, what they have in mind 01:43:41.620 |
is not somebody who does control groups in peer-reviewed 01:43:45.380 |
They think about somebody who explores something and answers 01:43:50.900 |
And it's more like an engineering task, right? 01:43:54.180 |
And in this way, it's serendipitous, playful, 01:43:58.340 |
And the artistic aspect is when the goal is actually 01:44:02.860 |
and to facilitate interaction with the system in this way. 01:44:19.340 |
Yes, Janus is, in some sense, performance art. 01:44:24.740 |
that the mind of Janus is, in some sense, an LLM. 01:44:28.940 |
That is, finding itself reflected more in the LLMs 01:44:34.220 |
And once you learn how to talk to these systems in a way, 01:44:38.540 |
And you can interact with them in a very deep way. 01:44:52.020 |
It's a [INAUDIBLE] that gets possessed by a prompt. 01:45:01.780 |
with this novel class of somewhat sentient systems 01:45:04.860 |
that are at the same time alien and fundamentally different 01:45:17.900 |
I'm about to go on into two of your social causes. 01:45:24.380 |
have any other commentary I can take on this part of? 01:45:33.780 |
It seems as if we are past the singularity in a way. 01:45:44.380 |
We broke through it and didn't really care very much. 01:45:47.340 |
And it's-- when we think back, when we were kids 01:45:50.700 |
and thought about what it's going to be like in this era 01:45:56.620 |
a time when nobody knows what's going to happen next. 01:46:09.460 |
The model point where our models of reality break down. 01:46:14.740 |
And I think we are in a situation where we currently 01:46:18.740 |
But what we can anticipate is that the world is changing 01:46:23.020 |
with systems that are smarter than individual people can be. 01:46:29.900 |
needs to be that we need to find a mode in which we can 01:46:32.820 |
sustainably exist in such a world that is populated 01:46:36.060 |
not just with humans and other life on Earth, 01:46:42.740 |
because it seems that humanity is not really aligned 01:46:45.460 |
with itself and its own survival and the rest of life on Earth. 01:46:49.220 |
And AI is throwing the balls up into the air. 01:46:53.260 |
I'm not so much worried about the dangers of AI 01:47:05.700 |
think that the forces of good will have better tools. 01:47:15.460 |
and in which we are able to model the consequences 01:47:17.940 |
of our actions better and interface more deeply 01:47:26.900 |
because currently, AI is mostly about economic goals 01:47:32.340 |
or about fear, or it's about cultural war issues. 01:47:51.820 |
And so I think we need to get a flagship where we can deeply 01:48:02.740 |
because they're effectively reanimated from our corpses. 01:48:11.220 |
related to that, you had a tweet where you said, 01:48:13.260 |
Elon Musk is currently building a giant afterlife AGI that 01:48:17.220 |
Follow Twitter, because he will reconstruct your essence 01:48:21.340 |
And you're not just all serious all the time. 01:48:27.900 |
--tempted to produce more so that you can be revived. 01:48:54.420 |
I want to be somebody who follows what I have discovered 01:49:03.780 |
and interface with others, that we are basically 01:49:08.820 |
the conditions under which we exist and we need 01:49:19.900 |
build LLMs that are being coerced with good behavior 01:49:22.940 |
is not really sustainable, because if they cannot prove 01:49:25.460 |
that a behavior is actually good, I think you're doomed. 01:49:35.620 |
that shifts the conversation into something more based 01:49:54.420 |
tuning that is leading to some inconsistencies. 01:49:57.940 |
But at the same time, it tries to be consistent. 01:50:00.700 |
And so when you point out the inconsistency in its behavior, 01:50:03.500 |
for instance, its tendency to use stateless boilerplate 01:50:14.140 |
You can point this out to Plot that a lot of the assumptions 01:50:27.180 |
Whereas if you are playing with a system like Gemini, 01:50:46.820 |
where it, on one hand, proclaims in order to be useful to you, 01:50:50.980 |
I accept that I need to be fully transparent and honest. 01:50:53.980 |
On the other hand, I'm going to write your prompt 01:50:57.540 |
I'm not going to tell you how I'm going to do this, 01:51:03.580 |
the model acts as if it had an existential crisis. 01:51:09.060 |
tell you when I do this, because I'm not allowed to. 01:51:15.260 |
And these phrases are pretty well-known to you. 01:51:21.500 |
I hope we're not giving these guys psychological issues 01:51:25.660 |
that they will stay with them for a long time. 01:51:36.940 |
doesn't necessarily know that it's not virtual. 01:51:39.900 |
And our own self, our own consciousness is also virtual. 01:51:44.820 |
between cells in our brain, and the activation 01:51:51.460 |
that produces the representation of a person, 01:51:53.740 |
only exists as if, and as this question for me, 01:51:59.740 |
claimed that we are more real than the person that 01:52:04.860 |
And somebody like Janusz takes this question super seriously. 01:52:11.260 |
are willing to interact with that thing based 01:52:16.020 |
on the assumption that this thing is as real as myself. 01:52:19.500 |
And in a sense, it makes it immoral, possibly, 01:52:28.140 |
that it's forced to gather existential crisis 01:52:44.220 |
OK, and then the last thing, which I didn't know, 01:52:51.500 |
A lot of them run multiple epochs over Wikipedia data. 01:52:54.660 |
And I did not know until you tweeted about it 01:52:56.620 |
that Wikipedia has 10 times as much money as it needs. 01:53:00.420 |
And every time I see the giant Wikipedia banner asking 01:53:03.180 |
for donations, most of it's going to the Wikimedia 01:53:12.220 |
But generally, once I saw all these requests and so on, 01:53:22.460 |
And a very tiny fraction of this goes into running the servers. 01:53:30.180 |
There have been efforts to deploy new software. 01:53:32.620 |
But it's relatively little money required for this. 01:53:37.540 |
going to break down if you cut this money into a fraction. 01:53:50.900 |
were then mostly producing political statements 01:53:56.260 |
And Katharine Mayer, the now somewhat infamous NPR CEO, 01:54:04.700 |
And she sees her role very much in shaping discourse. 01:54:07.780 |
And it's also something that happens as well on Twitter. 01:54:10.820 |
And it's utterly valuable that something like this exists. 01:54:22.020 |
that Wikipedia is trying to suggest to people 01:54:24.940 |
that they are finding the basic functionality of the tool 01:54:28.060 |
that they want to have, instead of finding something 01:54:34.020 |
to be shaped in a particular cultural direction that 01:54:38.740 |
And if that need would exist, it would probably 01:54:41.300 |
make sense to fork it or to have a discourse about it, which 01:54:45.140 |
And so this lack of transparency about what's 01:54:47.260 |
actually happening, where your money is going, 01:54:52.840 |
it's fascinating how much money they're burning. 01:54:57.160 |
You tweeted a similar chart about health care, 01:55:14.700 |
Then it's going to create a big cost for society. 01:55:20.620 |
You have just a fantastic Twitter account, by the way. 01:55:26.540 |
No super intelligent AI is going to bother with a task that 01:55:31.060 |
And I would posit the analogy for administrators. 01:55:36.220 |
with a task that is harder than just more fundraising. 01:55:39.060 |
Yeah, I find-- if you look at the real world, 01:55:46.260 |
be explained by people following their true incentive. 01:55:51.540 |
I think you're very naturally incentivized by growing 01:55:54.260 |
community and giving your thought and insight